dfl_pwq 257 kernel/workqueue.c struct pool_workqueue *dfl_pwq; /* PW: only for unbound wqs */ dfl_pwq 576 kernel/workqueue.c return wq->dfl_pwq; dfl_pwq 3874 kernel/workqueue.c struct pool_workqueue *dfl_pwq; dfl_pwq 3886 kernel/workqueue.c put_pwq_unlocked(ctx->dfl_pwq); dfl_pwq 3934 kernel/workqueue.c ctx->dfl_pwq = alloc_unbound_pwq(wq, new_attrs); dfl_pwq 3935 kernel/workqueue.c if (!ctx->dfl_pwq) dfl_pwq 3944 kernel/workqueue.c ctx->dfl_pwq->refcnt++; dfl_pwq 3945 kernel/workqueue.c ctx->pwq_tbl[node] = ctx->dfl_pwq; dfl_pwq 3981 kernel/workqueue.c link_pwq(ctx->dfl_pwq); dfl_pwq 3982 kernel/workqueue.c swap(ctx->wq->dfl_pwq, ctx->dfl_pwq); dfl_pwq 4114 kernel/workqueue.c if (wq_calc_node_cpumask(wq->dfl_pwq->pool->attrs, node, cpu_off, cpumask)) { dfl_pwq 4136 kernel/workqueue.c spin_lock_irq(&wq->dfl_pwq->pool->lock); dfl_pwq 4137 kernel/workqueue.c get_pwq(wq->dfl_pwq); dfl_pwq 4138 kernel/workqueue.c spin_unlock_irq(&wq->dfl_pwq->pool->lock); dfl_pwq 4139 kernel/workqueue.c old_pwq = numa_pwq_tbl_install(wq, node, wq->dfl_pwq); dfl_pwq 4174 kernel/workqueue.c WARN(!ret && (wq->pwqs.next != &wq->dfl_pwq->pwqs_node || dfl_pwq 4175 kernel/workqueue.c wq->pwqs.prev != &wq->dfl_pwq->pwqs_node), dfl_pwq 4373 kernel/workqueue.c if (WARN_ON((pwq != wq->dfl_pwq) && (pwq->refcnt > 1)) || dfl_pwq 4414 kernel/workqueue.c pwq = wq->dfl_pwq; dfl_pwq 4415 kernel/workqueue.c wq->dfl_pwq = NULL;