dfl_pwq           257 kernel/workqueue.c 	struct pool_workqueue	*dfl_pwq;	/* PW: only for unbound wqs */
dfl_pwq           576 kernel/workqueue.c 		return wq->dfl_pwq;
dfl_pwq          3874 kernel/workqueue.c 	struct pool_workqueue	*dfl_pwq;
dfl_pwq          3886 kernel/workqueue.c 		put_pwq_unlocked(ctx->dfl_pwq);
dfl_pwq          3934 kernel/workqueue.c 	ctx->dfl_pwq = alloc_unbound_pwq(wq, new_attrs);
dfl_pwq          3935 kernel/workqueue.c 	if (!ctx->dfl_pwq)
dfl_pwq          3944 kernel/workqueue.c 			ctx->dfl_pwq->refcnt++;
dfl_pwq          3945 kernel/workqueue.c 			ctx->pwq_tbl[node] = ctx->dfl_pwq;
dfl_pwq          3981 kernel/workqueue.c 	link_pwq(ctx->dfl_pwq);
dfl_pwq          3982 kernel/workqueue.c 	swap(ctx->wq->dfl_pwq, ctx->dfl_pwq);
dfl_pwq          4114 kernel/workqueue.c 	if (wq_calc_node_cpumask(wq->dfl_pwq->pool->attrs, node, cpu_off, cpumask)) {
dfl_pwq          4136 kernel/workqueue.c 	spin_lock_irq(&wq->dfl_pwq->pool->lock);
dfl_pwq          4137 kernel/workqueue.c 	get_pwq(wq->dfl_pwq);
dfl_pwq          4138 kernel/workqueue.c 	spin_unlock_irq(&wq->dfl_pwq->pool->lock);
dfl_pwq          4139 kernel/workqueue.c 	old_pwq = numa_pwq_tbl_install(wq, node, wq->dfl_pwq);
dfl_pwq          4174 kernel/workqueue.c 		WARN(!ret && (wq->pwqs.next != &wq->dfl_pwq->pwqs_node ||
dfl_pwq          4175 kernel/workqueue.c 			      wq->pwqs.prev != &wq->dfl_pwq->pwqs_node),
dfl_pwq          4373 kernel/workqueue.c 		if (WARN_ON((pwq != wq->dfl_pwq) && (pwq->refcnt > 1)) ||
dfl_pwq          4414 kernel/workqueue.c 		pwq = wq->dfl_pwq;
dfl_pwq          4415 kernel/workqueue.c 		wq->dfl_pwq = NULL;