Skip to content

Commit 4a6c560

Browse files
committed
workqueue: Make sure that wq_unbound_cpumask is never empty
During boot, depending on how the housekeeping and workqueue.unbound_cpus masks are set, wq_unbound_cpumask can end up empty. Since 8639ece ("workqueue: Implement non-strict affinity scope for unbound workqueues"), this may end up feeding -1 as a CPU number into scheduler leading to oopses. BUG: unable to handle page fault for address: ffffffff8305e9c0 #PF: supervisor read access in kernel mode #PF: error_code(0x0000) - not-present page ... Call Trace: <TASK> select_idle_sibling+0x79/0xaf0 select_task_rq_fair+0x1cb/0x7b0 try_to_wake_up+0x29c/0x5c0 wake_up_process+0x19/0x20 kick_pool+0x5e/0xb0 __queue_work+0x119/0x430 queue_work_on+0x29/0x30 ... An empty wq_unbound_cpumask is a clear misconfiguration and already disallowed once system is booted up. Let's warn on and ignore unbound_cpumask restrictions which lead to no unbound cpus. While at it, also remove now unncessary empty check on wq_unbound_cpumask in wq_select_unbound_cpu(). Signed-off-by: Tejun Heo <[email protected]> Reported-and-Tested-by: Yong He <[email protected]> Link: http://lkml.kernel.org/r/[email protected] Fixes: 8639ece ("workqueue: Implement non-strict affinity scope for unbound workqueues") Cc: [email protected] # v6.6+ Reviewed-by: Waiman Long <[email protected]>
1 parent a12deb4 commit 4a6c560

File tree

1 file changed

+15
-7
lines changed

1 file changed

+15
-7
lines changed

kernel/workqueue.c

Lines changed: 15 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1684,9 +1684,6 @@ static int wq_select_unbound_cpu(int cpu)
16841684
pr_warn_once("workqueue: round-robin CPU selection forced, expect performance impact\n");
16851685
}
16861686

1687-
if (cpumask_empty(wq_unbound_cpumask))
1688-
return cpu;
1689-
16901687
new_cpu = __this_cpu_read(wq_rr_cpu_last);
16911688
new_cpu = cpumask_next_and(new_cpu, wq_unbound_cpumask, cpu_online_mask);
16921689
if (unlikely(new_cpu >= nr_cpu_ids)) {
@@ -6515,6 +6512,17 @@ static inline void wq_watchdog_init(void) { }
65156512

65166513
#endif /* CONFIG_WQ_WATCHDOG */
65176514

6515+
static void __init restrict_unbound_cpumask(const char *name, const struct cpumask *mask)
6516+
{
6517+
if (!cpumask_intersects(wq_unbound_cpumask, mask)) {
6518+
pr_warn("workqueue: Restricting unbound_cpumask (%*pb) with %s (%*pb) leaves no CPU, ignoring\n",
6519+
cpumask_pr_args(wq_unbound_cpumask), name, cpumask_pr_args(mask));
6520+
return;
6521+
}
6522+
6523+
cpumask_and(wq_unbound_cpumask, wq_unbound_cpumask, mask);
6524+
}
6525+
65186526
/**
65196527
* workqueue_init_early - early init for workqueue subsystem
65206528
*
@@ -6534,11 +6542,11 @@ void __init workqueue_init_early(void)
65346542
BUILD_BUG_ON(__alignof__(struct pool_workqueue) < __alignof__(long long));
65356543

65366544
BUG_ON(!alloc_cpumask_var(&wq_unbound_cpumask, GFP_KERNEL));
6537-
cpumask_copy(wq_unbound_cpumask, housekeeping_cpumask(HK_TYPE_WQ));
6538-
cpumask_and(wq_unbound_cpumask, wq_unbound_cpumask, housekeeping_cpumask(HK_TYPE_DOMAIN));
6539-
6545+
cpumask_copy(wq_unbound_cpumask, cpu_possible_mask);
6546+
restrict_unbound_cpumask("HK_TYPE_WQ", housekeeping_cpumask(HK_TYPE_WQ));
6547+
restrict_unbound_cpumask("HK_TYPE_DOMAIN", housekeeping_cpumask(HK_TYPE_DOMAIN));
65406548
if (!cpumask_empty(&wq_cmdline_cpumask))
6541-
cpumask_and(wq_unbound_cpumask, wq_unbound_cpumask, &wq_cmdline_cpumask);
6549+
restrict_unbound_cpumask("workqueue.unbound_cpus", &wq_cmdline_cpumask);
65426550

65436551
pwq_cache = KMEM_CACHE(pool_workqueue, SLAB_PANIC);
65446552

0 commit comments

Comments
 (0)