Skip to content

Commit f97bb52

Browse files
author
Peter Zijlstra
committed
sched: Fix data-race in wakeup
Mel reported that on some ARM64 platforms loadavg goes bananas and Will tracked it down to the following race: CPU0 CPU1 schedule() prev->sched_contributes_to_load = X; deactivate_task(prev); try_to_wake_up() if (p->on_rq &&) // false if (smp_load_acquire(&p->on_cpu) && // true ttwu_queue_wakelist()) p->sched_remote_wakeup = Y; smp_store_release(prev->on_cpu, 0); where both p->sched_contributes_to_load and p->sched_remote_wakeup are in the same word, and thus the stores X and Y race (and can clobber one another's data). Whereas prior to commit c6e7bd7 ("sched/core: Optimize ttwu() spinning on p->on_cpu") the p->on_cpu handoff serialized access to p->sched_remote_wakeup (just as it still does with p->sched_contributes_to_load) that commit broke that by calling ttwu_queue_wakelist() with p->on_cpu != 0. However, due to p->XXX = X ttwu() schedule() if (p->on_rq && ...) // false smp_mb__after_spinlock() if (smp_load_acquire(&p->on_cpu) && deactivate_task() ttwu_queue_wakelist()) p->on_rq = 0; p->sched_remote_wakeup = Y; We can be sure any 'current' store is complete and 'current' is guaranteed asleep. Therefore we can move p->sched_remote_wakeup into the current flags word. Note: while the observed failure was loadavg accounting gone wrong due to ttwu() cobbering p->sched_contributes_to_load, the reverse problem is also possible where schedule() clobbers p->sched_remote_wakeup, this could result in enqueue_entity() wrecking ->vruntime and causing scheduling artifacts. Fixes: c6e7bd7 ("sched/core: Optimize ttwu() spinning on p->on_cpu") Reported-by: Mel Gorman <[email protected]> Debugged-by: Will Deacon <[email protected]> Signed-off-by: Peter Zijlstra (Intel) <[email protected]> Link: https://lkml.kernel.org/r/[email protected]
1 parent 8e1ac42 commit f97bb52

File tree

1 file changed

+15
-1
lines changed

1 file changed

+15
-1
lines changed

include/linux/sched.h

Lines changed: 15 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -769,7 +769,6 @@ struct task_struct {
769769
unsigned sched_reset_on_fork:1;
770770
unsigned sched_contributes_to_load:1;
771771
unsigned sched_migrated:1;
772-
unsigned sched_remote_wakeup:1;
773772
#ifdef CONFIG_PSI
774773
unsigned sched_psi_wake_requeue:1;
775774
#endif
@@ -779,6 +778,21 @@ struct task_struct {
779778

780779
/* Unserialized, strictly 'current' */
781780

781+
/*
782+
* This field must not be in the scheduler word above due to wakelist
783+
* queueing no longer being serialized by p->on_cpu. However:
784+
*
785+
* p->XXX = X; ttwu()
786+
* schedule() if (p->on_rq && ..) // false
787+
* smp_mb__after_spinlock(); if (smp_load_acquire(&p->on_cpu) && //true
788+
* deactivate_task() ttwu_queue_wakelist())
789+
* p->on_rq = 0; p->sched_remote_wakeup = Y;
790+
*
791+
* guarantees all stores of 'current' are visible before
792+
* ->sched_remote_wakeup gets used, so it can be in this word.
793+
*/
794+
unsigned sched_remote_wakeup:1;
795+
782796
/* Bit to tell LSMs we're in execve(): */
783797
unsigned in_execve:1;
784798
unsigned in_iowait:1;

0 commit comments

Comments
 (0)