diff options
author | Kyle Devir | 2020-02-01 23:58:35 +1000 |
---|---|---|
committer | Kyle Devir | 2020-02-01 23:58:35 +1000 |
commit | 39d73678a45ff0c20c66b6d8ab58328192dcb971 (patch) | |
tree | 969a4d03180d4ac62e7a8f57e4ac414766966a53 /0005-glitched-pds.patch | |
parent | f83525958f388fee4dd51de2b08f5858c0121a7a (diff) | |
download | aur-39d73678a45ff0c20c66b6d8ab58328192dcb971.tar.gz |
5.5.arch1
Diffstat (limited to '0005-glitched-pds.patch')
-rw-r--r-- | 0005-glitched-pds.patch | 55 |
1 files changed, 0 insertions, 55 deletions
diff --git a/0005-glitched-pds.patch b/0005-glitched-pds.patch index 21f2d696f687..ae589dd67e4b 100644 --- a/0005-glitched-pds.patch +++ b/0005-glitched-pds.patch @@ -89,61 +89,6 @@ index 9270a4370d54..30d01e647417 100644 * The total number of pages which are beyond the high watermark within all * zones. -diff --git a/kernel/sched/pds.c b/kernel/sched/pds.c -index c2d831b242b6d18a47e0d87a9f5433a7748b52ff..5bc8d7a8f920c21feab69b2706a3328dc8d39f9a 100644 ---- a/kernel/sched/pds.c -+++ b/kernel/sched/pds.c -@@ -409,12 +409,11 @@ struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) - * [L] ->on_rq - * RELEASE (rq->lock) - * -- * If we observe the old CPU in task_rq_lock(), the acquire of -+ * If we observe the old CPU in task_rq_lock, the acquire of - * the old rq->lock will fully serialize against the stores. - * -- * If we observe the new CPU in task_rq_lock(), the address -- * dependency headed by '[L] rq = task_rq()' and the acquire -- * will pair with the WMB to ensure we then also see migrating. -+ * If we observe the new CPU in task_rq_lock, the acquire will -+ * pair with the WMB to ensure we must then also see migrating. - */ - if (likely(rq == task_rq(p) && !task_on_rq_migrating(p))) { - return rq; -@@ -952,9 +953,9 @@ static inline void __set_task_cpu(struct task_struct *p, unsigned int cpu) - smp_wmb(); - - #ifdef CONFIG_THREAD_INFO_IN_TASK -- WRITE_ONCE(p->cpu, cpu); -+ p->cpu = cpu; - #else -- WRITE_ONCE(task_thread_info(p)->cpu, cpu); -+ task_thread_info(p)->cpu = cpu; - #endif - #endif - } -@@ -1035,7 +1036,7 @@ static void detach_task(struct rq *rq, struct task_struct *p, int target_cpu) - { - lockdep_assert_held(&rq->lock); - -- WRITE_ONCE(p->on_rq ,TASK_ON_RQ_MIGRATING); -+ p->on_rq = TASK_ON_RQ_MIGRATING; - if (task_contributes_to_load(p)) - rq->nr_uninterruptible++; - dequeue_task(p, rq, 0); -diff --git a/kernel/sched/pds_sched.h b/kernel/sched/pds_sched.h -index 20dcf19ea057627d91be07b4ec20f0827c30084c..24fa90ca63d144cc4f45d82d88407ea70d2d2edf 100644 ---- a/kernel/sched/pds_sched.h -+++ b/kernel/sched/pds_sched.h -@@ -56,7 +56,7 @@ static inline int task_on_rq_queued(struct task_struct *p) - - static inline int task_on_rq_migrating(struct task_struct *p) - { -- return READ_ONCE(p->on_rq) == TASK_ON_RQ_MIGRATING; -+ return p->on_rq == TASK_ON_RQ_MIGRATING; - } - - enum { - diff --git a/init/Kconfig b/init/Kconfig index 11fd9b502d06..e9bc34d3019b 100644 --- a/init/Kconfig |