summarylogtreecommitdiffstats
path: root/0005-glitched-pds.patch
diff options
context:
space:
mode:
authorKyle Devir2020-02-01 23:58:35 +1000
committerKyle Devir2020-02-01 23:58:35 +1000
commit39d73678a45ff0c20c66b6d8ab58328192dcb971 (patch)
tree969a4d03180d4ac62e7a8f57e4ac414766966a53 /0005-glitched-pds.patch
parentf83525958f388fee4dd51de2b08f5858c0121a7a (diff)
downloadaur-39d73678a45ff0c20c66b6d8ab58328192dcb971.tar.gz
5.5.arch1
Diffstat (limited to '0005-glitched-pds.patch')
-rw-r--r--0005-glitched-pds.patch55
1 files changed, 0 insertions, 55 deletions
diff --git a/0005-glitched-pds.patch b/0005-glitched-pds.patch
index 21f2d696f687..ae589dd67e4b 100644
--- a/0005-glitched-pds.patch
+++ b/0005-glitched-pds.patch
@@ -89,61 +89,6 @@ index 9270a4370d54..30d01e647417 100644
* The total number of pages which are beyond the high watermark within all
* zones.
-diff --git a/kernel/sched/pds.c b/kernel/sched/pds.c
-index c2d831b242b6d18a47e0d87a9f5433a7748b52ff..5bc8d7a8f920c21feab69b2706a3328dc8d39f9a 100644
---- a/kernel/sched/pds.c
-+++ b/kernel/sched/pds.c
-@@ -409,12 +409,11 @@ struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf)
- * [L] ->on_rq
- * RELEASE (rq->lock)
- *
-- * If we observe the old CPU in task_rq_lock(), the acquire of
-+ * If we observe the old CPU in task_rq_lock, the acquire of
- * the old rq->lock will fully serialize against the stores.
- *
-- * If we observe the new CPU in task_rq_lock(), the address
-- * dependency headed by '[L] rq = task_rq()' and the acquire
-- * will pair with the WMB to ensure we then also see migrating.
-+ * If we observe the new CPU in task_rq_lock, the acquire will
-+ * pair with the WMB to ensure we must then also see migrating.
- */
- if (likely(rq == task_rq(p) && !task_on_rq_migrating(p))) {
- return rq;
-@@ -952,9 +953,9 @@ static inline void __set_task_cpu(struct task_struct *p, unsigned int cpu)
- smp_wmb();
-
- #ifdef CONFIG_THREAD_INFO_IN_TASK
-- WRITE_ONCE(p->cpu, cpu);
-+ p->cpu = cpu;
- #else
-- WRITE_ONCE(task_thread_info(p)->cpu, cpu);
-+ task_thread_info(p)->cpu = cpu;
- #endif
- #endif
- }
-@@ -1035,7 +1036,7 @@ static void detach_task(struct rq *rq, struct task_struct *p, int target_cpu)
- {
- lockdep_assert_held(&rq->lock);
-
-- WRITE_ONCE(p->on_rq ,TASK_ON_RQ_MIGRATING);
-+ p->on_rq = TASK_ON_RQ_MIGRATING;
- if (task_contributes_to_load(p))
- rq->nr_uninterruptible++;
- dequeue_task(p, rq, 0);
-diff --git a/kernel/sched/pds_sched.h b/kernel/sched/pds_sched.h
-index 20dcf19ea057627d91be07b4ec20f0827c30084c..24fa90ca63d144cc4f45d82d88407ea70d2d2edf 100644
---- a/kernel/sched/pds_sched.h
-+++ b/kernel/sched/pds_sched.h
-@@ -56,7 +56,7 @@ static inline int task_on_rq_queued(struct task_struct *p)
-
- static inline int task_on_rq_migrating(struct task_struct *p)
- {
-- return READ_ONCE(p->on_rq) == TASK_ON_RQ_MIGRATING;
-+ return p->on_rq == TASK_ON_RQ_MIGRATING;
- }
-
- enum {
-
diff --git a/init/Kconfig b/init/Kconfig
index 11fd9b502d06..e9bc34d3019b 100644
--- a/init/Kconfig