1
2
3
4
5
6
7struct sched_core_cookie {
8 refcount_t refcnt;
9};
10
11static unsigned long sched_core_alloc_cookie(void)
12{
13 struct sched_core_cookie *ck = kmalloc(sizeof(*ck), GFP_KERNEL);
14 if (!ck)
15 return 0;
16
17 refcount_set(&ck->refcnt, 1);
18 sched_core_get();
19
20 return (unsigned long)ck;
21}
22
23static void sched_core_put_cookie(unsigned long cookie)
24{
25 struct sched_core_cookie *ptr = (void *)cookie;
26
27 if (ptr && refcount_dec_and_test(&ptr->refcnt)) {
28 kfree(ptr);
29 sched_core_put();
30 }
31}
32
33static unsigned long sched_core_get_cookie(unsigned long cookie)
34{
35 struct sched_core_cookie *ptr = (void *)cookie;
36
37 if (ptr)
38 refcount_inc(&ptr->refcnt);
39
40 return cookie;
41}
42
43
44
45
46
47
48
49
50
51
52
53static unsigned long sched_core_update_cookie(struct task_struct *p,
54 unsigned long cookie)
55{
56 unsigned long old_cookie;
57 struct rq_flags rf;
58 struct rq *rq;
59 bool enqueued;
60
61 rq = task_rq_lock(p, &rf);
62
63
64
65
66
67
68
69 SCHED_WARN_ON((p->core_cookie || cookie) && !sched_core_enabled(rq));
70
71 enqueued = sched_core_enqueued(p);
72 if (enqueued)
73 sched_core_dequeue(rq, p, DEQUEUE_SAVE);
74
75 old_cookie = p->core_cookie;
76 p->core_cookie = cookie;
77
78 if (enqueued)
79 sched_core_enqueue(rq, p);
80
81
82
83
84
85
86
87
88
89
90 if (task_running(rq, p))
91 resched_curr(rq);
92
93 task_rq_unlock(rq, p, &rf);
94
95 return old_cookie;
96}
97
98static unsigned long sched_core_clone_cookie(struct task_struct *p)
99{
100 unsigned long cookie, flags;
101
102 raw_spin_lock_irqsave(&p->pi_lock, flags);
103 cookie = sched_core_get_cookie(p->core_cookie);
104 raw_spin_unlock_irqrestore(&p->pi_lock, flags);
105
106 return cookie;
107}
108
109void sched_core_fork(struct task_struct *p)
110{
111 RB_CLEAR_NODE(&p->core_node);
112 p->core_cookie = sched_core_clone_cookie(current);
113}
114
115void sched_core_free(struct task_struct *p)
116{
117 sched_core_put_cookie(p->core_cookie);
118}
119
120static void __sched_core_set(struct task_struct *p, unsigned long cookie)
121{
122 cookie = sched_core_get_cookie(cookie);
123 cookie = sched_core_update_cookie(p, cookie);
124 sched_core_put_cookie(cookie);
125}
126
127
128int sched_core_share_pid(unsigned int cmd, pid_t pid, enum pid_type type,
129 unsigned long uaddr)
130{
131 unsigned long cookie = 0, id = 0;
132 struct task_struct *task, *p;
133 struct pid *grp;
134 int err = 0;
135
136 if (!static_branch_likely(&sched_smt_present))
137 return -ENODEV;
138
139 BUILD_BUG_ON(PR_SCHED_CORE_SCOPE_THREAD != PIDTYPE_PID);
140 BUILD_BUG_ON(PR_SCHED_CORE_SCOPE_THREAD_GROUP != PIDTYPE_TGID);
141 BUILD_BUG_ON(PR_SCHED_CORE_SCOPE_PROCESS_GROUP != PIDTYPE_PGID);
142
143 if (type > PIDTYPE_PGID || cmd >= PR_SCHED_CORE_MAX || pid < 0 ||
144 (cmd != PR_SCHED_CORE_GET && uaddr))
145 return -EINVAL;
146
147 rcu_read_lock();
148 if (pid == 0) {
149 task = current;
150 } else {
151 task = find_task_by_vpid(pid);
152 if (!task) {
153 rcu_read_unlock();
154 return -ESRCH;
155 }
156 }
157 get_task_struct(task);
158 rcu_read_unlock();
159
160
161
162
163
164 if (!ptrace_may_access(task, PTRACE_MODE_READ_REALCREDS)) {
165 err = -EPERM;
166 goto out;
167 }
168
169 switch (cmd) {
170 case PR_SCHED_CORE_GET:
171 if (type != PIDTYPE_PID || uaddr & 7) {
172 err = -EINVAL;
173 goto out;
174 }
175 cookie = sched_core_clone_cookie(task);
176 if (cookie) {
177
178 ptr_to_hashval((void *)cookie, &id);
179 }
180 err = put_user(id, (u64 __user *)uaddr);
181 goto out;
182
183 case PR_SCHED_CORE_CREATE:
184 cookie = sched_core_alloc_cookie();
185 if (!cookie) {
186 err = -ENOMEM;
187 goto out;
188 }
189 break;
190
191 case PR_SCHED_CORE_SHARE_TO:
192 cookie = sched_core_clone_cookie(current);
193 break;
194
195 case PR_SCHED_CORE_SHARE_FROM:
196 if (type != PIDTYPE_PID) {
197 err = -EINVAL;
198 goto out;
199 }
200 cookie = sched_core_clone_cookie(task);
201 __sched_core_set(current, cookie);
202 goto out;
203
204 default:
205 err = -EINVAL;
206 goto out;
207 };
208
209 if (type == PIDTYPE_PID) {
210 __sched_core_set(task, cookie);
211 goto out;
212 }
213
214 read_lock(&tasklist_lock);
215 grp = task_pid_type(task, type);
216
217 do_each_pid_thread(grp, type, p) {
218 if (!ptrace_may_access(p, PTRACE_MODE_READ_REALCREDS)) {
219 err = -EPERM;
220 goto out_tasklist;
221 }
222 } while_each_pid_thread(grp, type, p);
223
224 do_each_pid_thread(grp, type, p) {
225 __sched_core_set(p, cookie);
226 } while_each_pid_thread(grp, type, p);
227out_tasklist:
228 read_unlock(&tasklist_lock);
229
230out:
231 sched_core_put_cookie(cookie);
232 put_task_struct(task);
233 return err;
234}
235
236#ifdef CONFIG_SCHEDSTATS
237
238
239void __sched_core_account_forceidle(struct rq *rq)
240{
241 const struct cpumask *smt_mask = cpu_smt_mask(cpu_of(rq));
242 u64 delta, now = rq_clock(rq->core);
243 struct rq *rq_i;
244 struct task_struct *p;
245 int i;
246
247 lockdep_assert_rq_held(rq);
248
249 WARN_ON_ONCE(!rq->core->core_forceidle_count);
250
251 if (rq->core->core_forceidle_start == 0)
252 return;
253
254 delta = now - rq->core->core_forceidle_start;
255 if (unlikely((s64)delta <= 0))
256 return;
257
258 rq->core->core_forceidle_start = now;
259
260 if (WARN_ON_ONCE(!rq->core->core_forceidle_occupation)) {
261
262 } else if (rq->core->core_forceidle_count > 1 ||
263 rq->core->core_forceidle_occupation > 1) {
264
265
266
267
268
269 delta *= rq->core->core_forceidle_count;
270 delta = div_u64(delta, rq->core->core_forceidle_occupation);
271 }
272
273 for_each_cpu(i, smt_mask) {
274 rq_i = cpu_rq(i);
275 p = rq_i->core_pick ?: rq_i->curr;
276
277 if (p == rq_i->idle)
278 continue;
279
280 __schedstat_add(p->stats.core_forceidle_sum, delta);
281 }
282}
283
284void __sched_core_tick(struct rq *rq)
285{
286 if (!rq->core->core_forceidle_count)
287 return;
288
289 if (rq != rq->core)
290 update_rq_clock(rq->core);
291
292 __sched_core_account_forceidle(rq);
293}
294
295#endif
296