1
2#ifndef _LINUX_CGROUP_H
3#define _LINUX_CGROUP_H
4
5
6
7
8
9
10
11
12#include <linux/sched.h>
13#include <linux/cpumask.h>
14#include <linux/nodemask.h>
15#include <linux/rculist.h>
16#include <linux/cgroupstats.h>
17#include <linux/fs.h>
18#include <linux/seq_file.h>
19#include <linux/kernfs.h>
20#include <linux/jump_label.h>
21#include <linux/types.h>
22#include <linux/ns_common.h>
23#include <linux/nsproxy.h>
24#include <linux/user_namespace.h>
25#include <linux/refcount.h>
26#include <linux/kernel_stat.h>
27
28#include <linux/cgroup-defs.h>
29
30struct kernel_clone_args;
31
32#ifdef CONFIG_CGROUPS
33
34
35
36
37
38
39#define CGROUP_WEIGHT_MIN 1
40#define CGROUP_WEIGHT_DFL 100
41#define CGROUP_WEIGHT_MAX 10000
42
43
44#define CSS_TASK_ITER_PROCS (1U << 0)
45
46#define CSS_TASK_ITER_THREADED (1U << 1)
47
48
49#define CSS_TASK_ITER_SKIPPED (1U << 16)
50
51
52struct css_task_iter {
53 struct cgroup_subsys *ss;
54 unsigned int flags;
55
56 struct list_head *cset_pos;
57 struct list_head *cset_head;
58
59 struct list_head *tcset_pos;
60 struct list_head *tcset_head;
61
62 struct list_head *task_pos;
63
64 struct list_head *cur_tasks_head;
65 struct css_set *cur_cset;
66 struct css_set *cur_dcset;
67 struct task_struct *cur_task;
68 struct list_head iters_node;
69};
70
71extern struct cgroup_root cgrp_dfl_root;
72extern struct css_set init_css_set;
73
74#define SUBSYS(_x) extern struct cgroup_subsys _x ## _cgrp_subsys;
75#include <linux/cgroup_subsys.h>
76#undef SUBSYS
77
78#define SUBSYS(_x) \
79 extern struct static_key_true _x ## _cgrp_subsys_enabled_key; \
80 extern struct static_key_true _x ## _cgrp_subsys_on_dfl_key;
81#include <linux/cgroup_subsys.h>
82#undef SUBSYS
83
84
85
86
87
88#define cgroup_subsys_enabled(ss) \
89 static_branch_likely(&ss ## _enabled_key)
90
91
92
93
94
95#define cgroup_subsys_on_dfl(ss) \
96 static_branch_likely(&ss ## _on_dfl_key)
97
98bool css_has_online_children(struct cgroup_subsys_state *css);
99struct cgroup_subsys_state *css_from_id(int id, struct cgroup_subsys *ss);
100struct cgroup_subsys_state *cgroup_e_css(struct cgroup *cgroup,
101 struct cgroup_subsys *ss);
102struct cgroup_subsys_state *cgroup_get_e_css(struct cgroup *cgroup,
103 struct cgroup_subsys *ss);
104struct cgroup_subsys_state *css_tryget_online_from_dir(struct dentry *dentry,
105 struct cgroup_subsys *ss);
106
107struct cgroup *cgroup_get_from_path(const char *path);
108struct cgroup *cgroup_get_from_fd(int fd);
109
110int cgroup_attach_task_all(struct task_struct *from, struct task_struct *);
111int cgroup_transfer_tasks(struct cgroup *to, struct cgroup *from);
112
113int cgroup_add_dfl_cftypes(struct cgroup_subsys *ss, struct cftype *cfts);
114int cgroup_add_legacy_cftypes(struct cgroup_subsys *ss, struct cftype *cfts);
115int cgroup_rm_cftypes(struct cftype *cfts);
116void cgroup_file_notify(struct cgroup_file *cfile);
117
118int task_cgroup_path(struct task_struct *task, char *buf, size_t buflen);
119int cgroupstats_build(struct cgroupstats *stats, struct dentry *dentry);
120int proc_cgroup_show(struct seq_file *m, struct pid_namespace *ns,
121 struct pid *pid, struct task_struct *tsk);
122
123void cgroup_fork(struct task_struct *p);
124extern int cgroup_can_fork(struct task_struct *p,
125 struct kernel_clone_args *kargs);
126extern void cgroup_cancel_fork(struct task_struct *p,
127 struct kernel_clone_args *kargs);
128extern void cgroup_post_fork(struct task_struct *p,
129 struct kernel_clone_args *kargs);
130void cgroup_exit(struct task_struct *p);
131void cgroup_release(struct task_struct *p);
132void cgroup_free(struct task_struct *p);
133
134int cgroup_init_early(void);
135int cgroup_init(void);
136
137int cgroup_parse_float(const char *input, unsigned dec_shift, s64 *v);
138
139
140
141
142
143struct cgroup_subsys_state *css_next_child(struct cgroup_subsys_state *pos,
144 struct cgroup_subsys_state *parent);
145struct cgroup_subsys_state *css_next_descendant_pre(struct cgroup_subsys_state *pos,
146 struct cgroup_subsys_state *css);
147struct cgroup_subsys_state *css_rightmost_descendant(struct cgroup_subsys_state *pos);
148struct cgroup_subsys_state *css_next_descendant_post(struct cgroup_subsys_state *pos,
149 struct cgroup_subsys_state *css);
150
151struct task_struct *cgroup_taskset_first(struct cgroup_taskset *tset,
152 struct cgroup_subsys_state **dst_cssp);
153struct task_struct *cgroup_taskset_next(struct cgroup_taskset *tset,
154 struct cgroup_subsys_state **dst_cssp);
155
156void css_task_iter_start(struct cgroup_subsys_state *css, unsigned int flags,
157 struct css_task_iter *it);
158struct task_struct *css_task_iter_next(struct css_task_iter *it);
159void css_task_iter_end(struct css_task_iter *it);
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179#define css_for_each_child(pos, parent) \
180 for ((pos) = css_next_child(NULL, (parent)); (pos); \
181 (pos) = css_next_child((pos), (parent)))
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239#define css_for_each_descendant_pre(pos, css) \
240 for ((pos) = css_next_descendant_pre(NULL, (css)); (pos); \
241 (pos) = css_next_descendant_pre((pos), (css)))
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262#define css_for_each_descendant_post(pos, css) \
263 for ((pos) = css_next_descendant_post(NULL, (css)); (pos); \
264 (pos) = css_next_descendant_post((pos), (css)))
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284#define cgroup_taskset_for_each(task, dst_css, tset) \
285 for ((task) = cgroup_taskset_first((tset), &(dst_css)); \
286 (task); \
287 (task) = cgroup_taskset_next((tset), &(dst_css)))
288
289
290
291
292
293
294
295
296
297
298#define cgroup_taskset_for_each_leader(leader, dst_css, tset) \
299 for ((leader) = cgroup_taskset_first((tset), &(dst_css)); \
300 (leader); \
301 (leader) = cgroup_taskset_next((tset), &(dst_css))) \
302 if ((leader) != (leader)->group_leader) \
303 ; \
304 else
305
306
307
308
309
310static inline u64 cgroup_id(const struct cgroup *cgrp)
311{
312 return cgrp->kn->id;
313}
314
315
316
317
318
319
320
321static inline void css_get(struct cgroup_subsys_state *css)
322{
323 if (!(css->flags & CSS_NO_REF))
324 percpu_ref_get(&css->refcnt);
325}
326
327
328
329
330
331
332
333
334static inline void css_get_many(struct cgroup_subsys_state *css, unsigned int n)
335{
336 if (!(css->flags & CSS_NO_REF))
337 percpu_ref_get_many(&css->refcnt, n);
338}
339
340
341
342
343
344
345
346
347
348
349
350
351static inline bool css_tryget(struct cgroup_subsys_state *css)
352{
353 if (!(css->flags & CSS_NO_REF))
354 return percpu_ref_tryget(&css->refcnt);
355 return true;
356}
357
358
359
360
361
362
363
364
365
366
367
368static inline bool css_tryget_online(struct cgroup_subsys_state *css)
369{
370 if (!(css->flags & CSS_NO_REF))
371 return percpu_ref_tryget_live(&css->refcnt);
372 return true;
373}
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390static inline bool css_is_dying(struct cgroup_subsys_state *css)
391{
392 return !(css->flags & CSS_NO_REF) && percpu_ref_is_dying(&css->refcnt);
393}
394
395
396
397
398
399
400
401static inline void css_put(struct cgroup_subsys_state *css)
402{
403 if (!(css->flags & CSS_NO_REF))
404 percpu_ref_put(&css->refcnt);
405}
406
407
408
409
410
411
412
413
414static inline void css_put_many(struct cgroup_subsys_state *css, unsigned int n)
415{
416 if (!(css->flags & CSS_NO_REF))
417 percpu_ref_put_many(&css->refcnt, n);
418}
419
420static inline void cgroup_get(struct cgroup *cgrp)
421{
422 css_get(&cgrp->self);
423}
424
425static inline bool cgroup_tryget(struct cgroup *cgrp)
426{
427 return css_tryget(&cgrp->self);
428}
429
430static inline void cgroup_put(struct cgroup *cgrp)
431{
432 css_put(&cgrp->self);
433}
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448#ifdef CONFIG_PROVE_RCU
449extern struct mutex cgroup_mutex;
450extern spinlock_t css_set_lock;
451#define task_css_set_check(task, __c) \
452 rcu_dereference_check((task)->cgroups, \
453 lockdep_is_held(&cgroup_mutex) || \
454 lockdep_is_held(&css_set_lock) || \
455 ((task)->flags & PF_EXITING) || (__c))
456#else
457#define task_css_set_check(task, __c) \
458 rcu_dereference((task)->cgroups)
459#endif
460
461
462
463
464
465
466
467
468
469
470#define task_css_check(task, subsys_id, __c) \
471 task_css_set_check((task), (__c))->subsys[(subsys_id)]
472
473
474
475
476
477
478
479static inline struct css_set *task_css_set(struct task_struct *task)
480{
481 return task_css_set_check(task, false);
482}
483
484
485
486
487
488
489
490
491static inline struct cgroup_subsys_state *task_css(struct task_struct *task,
492 int subsys_id)
493{
494 return task_css_check(task, subsys_id, false);
495}
496
497
498
499
500
501
502
503
504
505
506static inline struct cgroup_subsys_state *
507task_get_css(struct task_struct *task, int subsys_id)
508{
509 struct cgroup_subsys_state *css;
510
511 rcu_read_lock();
512 while (true) {
513 css = task_css(task, subsys_id);
514
515
516
517
518
519
520 if (likely(css_tryget(css)))
521 break;
522 cpu_relax();
523 }
524 rcu_read_unlock();
525 return css;
526}
527
528
529
530
531
532
533
534
535
536static inline bool task_css_is_root(struct task_struct *task, int subsys_id)
537{
538 return task_css_check(task, subsys_id, true) ==
539 init_css_set.subsys[subsys_id];
540}
541
542static inline struct cgroup *task_cgroup(struct task_struct *task,
543 int subsys_id)
544{
545 return task_css(task, subsys_id)->cgroup;
546}
547
548static inline struct cgroup *task_dfl_cgroup(struct task_struct *task)
549{
550 return task_css_set(task)->dfl_cgrp;
551}
552
553static inline struct cgroup *cgroup_parent(struct cgroup *cgrp)
554{
555 struct cgroup_subsys_state *parent_css = cgrp->self.parent;
556
557 if (parent_css)
558 return container_of(parent_css, struct cgroup, self);
559 return NULL;
560}
561
562
563
564
565
566
567
568
569
570
571static inline bool cgroup_is_descendant(struct cgroup *cgrp,
572 struct cgroup *ancestor)
573{
574 if (cgrp->root != ancestor->root || cgrp->level < ancestor->level)
575 return false;
576 return cgrp->ancestor_ids[ancestor->level] == cgroup_id(ancestor);
577}
578
579
580
581
582
583
584
585
586
587
588
589
590static inline struct cgroup *cgroup_ancestor(struct cgroup *cgrp,
591 int ancestor_level)
592{
593 if (cgrp->level < ancestor_level)
594 return NULL;
595 while (cgrp && cgrp->level > ancestor_level)
596 cgrp = cgroup_parent(cgrp);
597 return cgrp;
598}
599
600
601
602
603
604
605
606
607
608
609static inline bool task_under_cgroup_hierarchy(struct task_struct *task,
610 struct cgroup *ancestor)
611{
612 struct css_set *cset = task_css_set(task);
613
614 return cgroup_is_descendant(cset->dfl_cgrp, ancestor);
615}
616
617
618static inline bool cgroup_is_populated(struct cgroup *cgrp)
619{
620 return cgrp->nr_populated_csets + cgrp->nr_populated_domain_children +
621 cgrp->nr_populated_threaded_children;
622}
623
624
625static inline ino_t cgroup_ino(struct cgroup *cgrp)
626{
627 return kernfs_ino(cgrp->kn);
628}
629
630
631static inline struct cftype *of_cft(struct kernfs_open_file *of)
632{
633 return of->kn->priv;
634}
635
636struct cgroup_subsys_state *of_css(struct kernfs_open_file *of);
637
638
639static inline struct cftype *seq_cft(struct seq_file *seq)
640{
641 return of_cft(seq->private);
642}
643
644static inline struct cgroup_subsys_state *seq_css(struct seq_file *seq)
645{
646 return of_css(seq->private);
647}
648
649
650
651
652
653
654static inline int cgroup_name(struct cgroup *cgrp, char *buf, size_t buflen)
655{
656 return kernfs_name(cgrp->kn, buf, buflen);
657}
658
659static inline int cgroup_path(struct cgroup *cgrp, char *buf, size_t buflen)
660{
661 return kernfs_path(cgrp->kn, buf, buflen);
662}
663
664static inline void pr_cont_cgroup_name(struct cgroup *cgrp)
665{
666 pr_cont_kernfs_name(cgrp->kn);
667}
668
669static inline void pr_cont_cgroup_path(struct cgroup *cgrp)
670{
671 pr_cont_kernfs_path(cgrp->kn);
672}
673
674static inline struct psi_group *cgroup_psi(struct cgroup *cgrp)
675{
676 return &cgrp->psi;
677}
678
679bool cgroup_psi_enabled(void);
680
681static inline void cgroup_init_kthreadd(void)
682{
683
684
685
686
687
688 current->no_cgroup_migration = 1;
689}
690
691static inline void cgroup_kthread_ready(void)
692{
693
694
695
696
697 current->no_cgroup_migration = 0;
698}
699
700void cgroup_path_from_kernfs_id(u64 id, char *buf, size_t buflen);
701struct cgroup *cgroup_get_from_id(u64 id);
702#else
703
704struct cgroup_subsys_state;
705struct cgroup;
706
707static inline u64 cgroup_id(const struct cgroup *cgrp) { return 1; }
708static inline void css_get(struct cgroup_subsys_state *css) {}
709static inline void css_put(struct cgroup_subsys_state *css) {}
710static inline int cgroup_attach_task_all(struct task_struct *from,
711 struct task_struct *t) { return 0; }
712static inline int cgroupstats_build(struct cgroupstats *stats,
713 struct dentry *dentry) { return -EINVAL; }
714
715static inline void cgroup_fork(struct task_struct *p) {}
716static inline int cgroup_can_fork(struct task_struct *p,
717 struct kernel_clone_args *kargs) { return 0; }
718static inline void cgroup_cancel_fork(struct task_struct *p,
719 struct kernel_clone_args *kargs) {}
720static inline void cgroup_post_fork(struct task_struct *p,
721 struct kernel_clone_args *kargs) {}
722static inline void cgroup_exit(struct task_struct *p) {}
723static inline void cgroup_release(struct task_struct *p) {}
724static inline void cgroup_free(struct task_struct *p) {}
725
726static inline int cgroup_init_early(void) { return 0; }
727static inline int cgroup_init(void) { return 0; }
728static inline void cgroup_init_kthreadd(void) {}
729static inline void cgroup_kthread_ready(void) {}
730
731static inline struct cgroup *cgroup_parent(struct cgroup *cgrp)
732{
733 return NULL;
734}
735
736static inline struct psi_group *cgroup_psi(struct cgroup *cgrp)
737{
738 return NULL;
739}
740
741static inline bool cgroup_psi_enabled(void)
742{
743 return false;
744}
745
746static inline bool task_under_cgroup_hierarchy(struct task_struct *task,
747 struct cgroup *ancestor)
748{
749 return true;
750}
751
752static inline void cgroup_path_from_kernfs_id(u64 id, char *buf, size_t buflen)
753{}
754
755static inline struct cgroup *cgroup_get_from_id(u64 id)
756{
757 return NULL;
758}
759#endif
760
761#ifdef CONFIG_CGROUPS
762
763
764
765void cgroup_rstat_updated(struct cgroup *cgrp, int cpu);
766void cgroup_rstat_flush(struct cgroup *cgrp);
767void cgroup_rstat_flush_irqsafe(struct cgroup *cgrp);
768void cgroup_rstat_flush_hold(struct cgroup *cgrp);
769void cgroup_rstat_flush_release(void);
770
771
772
773
774#ifdef CONFIG_CGROUP_CPUACCT
775void cpuacct_charge(struct task_struct *tsk, u64 cputime);
776void cpuacct_account_field(struct task_struct *tsk, int index, u64 val);
777#else
778static inline void cpuacct_charge(struct task_struct *tsk, u64 cputime) {}
779static inline void cpuacct_account_field(struct task_struct *tsk, int index,
780 u64 val) {}
781#endif
782
783void __cgroup_account_cputime(struct cgroup *cgrp, u64 delta_exec);
784void __cgroup_account_cputime_field(struct cgroup *cgrp,
785 enum cpu_usage_stat index, u64 delta_exec);
786
787static inline void cgroup_account_cputime(struct task_struct *task,
788 u64 delta_exec)
789{
790 struct cgroup *cgrp;
791
792 cpuacct_charge(task, delta_exec);
793
794 rcu_read_lock();
795 cgrp = task_dfl_cgroup(task);
796 if (cgroup_parent(cgrp))
797 __cgroup_account_cputime(cgrp, delta_exec);
798 rcu_read_unlock();
799}
800
801static inline void cgroup_account_cputime_field(struct task_struct *task,
802 enum cpu_usage_stat index,
803 u64 delta_exec)
804{
805 struct cgroup *cgrp;
806
807 cpuacct_account_field(task, index, delta_exec);
808
809 rcu_read_lock();
810 cgrp = task_dfl_cgroup(task);
811 if (cgroup_parent(cgrp))
812 __cgroup_account_cputime_field(cgrp, index, delta_exec);
813 rcu_read_unlock();
814}
815
816#else
817
818static inline void cgroup_account_cputime(struct task_struct *task,
819 u64 delta_exec) {}
820static inline void cgroup_account_cputime_field(struct task_struct *task,
821 enum cpu_usage_stat index,
822 u64 delta_exec) {}
823
824#endif
825
826
827
828
829
830#ifdef CONFIG_SOCK_CGROUP_DATA
831
832void cgroup_sk_alloc(struct sock_cgroup_data *skcd);
833void cgroup_sk_clone(struct sock_cgroup_data *skcd);
834void cgroup_sk_free(struct sock_cgroup_data *skcd);
835
836static inline struct cgroup *sock_cgroup_ptr(struct sock_cgroup_data *skcd)
837{
838 return skcd->cgroup;
839}
840
841#else
842
843static inline void cgroup_sk_alloc(struct sock_cgroup_data *skcd) {}
844static inline void cgroup_sk_clone(struct sock_cgroup_data *skcd) {}
845static inline void cgroup_sk_free(struct sock_cgroup_data *skcd) {}
846
847#endif
848
849struct cgroup_namespace {
850 struct ns_common ns;
851 struct user_namespace *user_ns;
852 struct ucounts *ucounts;
853 struct css_set *root_cset;
854};
855
856extern struct cgroup_namespace init_cgroup_ns;
857
858#ifdef CONFIG_CGROUPS
859
860void free_cgroup_ns(struct cgroup_namespace *ns);
861
862struct cgroup_namespace *copy_cgroup_ns(unsigned long flags,
863 struct user_namespace *user_ns,
864 struct cgroup_namespace *old_ns);
865
866int cgroup_path_ns(struct cgroup *cgrp, char *buf, size_t buflen,
867 struct cgroup_namespace *ns);
868
869#else
870
871static inline void free_cgroup_ns(struct cgroup_namespace *ns) { }
872static inline struct cgroup_namespace *
873copy_cgroup_ns(unsigned long flags, struct user_namespace *user_ns,
874 struct cgroup_namespace *old_ns)
875{
876 return old_ns;
877}
878
879#endif
880
881static inline void get_cgroup_ns(struct cgroup_namespace *ns)
882{
883 if (ns)
884 refcount_inc(&ns->ns.count);
885}
886
887static inline void put_cgroup_ns(struct cgroup_namespace *ns)
888{
889 if (ns && refcount_dec_and_test(&ns->ns.count))
890 free_cgroup_ns(ns);
891}
892
893#ifdef CONFIG_CGROUPS
894
895void cgroup_enter_frozen(void);
896void cgroup_leave_frozen(bool always_leave);
897void cgroup_update_frozen(struct cgroup *cgrp);
898void cgroup_freeze(struct cgroup *cgrp, bool freeze);
899void cgroup_freezer_migrate_task(struct task_struct *task, struct cgroup *src,
900 struct cgroup *dst);
901
902static inline bool cgroup_task_frozen(struct task_struct *task)
903{
904 return task->frozen;
905}
906
907#else
908
909static inline void cgroup_enter_frozen(void) { }
910static inline void cgroup_leave_frozen(bool always_leave) { }
911static inline bool cgroup_task_frozen(struct task_struct *task)
912{
913 return false;
914}
915
916#endif
917
918#ifdef CONFIG_CGROUP_BPF
919static inline void cgroup_bpf_get(struct cgroup *cgrp)
920{
921 percpu_ref_get(&cgrp->bpf.refcnt);
922}
923
924static inline void cgroup_bpf_put(struct cgroup *cgrp)
925{
926 percpu_ref_put(&cgrp->bpf.refcnt);
927}
928
929#else
930
931static inline void cgroup_bpf_get(struct cgroup *cgrp) {}
932static inline void cgroup_bpf_put(struct cgroup *cgrp) {}
933
934#endif
935
936#endif
937