dwww Home | Manual pages | Find package

cgroups(7)             Miscellaneous Information Manual             cgroups(7)

NAME
       cgroups - Linux control groups

DESCRIPTION
       Control groups, usually referred to as cgroups, are a Linux kernel fea-
       ture which allow processes to be  organized  into  hierarchical  groups
       whose usage of various types of resources can then be limited and moni-
       tored.  The kernel's cgroup interface is  provided  through  a  pseudo-
       filesystem called cgroupfs.  Grouping is implemented in the core cgroup
       kernel code, while resource tracking and limits are  implemented  in  a
       set of per-resource-type subsystems (memory, CPU, and so on).

   Terminology
       A cgroup is a collection of processes that are bound to a set of limits
       or parameters defined via the cgroup filesystem.

       A subsystem is a kernel component that modifies  the  behavior  of  the
       processes  in a cgroup.  Various subsystems have been implemented, mak-
       ing it possible to do things such as limiting the amount  of  CPU  time
       and memory available to a cgroup, accounting for the CPU time used by a
       cgroup, and freezing and resuming  execution  of  the  processes  in  a
       cgroup.   Subsystems  are  sometimes also known as resource controllers
       (or simply, controllers).

       The cgroups for a controller are arranged in a hierarchy.  This hierar-
       chy  is  defined  by  creating,  removing,  and renaming subdirectories
       within the cgroup filesystem.  At each  level  of  the  hierarchy,  at-
       tributes  (e.g.,  limits) can be defined.  The limits, control, and ac-
       counting provided by cgroups generally have effect throughout the  sub-
       hierarchy  underneath  the  cgroup  where  the  attributes are defined.
       Thus, for example, the limits placed on a cgroup at a higher  level  in
       the hierarchy cannot be exceeded by descendant cgroups.

   Cgroups version 1 and version 2
       The  initial release of the cgroups implementation was in Linux 2.6.24.
       Over time, various cgroup controllers have been added to allow the man-
       agement  of  various  types  of resources.  However, the development of
       these controllers was largely uncoordinated, with the result that  many
       inconsistencies  arose between controllers and management of the cgroup
       hierarchies became rather complex.  A longer description of these prob-
       lems   can  be  found  in  the  kernel  source  file  Documentation/ad-
       min-guide/cgroup-v2.rst (or Documentation/cgroup-v2.txt in  Linux  4.17
       and earlier).

       Because  of  the  problems  with  the  initial  cgroups  implementation
       (cgroups version 1), starting in Linux 3.10, work began on a  new,  or-
       thogonal implementation to remedy these problems.  Initially marked ex-
       perimental, and hidden behind the -o __DEVEL__sane_behavior  mount  op-
       tion,  the new version (cgroups version 2) was eventually made official
       with the release of Linux 4.5.  Differences between  the  two  versions
       are  described  in  the  text  below.   The  file cgroup.sane_behavior,
       present in cgroups v1, is a relic of this mount option.  The  file  al-
       ways reports "0" and is only retained for backward compatibility.

       Although  cgroups  v2  is intended as a replacement for cgroups v1, the
       older system continues to exist (and for compatibility reasons  is  un-
       likely  to be removed).  Currently, cgroups v2 implements only a subset
       of the controllers available in cgroups v1.  The two systems are imple-
       mented so that both v1 controllers and v2 controllers can be mounted on
       the same system.  Thus, for example, it is possible to use  those  con-
       trollers that are supported under version 2, while also using version 1
       controllers where version 2 does not  yet  support  those  controllers.
       The  only restriction here is that a controller can't be simultaneously
       employed in both a cgroups v1 hierarchy and in the cgroups  v2  hierar-
       chy.

CGROUPS VERSION 1
       Under  cgroups  v1,  each  controller may be mounted against a separate
       cgroup filesystem that provides its own  hierarchical  organization  of
       the  processes  on the system.  It is also possible to comount multiple
       (or even all) cgroups v1 controllers against the same  cgroup  filesys-
       tem,  meaning that the comounted controllers manage the same hierarchi-
       cal organization of processes.

       For each mounted hierarchy, the  directory  tree  mirrors  the  control
       group  hierarchy.   Each  control  group is represented by a directory,
       with each of its child control cgroups represented as  a  child  direc-
       tory.   For  instance,  /user/joe/1.session  represents  control  group
       1.session, which is a child of cgroup joe, which is a child  of  /user.
       Under  each  cgroup  directory  is  a set of files which can be read or
       written to, reflecting resource limits and a few general cgroup proper-
       ties.

   Tasks (threads) versus processes
       In  cgroups v1, a distinction is drawn between processes and tasks.  In
       this view, a process can  consist  of  multiple  tasks  (more  commonly
       called  threads,  from a user-space perspective, and called such in the
       remainder of this man page).  In cgroups v1, it is possible to indepen-
       dently manipulate the cgroup memberships of the threads in a process.

       The cgroups v1 ability to split threads across different cgroups caused
       problems in some cases.  For example, it made no sense for  the  memory
       controller,  since  all  of the threads of a process share a single ad-
       dress space.  Because of these problems, the ability  to  independently
       manipulate  the  cgroup memberships of the threads in a process was re-
       moved in the initial cgroups v2 implementation,  and  subsequently  re-
       stored  in a more limited form (see the discussion of "thread mode" be-
       low).

   Mounting v1 controllers
       The use of cgroups requires a kernel built with the  CONFIG_CGROUP  op-
       tion.   In  addition, each of the v1 controllers has an associated con-
       figuration option that must be set in order to employ that controller.

       In order to use a v1 controller, it must be mounted  against  a  cgroup
       filesystem.   The  usual  place  for  such  mounts  is under a tmpfs(5)
       filesystem mounted at /sys/fs/cgroup.  Thus, one might  mount  the  cpu
       controller as follows:

           mount -t cgroup -o cpu none /sys/fs/cgroup/cpu

       It is possible to comount multiple controllers against the same hierar-
       chy.  For example, here the cpu and cpuacct controllers  are  comounted
       against a single hierarchy:

           mount -t cgroup -o cpu,cpuacct none /sys/fs/cgroup/cpu,cpuacct

       Comounting  controllers  has  the  effect that a process is in the same
       cgroup for all of the comounted controllers.  Separately mounting  con-
       trollers  allows  a  process  to  be in cgroup /foo1 for one controller
       while being in /foo2/foo3 for another.

       It is possible to comount all v1 controllers against the  same  hierar-
       chy:

           mount -t cgroup -o all cgroup /sys/fs/cgroup

       (One  can  achieve  the same result by omitting -o all, since it is the
       default if no controllers are explicitly specified.)

       It is not possible to mount the same controller against multiple cgroup
       hierarchies.  For example, it is not possible to mount both the cpu and
       cpuacct controllers against one hierarchy, and to mount  the  cpu  con-
       troller alone against another hierarchy.  It is possible to create mul-
       tiple mount with exactly the same set of comounted  controllers.   How-
       ever,  in this case all that results is multiple mount points providing
       a view of the same hierarchy.

       Note that on many systems, the v1 controllers are automatically mounted
       under  /sys/fs/cgroup;  in particular, systemd(1) automatically creates
       such mounts.

   Unmounting v1 controllers
       A mounted cgroup filesystem can be unmounted using the  umount(8)  com-
       mand, as in the following example:

           umount /sys/fs/cgroup/pids

       But note well: a cgroup filesystem is unmounted only if it is not busy,
       that is, it has no child cgroups.  If this is not the  case,  then  the
       only  effect of the umount(8) is to make the mount invisible.  Thus, to
       ensure that the mount is really removed,  one  must  first  remove  all
       child  cgroups,  which  in  turn can be done only after all member pro-
       cesses have been moved from those cgroups to the root cgroup.

   Cgroups version 1 controllers
       Each of the cgroups version 1 controllers is governed by a kernel  con-
       figuration  option  (listed  below).  Additionally, the availability of
       the cgroups feature is governed by the CONFIG_CGROUPS kernel configura-
       tion option.

       cpu (since Linux 2.6.24; CONFIG_CGROUP_SCHED)
              Cgroups  can be guaranteed a minimum number of "CPU shares" when
              a system is busy.  This does not limit a cgroup's CPU  usage  if
              the  CPUs are not busy.  For further information, see Documenta-
              tion/scheduler/sched-design-CFS.rst   (or   Documentation/sched-
              uler/sched-design-CFS.txt in Linux 5.2 and earlier).

              In Linux 3.2, this controller was extended to provide CPU "band-
              width"  control.   If  the  kernel  is  configured   with   CON-
              FIG_CFS_BANDWIDTH,  then  within each scheduling period (defined
              via a file in the cgroup directory), it is possible to define an
              upper  limit  on  the  CPU  time allocated to the processes in a
              cgroup.  This upper limit applies even if there is no other com-
              petition  for  the CPU.  Further information can be found in the
              kernel  source  file  Documentation/scheduler/sched-bwc.rst  (or
              Documentation/scheduler/sched-bwc.txt in Linux 5.2 and earlier).

       cpuacct (since Linux 2.6.24; CONFIG_CGROUP_CPUACCT)
              This provides accounting for CPU usage by groups of processes.

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/cpuacct.rst    (or    Documenta-
              tion/cgroup-v1/cpuacct.txt in Linux 5.2 and earlier).

       cpuset (since Linux 2.6.24; CONFIG_CPUSETS)
              This  cgroup  can be used to bind the processes in a cgroup to a
              specified set of CPUs and NUMA nodes.

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/cpusets.rst    (or    Documenta-
              tion/cgroup-v1/cpusets.txt in Linux 5.2 and earlier).

       memory (since Linux 2.6.25; CONFIG_MEMCG)
              The memory controller supports reporting and limiting of process
              memory, kernel memory, and swap used by cgroups.

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/memory.rst    (or     Documenta-
              tion/cgroup-v1/memory.txt in Linux 5.2 and earlier).

       devices (since Linux 2.6.26; CONFIG_CGROUP_DEVICE)
              This supports controlling which processes may create (mknod) de-
              vices as well as open them for reading or writing.  The policies
              may  be  specified  as allow-lists and deny-lists.  Hierarchy is
              enforced, so new rules must not violate existing rules  for  the
              target or ancestor cgroups.

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/devices.rst    (or    Documenta-
              tion/cgroup-v1/devices.txt in Linux 5.2 and earlier).

       freezer (since Linux 2.6.28; CONFIG_CGROUP_FREEZER)
              The  freezer  cgroup  can  suspend and restore (resume) all pro-
              cesses in a cgroup.  Freezing a cgroup /A also causes its  chil-
              dren, for example, processes in /A/B, to be frozen.

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/freezer-subsystem.rst (or  Docu-
              mentation/cgroup-v1/freezer-subsystem.txt  in Linux 5.2 and ear-
              lier).

       net_cls (since Linux 2.6.29; CONFIG_CGROUP_NET_CLASSID)
              This places a classid, specified  for  the  cgroup,  on  network
              packets created by a cgroup.  These classids can then be used in
              firewall rules, as well as used to shape  traffic  using  tc(8).
              This  applies only to packets leaving the cgroup, not to traffic
              arriving at the cgroup.

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/net_cls.rst    (or    Documenta-
              tion/cgroup-v1/net_cls.txt in Linux 5.2 and earlier).

       blkio (since Linux 2.6.33; CONFIG_BLK_CGROUP)
              The blkio cgroup controls and limits access to  specified  block
              devices by applying IO control in the form of throttling and up-
              per limits against leaf nodes  and  intermediate  nodes  in  the
              storage hierarchy.

              Two  policies are available.  The first is a proportional-weight
              time-based division of disk implemented with CFQ.   This  is  in
              effect  for  leaf  nodes  using CFQ.  The second is a throttling
              policy which specifies upper I/O rate limits on a device.

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/blkio-controller.rst  (or  Docu-
              mentation/cgroup-v1/blkio-controller.txt in Linux 5.2  and  ear-
              lier).

       perf_event (since Linux 2.6.39; CONFIG_CGROUP_PERF)
              This  controller  allows perf monitoring of the set of processes
              grouped in a cgroup.

              Further information can be found in the kernel source files

       net_prio (since Linux 3.3; CONFIG_CGROUP_NET_PRIO)
              This allows priorities to be specified, per  network  interface,
              for cgroups.

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/net_prio.rst   (or    Documenta-
              tion/cgroup-v1/net_prio.txt in Linux 5.2 and earlier).

       hugetlb (since Linux 3.5; CONFIG_CGROUP_HUGETLB)
              This supports limiting the use of huge pages by cgroups.

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/hugetlb.rst    (or    Documenta-
              tion/cgroup-v1/hugetlb.txt in Linux 5.2 and earlier).

       pids (since Linux 4.3; CONFIG_CGROUP_PIDS)
              This  controller permits limiting the number of process that may
              be created in a cgroup (and its descendants).

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/pids.rst      (or     Documenta-
              tion/cgroup-v1/pids.txt in Linux 5.2 and earlier).

       rdma (since Linux 4.11; CONFIG_CGROUP_RDMA)
              The RDMA controller permits limiting the use of RDMA/IB-specific
              resources per cgroup.

              Further information can be found in the kernel source file Docu-
              mentation/admin-guide/cgroup-v1/rdma.rst     (or      Documenta-
              tion/cgroup-v1/rdma.txt in Linux 5.2 and earlier).

   Creating cgroups and moving processes
       A cgroup filesystem initially contains a single root cgroup, '/', which
       all processes belong to.  A new cgroup is created by creating a  direc-
       tory in the cgroup filesystem:

           mkdir /sys/fs/cgroup/cpu/cg1

       This creates a new empty cgroup.

       A  process  may  be  moved  to  this cgroup by writing its PID into the
       cgroup's cgroup.procs file:

           echo $$ > /sys/fs/cgroup/cpu/cg1/cgroup.procs

       Only one PID at a time should be written to this file.

       Writing the value 0 to a cgroup.procs file causes the  writing  process
       to be moved to the corresponding cgroup.

       When  writing  a  PID into the cgroup.procs, all threads in the process
       are moved into the new cgroup at once.

       Within a hierarchy, a process can be a member of  exactly  one  cgroup.
       Writing a process's PID to a cgroup.procs file automatically removes it
       from the cgroup of which it was previously a member.

       The cgroup.procs file can be read to obtain a  list  of  the  processes
       that are members of a cgroup.  The returned list of PIDs is not guaran-
       teed to be in order.  Nor is it guaranteed to be  free  of  duplicates.
       (For example, a PID may be recycled while reading from the list.)

       In  cgroups  v1, an individual thread can be moved to another cgroup by
       writing its thread ID (i.e., the kernel thread ID returned by  clone(2)
       and  gettid(2)) to the tasks file in a cgroup directory.  This file can
       be read to discover the set of threads that are members of the cgroup.

   Removing cgroups
       To remove a cgroup, it must first have no child cgroups and contain  no
       (nonzombie) processes.  So long as that is the case, one can simply re-
       move the corresponding directory pathname.  Note that files in a cgroup
       directory cannot and need not be removed.

   Cgroups v1 release notification
       Two  files can be used to determine whether the kernel provides notifi-
       cations when a cgroup becomes empty.  A  cgroup  is  considered  to  be
       empty when it contains no child cgroups and no member processes.

       A  special  file  in  the  root directory of each cgroup hierarchy, re-
       lease_agent, can be used to register the pathname of a program that may
       be  invoked when a cgroup in the hierarchy becomes empty.  The pathname
       of the newly empty cgroup (relative to the cgroup mount point) is  pro-
       vided  as the sole command-line argument when the release_agent program
       is invoked.  The release_agent program might remove the  cgroup  direc-
       tory, or perhaps repopulate it with a process.

       The  default  value of the release_agent file is empty, meaning that no
       release agent is invoked.

       The content of the release_agent file can also be specified via a mount
       option when the cgroup filesystem is mounted:

           mount -o release_agent=pathname ...

       Whether  or  not the release_agent program is invoked when a particular
       cgroup becomes empty is determined by the value  in  the  notify_on_re-
       lease  file  in  the corresponding cgroup directory.  If this file con-
       tains the value 0, then the release_agent program is not  invoked.   If
       it contains the value 1, the release_agent program is invoked.  The de-
       fault value for this file in the root cgroup is 0.  At the time when  a
       new  cgroup  is  created,  the value in this file is inherited from the
       corresponding file in the parent cgroup.

   Cgroup v1 named hierarchies
       In cgroups v1, it is possible to mount a cgroup hierarchy that  has  no
       attached controllers:

           mount -t cgroup -o none,name=somename none /some/mount/point

       Multiple  instances  of such hierarchies can be mounted; each hierarchy
       must have a unique name.  The only purpose of such  hierarchies  is  to
       track  processes.   (See the discussion of release notification below.)
       An example of this is the name=systemd cgroup hierarchy that is used by
       systemd(1) to track services and user sessions.

       Since  Linux 5.0, the cgroup_no_v1 kernel boot option (described below)
       can be used to disable  cgroup  v1  named  hierarchies,  by  specifying
       cgroup_no_v1=named.

CGROUPS VERSION 2
       In cgroups v2, all mounted controllers reside in a single unified hier-
       archy.  While (different) controllers may be simultaneously mounted un-
       der  the  v1  and  v2 hierarchies, it is not possible to mount the same
       controller simultaneously under both the v1 and the v2 hierarchies.

       The new behaviors in cgroups v2 are summarized here, and in some  cases
       elaborated in the following subsections.

       •  Cgroups  v2  provides  a  unified  hierarchy  against which all con-
          trollers are mounted.

       •  "Internal" processes are not permitted.  With the exception  of  the
          root  cgroup,  processes may reside only in leaf nodes (cgroups that
          do not themselves contain child cgroups).  The details are  somewhat
          more subtle than this, and are described below.

       •  Active  cgroups  must  be specified via the files cgroup.controllers
          and cgroup.subtree_control.

       •  The   tasks   file   has   been   removed.    In    addition,    the
          cgroup.clone_children file that is employed by the cpuset controller
          has been removed.

       •  An improved mechanism for notification of empty cgroups is  provided
          by the cgroup.events file.

       For  more changes, see the Documentation/admin-guide/cgroup-v2.rst file
       in the kernel source (or Documentation/cgroup-v2.txt in Linux 4.17  and
       earlier).

       Some of the new behaviors listed above saw subsequent modification with
       the addition in Linux 4.14 of "thread mode" (described below).

   Cgroups v2 unified hierarchy
       In cgroups v1, the ability to mount different controllers against  dif-
       ferent hierarchies was intended to allow great flexibility for applica-
       tion design.  In practice, though, the flexibility  turned  out  to  be
       less  useful than expected, and in many cases added complexity.  There-
       fore, in cgroups v2, all available controllers are  mounted  against  a
       single hierarchy.  The available controllers are automatically mounted,
       meaning that it is not necessary (or  possible)  to  specify  the  con-
       trollers when mounting the cgroup v2 filesystem using a command such as
       the following:

           mount -t cgroup2 none /mnt/cgroup2

       A cgroup v2 controller is available only if it is not currently in  use
       via  a  mount against a cgroup v1 hierarchy.  Or, to put things another
       way, it is not possible to employ the same controller against both a v1
       hierarchy and the unified v2 hierarchy.  This means that it may be nec-
       essary first to unmount a v1 controller  (as  described  above)  before
       that  controller  is available in v2.  Since systemd(1) makes heavy use
       of some v1 controllers by default, it can in some cases be  simpler  to
       boot  the  system  with  selected v1 controllers disabled.  To do this,
       specify the cgroup_no_v1=list option on the kernel boot  command  line;
       list  is a comma-separated list of the names of the controllers to dis-
       able, or the word all to disable all v1 controllers.   (This  situation
       is correctly handled by systemd(1), which falls back to operating with-
       out the specified controllers.)

       Note that on many modern systems, systemd(1) automatically  mounts  the
       cgroup2 filesystem at /sys/fs/cgroup/unified during the boot process.

   Cgroups v2 mount options
       The  following  options  (mount  -o) can be specified when mounting the
       group v2 filesystem:

       nsdelegate (since Linux 4.15)
              Treat cgroup namespaces as delegation boundaries.  For  details,
              see below.

       memory_localevents (since Linux 5.2)
              The memory.events should show statistics only for the cgroup it-
              self, and not for any descendant cgroups.  This was the behavior
              before  Linux  5.2.  Starting in Linux 5.2, the default behavior
              is  to  include  statistics  for  descendant  cgroups  in   mem-
              ory.events,  and  this mount option can be used to revert to the
              legacy behavior.  This option is system wide and can be  set  on
              mount  or  modified  through remount only from the initial mount
              namespace; it is silently ignored in noninitial namespaces.

   Cgroups v2 controllers
       The following controllers, documented in the kernel source  file  Docu-
       mentation/admin-guide/cgroup-v2.rst  (or Documentation/cgroup-v2.txt in
       Linux 4.17 and earlier), are supported in cgroups version 2:

       cpu (since Linux 4.15)
              This is the successor to the version  1  cpu  and  cpuacct  con-
              trollers.

       cpuset (since Linux 5.0)
              This is the successor of the version 1 cpuset controller.

       freezer (since Linux 5.2)
              This is the successor of the version 1 freezer controller.

       hugetlb (since Linux 5.6)
              This is the successor of the version 1 hugetlb controller.

       io (since Linux 4.5)
              This is the successor of the version 1 blkio controller.

       memory (since Linux 4.5)
              This is the successor of the version 1 memory controller.

       perf_event (since Linux 4.11)
              This is the same as the version 1 perf_event controller.

       pids (since Linux 4.5)
              This is the same as the version 1 pids controller.

       rdma (since Linux 4.11)
              This is the same as the version 1 rdma controller.

       There  is  no direct equivalent of the net_cls and net_prio controllers
       from cgroups version 1.  Instead, support has been added to iptables(8)
       to  allow  eBPF  filters that hook on cgroup v2 pathnames to make deci-
       sions about network traffic on a per-cgroup basis.

       The v2 devices controller provides no interface files; instead,  device
       control  is gated by attaching an eBPF (BPF_CGROUP_DEVICE) program to a
       v2 cgroup.

   Cgroups v2 subtree control
       Each cgroup in the v2 hierarchy contains the following two files:

       cgroup.controllers
              This read-only file exposes a list of the controllers  that  are
              available  in  this cgroup.  The contents of this file match the
              contents  of  the  cgroup.subtree_control  file  in  the  parent
              cgroup.

       cgroup.subtree_control
              This  is  a list of controllers that are active (enabled) in the
              cgroup.  The set of controllers in this file is a subset of  the
              set in the cgroup.controllers of this cgroup.  The set of active
              controllers is modified by writing strings to this file contain-
              ing  space-delimited  controller names, each preceded by '+' (to
              enable a controller) or '-' (to disable a controller), as in the
              following example:

                  echo '+pids -memory' > x/y/cgroup.subtree_control

              An  attempt  to  enable  a  controller  that  is  not present in
              cgroup.controllers leads to an ENOENT error when writing to  the
              cgroup.subtree_control file.

       Because  the  list of controllers in cgroup.subtree_control is a subset
       of those cgroup.controllers, a controller that has been disabled in one
       cgroup  in  the  hierarchy can never be re-enabled in the subtree below
       that cgroup.

       A cgroup's cgroup.subtree_control  file  determines  the  set  of  con-
       trollers  that  are  exercised in the child cgroups.  When a controller
       (e.g., pids) is present in the cgroup.subtree_control file of a  parent
       cgroup,   then  the  corresponding  controller-interface  files  (e.g.,
       pids.max) are automatically created in the children of that cgroup  and
       can be used to exert resource control in the child cgroups.

   Cgroups v2 "no internal processes" rule
       Cgroups  v2 enforces a so-called "no internal processes" rule.  Roughly
       speaking, this rule means that, with the exception of the root  cgroup,
       processes may reside only in leaf nodes (cgroups that do not themselves
       contain child cgroups).  This avoids the need to decide how  to  parti-
       tion resources between processes which are members of cgroup A and pro-
       cesses in child cgroups of A.

       For instance, if cgroup /cg1/cg2 exists, then a process may  reside  in
       /cg1/cg2, but not in /cg1.  This is to avoid an ambiguity in cgroups v1
       with respect to the delegation of resources between processes  in  /cg1
       and  its  child  cgroups.  The recommended approach in cgroups v2 is to
       create a subdirectory called leaf for any nonleaf cgroup  which  should
       contain  processes, but no child cgroups.  Thus, processes which previ-
       ously would have gone into /cg1 would now go into /cg1/leaf.  This  has
       the  advantage of making explicit the relationship between processes in
       /cg1/leaf and /cg1's other children.

       The "no internal processes" rule is in fact  more  subtle  than  stated
       above.   More precisely, the rule is that a (nonroot) cgroup can't both
       (1) have member processes, and  (2)  distribute  resources  into  child
       cgroups—that is, have a nonempty cgroup.subtree_control file.  Thus, it
       is possible for a cgroup  to  have  both  member  processes  and  child
       cgroups,  but  before  controllers  can be enabled for that cgroup, the
       member processes must be moved out of the cgroup  (e.g.,  perhaps  into
       the child cgroups).

       With  the  Linux  4.14 addition of "thread mode" (described below), the
       "no internal processes" rule has been relaxed in some cases.

   Cgroups v2 cgroup.events file
       Each nonroot cgroup in the v2  hierarchy  contains  a  read-only  file,
       cgroup.events, whose contents are key-value pairs (delimited by newline
       characters, with the key and value separated by spaces) providing state
       information about the cgroup:

           $ cat mygrp/cgroup.events
           populated 1
           frozen 0

       The following keys may appear in this file:

       populated
              The  value of this key is either 1, if this cgroup or any of its
              descendants has member processes, or otherwise 0.

       frozen (since Linux 5.2)
              The value of this key is 1 if this cgroup is  currently  frozen,
              or 0 if it is not.

       The  cgroup.events file can be monitored, in order to receive notifica-
       tion when the value of one of its keys changes.  Such monitoring can be
       done  using  inotify(7), which notifies changes as IN_MODIFY events, or
       poll(2), which notifies changes by returning the  POLLPRI  and  POLLERR
       bits in the revents field.

   Cgroup v2 release notification
       Cgroups  v2  provides a new mechanism for obtaining notification when a
       cgroup becomes empty.  The cgroups v1 release_agent  and  notify_on_re-
       lease  files  are  removed,  and  replaced  by the populated key in the
       cgroup.events file.  This key either has the value 0, meaning that  the
       cgroup  (and  its descendants) contain no (nonzombie) member processes,
       or 1, meaning that the cgroup (or one of its descendants) contains mem-
       ber processes.

       The  cgroups v2 release-notification mechanism offers the following ad-
       vantages over the cgroups v1 release_agent mechanism:

       •  It allows for cheaper notification, since a single process can moni-
          tor  multiple  cgroup.events  files  (using the techniques described
          earlier).  By contrast, the cgroups v1 mechanism  requires  the  ex-
          pense of creating a process for each notification.

       •  Notification for different cgroup subhierarchies can be delegated to
          different processes.  By contrast, the cgroups v1  mechanism  allows
          only one release agent for an entire hierarchy.

   Cgroups v2 cgroup.stat file
       Each  cgroup  in the v2 hierarchy contains a read-only cgroup.stat file
       (first introduced in Linux 4.14) that consists of lines containing key-
       value pairs.  The following keys currently appear in this file:

       nr_descendants
              This  is  the  total number of visible (i.e., living) descendant
              cgroups underneath this cgroup.

       nr_dying_descendants
              This is the total number of dying descendant cgroups  underneath
              this  cgroup.   A  cgroup  enters  the  dying  state after being
              deleted.  It remains in  that  state  for  an  undefined  period
              (which will depend on system load) while resources are freed be-
              fore the cgroup is destroyed.  Note that the  presence  of  some
              cgroups  in  the dying state is normal, and is not indicative of
              any problem.

              A process can't be made a member of a dying cgroup, and a  dying
              cgroup can't be brought back to life.

   Limiting the number of descendant cgroups
       Each cgroup in the v2 hierarchy contains the following files, which can
       be used to view and set limits on the number of descendant cgroups  un-
       der that cgroup:

       cgroup.max.depth (since Linux 4.14)
              This  file defines a limit on the depth of nesting of descendant
              cgroups.  A value of 0 in this file  means  that  no  descendant
              cgroups can be created.  An attempt to create a descendant whose
              nesting level exceeds the limit fails (mkdir(2) fails  with  the
              error EAGAIN).

              Writing the string "max" to this file means that no limit is im-
              posed.  The default value in this file is "max" .

       cgroup.max.descendants (since Linux 4.14)
              This file defines a limit  on  the  number  of  live  descendant
              cgroups  that  this  cgroup may have.  An attempt to create more
              descendants than allowed by the limit fails (mkdir(2) fails with
              the error EAGAIN).

              Writing the string "max" to this file means that no limit is im-
              posed.  The default value in this file is "max".

CGROUPS DELEGATION: DELEGATING A HIERARCHY TO A LESS PRIVILEGED USER
       In the context of cgroups, delegation means passing management of  some
       subtree  of  the  cgroup hierarchy to a nonprivileged user.  Cgroups v1
       provides support for delegation based on file permissions in the cgroup
       hierarchy  but with less strict containment rules than v2 (as noted be-
       low).  Cgroups v2 supports delegation with containment by explicit  de-
       sign.   The focus of the discussion in this section is on delegation in
       cgroups v2, with some differences for cgroups v1 noted along the way.

       Some terminology is required in order to describe delegation.  A  dele-
       gater  is  a  privileged user (i.e., root) who owns a parent cgroup.  A
       delegatee is a nonprivileged user who will be granted  the  permissions
       needed  to  manage some subhierarchy under that parent cgroup, known as
       the delegated subtree.

       To perform delegation, the  delegater  makes  certain  directories  and
       files writable by the delegatee, typically by changing the ownership of
       the objects to be the user ID of the delegatee.  Assuming that we  want
       to  delegate the hierarchy rooted at (say) /dlgt_grp and that there are
       not yet any child cgroups under that cgroup, the ownership of the  fol-
       lowing is changed to the user ID of the delegatee:

       /dlgt_grp
              Changing the ownership of the root of the subtree means that any
              new cgroups created under the subtree (and the files  they  con-
              tain) will also be owned by the delegatee.

       /dlgt_grp/cgroup.procs
              Changing the ownership of this file means that the delegatee can
              move processes into the root of the delegated subtree.

       /dlgt_grp/cgroup.subtree_control (cgroups v2 only)
              Changing the ownership of this file means that the delegatee can
              enable  controllers  (that  are present in /dlgt_grp/cgroup.con-
              trollers) in order to further redistribute  resources  at  lower
              levels  in the subtree.  (As an alternative to changing the own-
              ership of this file, the delegater might  instead  add  selected
              controllers to this file.)

       /dlgt_grp/cgroup.threads (cgroups v2 only)
              Changing  the  ownership of this file is necessary if a threaded
              subtree is being  delegated  (see  the  description  of  "thread
              mode",  below).   This permits the delegatee to write thread IDs
              to the file.  (The ownership of this file can  also  be  changed
              when  delegating  a domain subtree, but currently this serves no
              purpose, since, as described below, it is not possible to move a
              thread  between  domain  cgroups by writing its thread ID to the
              cgroup.threads file.)

              In cgroups v1, the corresponding file  that  should  instead  be
              delegated is the tasks file.

       The  delegater should not change the ownership of any of the controller
       interfaces files (e.g.,  pids.max,  memory.high)  in  dlgt_grp.   Those
       files are used from the next level above the delegated subtree in order
       to distribute resources into the subtree, and the delegatee should  not
       have  permission  to change the resources that are distributed into the
       delegated subtree.

       See also the discussion  of  the  /sys/kernel/cgroup/delegate  file  in
       NOTES for information about further delegatable files in cgroups v2.

       After  the  aforementioned steps have been performed, the delegatee can
       create child cgroups within the delegated subtree (the cgroup subdirec-
       tories  and  the files they contain will be owned by the delegatee) and
       move processes between cgroups in the subtree.  If some controllers are
       present  in  dlgt_grp/cgroup.subtree_control,  or the ownership of that
       file was passed to the delegatee, the delegatee can  also  control  the
       further  redistribution  of  the corresponding resources into the dele-
       gated subtree.

   Cgroups v2 delegation: nsdelegate and cgroup namespaces
       Starting with Linux 4.13, there is a second way to perform cgroup dele-
       gation  in  the  cgroups v2 hierarchy.  This is done by mounting or re-
       mounting the cgroup v2 filesystem with  the  nsdelegate  mount  option.
       For  example,  if the cgroup v2 filesystem has already been mounted, we
       can remount it with the nsdelegate option as follows:

           mount -t cgroup2 -o remount,nsdelegate \
                            none /sys/fs/cgroup/unified

       The effect of this mount option is to cause cgroup namespaces to  auto-
       matically become delegation boundaries.  More specifically, the follow-
       ing restrictions apply for processes inside the cgroup namespace:

       •  Writes to controller interface files in the root  directory  of  the
          namespace  will  fail  with  the  error EPERM.  Processes inside the
          cgroup namespace can still write to delegatable files  in  the  root
          directory   of   the  cgroup  namespace  such  as  cgroup.procs  and
          cgroup.subtree_control, and can create subhierarchy  underneath  the
          root directory.

       •  Attempts  to migrate processes across the namespace boundary are de-
          nied (with the error ENOENT).  Processes inside the cgroup namespace
          can  still  (subject  to the containment rules described below) move
          processes between cgroups within the subhierarchy  under  the  name-
          space root.

       The  ability to define cgroup namespaces as delegation boundaries makes
       cgroup namespaces more useful.  To understand why, suppose that we  al-
       ready  have one cgroup hierarchy that has been delegated to a nonprivi-
       leged user, cecilia, using the  older  delegation  technique  described
       above.   Suppose further that cecilia wanted to further delegate a sub-
       hierarchy under the existing delegated hierarchy.   (For  example,  the
       delegated  hierarchy might be associated with an unprivileged container
       run by cecilia.)  Even if a cgroup namespace was employed, because both
       hierarchies  are  owned by the unprivileged user cecilia, the following
       illegitimate actions could be performed:

       •  A process in the inferior hierarchy could change the  resource  con-
          troller  settings  in  the root directory of that hierarchy.  (These
          resource controller settings are intended to allow control to be ex-
          ercised  from  the  parent cgroup; a process inside the child cgroup
          should not be allowed to modify them.)

       •  A process inside the inferior hierarchy could  move  processes  into
          and out of the inferior hierarchy if the cgroups in the superior hi-
          erarchy were somehow visible.

       Employing the nsdelegate mount option prevents both of these possibili-
       ties.

       The  nsdelegate  mount  option only has an effect when performed in the
       initial mount namespace; in  other  mount  namespaces,  the  option  is
       silently ignored.

       Note:  On  some  systems, systemd(1) automatically mounts the cgroup v2
       filesystem.  In order to experiment with the nsdelegate  operation,  it
       may  be  useful  to boot the kernel with the following command-line op-
       tions:

           cgroup_no_v1=all systemd.legacy_systemd_cgroup_controller

       These options cause the kernel to boot with the cgroups v1  controllers
       disabled  (meaning that the controllers are available in the v2 hierar-
       chy), and tells systemd(1) not to mount and use the cgroup  v2  hierar-
       chy,  so that the v2 hierarchy can be manually mounted with the desired
       options after boot-up.

   Cgroup delegation containment rules
       Some delegation containment rules ensure that the  delegatee  can  move
       processes  between cgroups within the delegated subtree, but can't move
       processes from outside the delegated subtree into the subtree  or  vice
       versa.  A nonprivileged process (i.e., the delegatee) can write the PID
       of a "target" process into a cgroup.procs file only if all of the  fol-
       lowing are true:

       •  The writer has write permission on the cgroup.procs file in the des-
          tination cgroup.

       •  The writer has write permission on  the  cgroup.procs  file  in  the
          nearest common ancestor of the source and destination cgroups.  Note
          that in some cases, the nearest common ancestor may be the source or
          destination  cgroup  itself.   This  requirement is not enforced for
          cgroups v1 hierarchies, with the consequence that containment in  v1
          is  less  strict  than  in v2.  (For example, in cgroups v1 the user
          that owns two distinct delegated subhierarchies can move  a  process
          between the hierarchies.)

       •  If  the cgroup v2 filesystem was mounted with the nsdelegate option,
          the writer must be able to see the source  and  destination  cgroups
          from its cgroup namespace.

       •  In cgroups v1: the effective UID of the writer (i.e., the delegatee)
          matches the real user ID or the  saved  set-user-ID  of  the  target
          process.   Before  Linux  4.11,  this  requirement  also  applied in
          cgroups v2 (This was a historical requirement inherited from cgroups
          v1  that was later deemed unnecessary, since the other rules suffice
          for containment in cgroups v2.)

       Note: one consequence of these delegation containment rules is that the
       unprivileged delegatee can't place the first process into the delegated
       subtree; instead, the delegater must place the first process (a process
       owned by the delegatee) into the delegated subtree.

CGROUPS VERSION 2 THREAD MODE
       Among  the  restrictions imposed by cgroups v2 that were not present in
       cgroups v1 are the following:

       •  No thread-granularity control: all of the threads of a process  must
          be in the same cgroup.

       •  No internal processes: a cgroup can't both have member processes and
          exercise controllers on child cgroups.

       Both of these restrictions were added because the  lack  of  these  re-
       strictions  had  caused  problems  in  cgroups  v1.  In particular, the
       cgroups v1 ability to allow thread-level granularity for cgroup member-
       ship  made  no  sense for some controllers.  (A notable example was the
       memory controller: since threads share an address  space,  it  made  no
       sense to split threads across different memory cgroups.)

       Notwithstanding  the  initial design decision in cgroups v2, there were
       use cases for certain controllers,  notably  the  cpu  controller,  for
       which  thread-level  granularity  of control was meaningful and useful.
       To accommodate such use cases, Linux 4.14 added thread mode for cgroups
       v2.

       Thread mode allows the following:

       •  The  creation of threaded subtrees in which the threads of a process
          may be spread across cgroups inside the tree.  (A  threaded  subtree
          may contain multiple multithreaded processes.)

       •  The  concept of threaded controllers, which can distribute resources
          across the cgroups in a threaded subtree.

       •  A relaxation of the "no internal processes rule", so that, within  a
          threaded subtree, a cgroup can both contain member threads and exer-
          cise resource control over child cgroups.

       With the addition of thread mode, each nonroot cgroup  now  contains  a
       new  file,  cgroup.type, that exposes, and in some circumstances can be
       used to change, the "type" of a cgroup.  This file contains one of  the
       following type values:

       domain This  is  a  normal  v2 cgroup that provides process-granularity
              control.  If a process is a member  of  this  cgroup,  then  all
              threads  of  the process are (by definition) in the same cgroup.
              This is the default cgroup type, and provides the same  behavior
              that  was  provided for cgroups in the initial cgroups v2 imple-
              mentation.

       threaded
              This cgroup is a member of a threaded subtree.  Threads  can  be
              added  to  this  cgroup,  and controllers can be enabled for the
              cgroup.

       domain threaded
              This is a domain cgroup that serves as the root  of  a  threaded
              subtree.  This cgroup type is also known as "threaded root".

       domain invalid
              This  is  a  cgroup inside a threaded subtree that is in an "in-
              valid" state.  Processes can't be added to the cgroup, and  con-
              trollers  can't  be enabled for the cgroup.  The only thing that
              can be done with this cgroup (other than deleting it) is to con-
              vert it to a threaded cgroup by writing the string "threaded" to
              the cgroup.type file.

              The rationale for the existence of this  "interim"  type  during
              the  creation of a threaded subtree (rather than the kernel sim-
              ply immediately converting all cgroups under the  threaded  root
              to the type threaded) is to allow for possible future extensions
              to the thread mode model

   Threaded versus domain controllers
       With the addition of threads mode, cgroups  v2  now  distinguishes  two
       types of resource controllers:

       •  Threaded  controllers:  these controllers support thread-granularity
          for resource control and can be enabled  inside  threaded  subtrees,
          with  the  result  that the corresponding controller-interface files
          appear inside the cgroups in the  threaded  subtree.   As  at  Linux
          4.19,  the  following controllers are threaded: cpu, perf_event, and
          pids.

       •  Domain controllers: these controllers support only process granular-
          ity  for  resource  control.   From the perspective of a domain con-
          troller, all threads of a process are always  in  the  same  cgroup.
          Domain controllers can't be enabled inside a threaded subtree.

   Creating a threaded subtree
       There are two pathways that lead to the creation of a threaded subtree.
       The first pathway proceeds as follows:

       (1)  We write the string "threaded" to the cgroup.type file of a cgroup
            y/z  that  currently  has the type domain.  This has the following
            effects:

            •  The type of the cgroup y/z becomes threaded.

            •  The type of the parent cgroup, y, becomes domain threaded.  The
               parent  cgroup is the root of a threaded subtree (also known as
               the "threaded root").

            •  All other cgroups  under  y  that  were  not  already  of  type
               threaded  (because  they  were inside already existing threaded
               subtrees under the new threaded root) are converted to type do-
               main  invalid.   Any  subsequently created cgroups under y will
               also have the type domain invalid.

       (2)  We write the string "threaded"  to  each  of  the  domain  invalid
            cgroups  under  y,  in order to convert them to the type threaded.
            As a consequence of this step, all threads under the threaded root
            now  have  the type threaded and the threaded subtree is now fully
            usable.  The requirement to write  "threaded"  to  each  of  these
            cgroups is somewhat cumbersome, but allows for possible future ex-
            tensions to the thread-mode model.

       The second way of creating a threaded subtree is as follows:

       (1)  In an existing cgroup, z, that currently has the type  domain,  we
            (1.1)  enable  one  or  more threaded controllers and (1.2) make a
            process a member of z.  (These two steps can be done in either or-
            der.)  This has the following consequences:

            •  The type of z becomes domain threaded.

            •  All  of  the  descendant  cgroups of x that were not already of
               type threaded are converted to type domain invalid.

       (2)  As before, we make the threaded  subtree  usable  by  writing  the
            string  "threaded"  to each of the domain invalid cgroups under y,
            in order to convert them to the type threaded.

       One of the consequences of the above pathways to  creating  a  threaded
       subtree  is  that  the  threaded  root  cgroup  can be a parent only to
       threaded (and domain invalid) cgroups.  The threaded root cgroup  can't
       be  a  parent  of  a domain cgroups, and a threaded cgroup can't have a
       sibling that is a domain cgroup.

   Using a threaded subtree
       Within a threaded subtree, threaded controllers can be enabled in  each
       subgroup  whose  type  has been changed to threaded; upon doing so, the
       corresponding controller interface files appear in the children of that
       cgroup.

       A  process  can  be moved into a threaded subtree by writing its PID to
       the cgroup.procs file in one of the cgroups inside the tree.  This  has
       the  effect  of making all of the threads in the process members of the
       corresponding cgroup and makes the process a  member  of  the  threaded
       subtree.   The  threads  of  the  process can then be spread across the
       threaded subtree by writing their thread IDs  (see  gettid(2))  to  the
       cgroup.threads  files  in  different  cgroups  inside the subtree.  The
       threads of a process must all reside in the same threaded subtree.

       As with writing to cgroup.procs,  some  containment  rules  apply  when
       writing to the cgroup.threads file:

       •  The  writer must have write permission on the cgroup.threads file in
          the destination cgroup.

       •  The writer must have write permission on the  cgroup.procs  file  in
          the common ancestor of the source and destination cgroups.  (In some
          cases, the common ancestor may be the source or  destination  cgroup
          itself.)

       •  The source and destination cgroups must be in the same threaded sub-
          tree.  (Outside a threaded subtree, an attempt to move a  thread  by
          writing  its thread ID to the cgroup.threads file in a different do-
          main cgroup fails with the error EOPNOTSUPP.)

       The cgroup.threads file is present in  each  cgroup  (including  domain
       cgroups)  and  can be read in order to discover the set of threads that
       is present in the cgroup.  The set of thread IDs obtained when  reading
       this file is not guaranteed to be ordered or free of duplicates.

       The  cgroup.procs  file in the threaded root shows the PIDs of all pro-
       cesses that are members of  the  threaded  subtree.   The  cgroup.procs
       files in the other cgroups in the subtree are not readable.

       Domain  controllers  can't  be  enabled  in a threaded subtree; no con-
       troller-interface  files  appear  inside  the  cgroups  underneath  the
       threaded root.  From the point of view of a domain controller, threaded
       subtrees are invisible: a multithreaded process inside a threaded  sub-
       tree  appears  to  a domain controller as a process that resides in the
       threaded root cgroup.

       Within a threaded subtree, the "no internal processes"  rule  does  not
       apply: a cgroup can both contain member processes (or thread) and exer-
       cise controllers on child cgroups.

   Rules for writing to cgroup.type and creating threaded subtrees
       A number of rules apply when writing to the cgroup.type file:

       •  Only the string "threaded" may be written.  In other words, the only
          explicit  transition  that is possible is to convert a domain cgroup
          to type threaded.

       •  The effect of writing "threaded" depends on  the  current  value  in
          cgroup.type, as follows:

          •  domain  or domain threaded: start the creation of a threaded sub-
             tree (whose root is the parent of this cgroup) via the  first  of
             the pathways described above;

          •  domain invalid:  convert  this cgroup (which is inside a threaded
             subtree) to a usable (i.e., threaded) state;

          •  threaded: no effect (a "no-op").

       •  We can't write to a cgroup.type file if the parent's type is  domain
          invalid.   In other words, the cgroups of a threaded subtree must be
          converted to the threaded state in a top-down manner.

       There are also some constraints that must be satisfied in order to cre-
       ate a threaded subtree rooted at the cgroup x:

       •  There  can  be  no  member processes in the descendant cgroups of x.
          (The cgroup x can itself have member processes.)

       •  No domain controllers may be enabled in  x's  cgroup.subtree_control
          file.

       If  any  of the above constraints is violated, then an attempt to write
       "threaded" to a cgroup.type file fails with the error ENOTSUP.

   The "domain threaded" cgroup type
       According to the pathways described above, the type  of  a  cgroup  can
       change to domain threaded in either of the following cases:

       •  The string "threaded" is written to a child cgroup.

       •  A  threaded controller is enabled inside the cgroup and a process is
          made a member of the cgroup.

       A domain threaded cgroup, x, can revert to the type domain if the above
       conditions  no  longer hold true—that is, if all threaded child cgroups
       of x are removed and either x no longer has  threaded  controllers  en-
       abled or no longer has member processes.

       When a domain threaded cgroup x reverts to the type domain:

       •  All  domain  invalid  descendants  of  x that are not in lower-level
          threaded subtrees revert to the type domain.

       •  The root cgroups in any lower-level threaded subtrees revert to  the
          type domain threaded.

   Exceptions for the root cgroup
       The root cgroup of the v2 hierarchy is treated exceptionally: it can be
       the parent  of  both  domain  and  threaded  cgroups.   If  the  string
       "threaded" is written to the cgroup.type file of one of the children of
       the root cgroup, then

       •  The type of that cgroup becomes threaded.

       •  The type of any descendants of that cgroup  that  are  not  part  of
          lower-level threaded subtrees changes to domain invalid.

       Note  that  in  this case, there is no cgroup whose type becomes domain
       threaded.  (Notionally, the  root  cgroup  can  be  considered  as  the
       threaded root for the cgroup whose type was changed to threaded.)

       The aim of this exceptional treatment for the root cgroup is to allow a
       threaded cgroup that employs the cpu controller to be placed as high as
       possible  in  the  hierarchy,  so  as  to  minimize the (small) cost of
       traversing the cgroup hierarchy.

   The cgroups v2 "cpu" controller and realtime threads
       As at Linux 4.19, the cgroups v2 cpu controller does not  support  con-
       trol  of  realtime threads (specifically threads scheduled under any of
       the  policies  SCHED_FIFO,  SCHED_RR,  described  SCHED_DEADLINE;   see
       sched(7)).   Therefore,  the  cpu controller can be enabled in the root
       cgroup only if all realtime threads are in the root cgroup.  (If  there
       are  realtime threads in nonroot cgroups, then a write(2) of the string
       "+cpu" to the cgroup.subtree_control file fails with the error EINVAL.)

       On some systems, systemd(1) places certain realtime threads in  nonroot
       cgroups in the v2 hierarchy.  On such systems, these threads must first
       be moved to the root cgroup before the cpu controller can be enabled.

ERRORS
       The following errors can occur for mount(2):

       EBUSY  An attempt to mount a cgroup version 1 filesystem specified nei-
              ther  the  name=  option (to mount a named hierarchy) nor a con-
              troller name (or all).

NOTES
       A child process created via fork(2) inherits its parent's  cgroup  mem-
       berships.   A  process's  cgroup  memberships  are preserved across ex-
       ecve(2).

       The clone3(2) CLONE_INTO_CGROUP flag can be  used  to  create  a  child
       process  that  begins its life in a different version 2 cgroup from the
       parent process.

   /proc files
       /proc/cgroups (since Linux 2.6.24)
              This file contains information about the  controllers  that  are
              compiled  into  the  kernel.  An example of the contents of this
              file (reformatted for readability) is the following:

                  #subsys_name    hierarchy      num_cgroups    enabled
                  cpuset          4              1              1
                  cpu             8              1              1
                  cpuacct         8              1              1
                  blkio           6              1              1
                  memory          3              1              1
                  devices         10             84             1
                  freezer         7              1              1
                  net_cls         9              1              1
                  perf_event      5              1              1
                  net_prio        9              1              1
                  hugetlb         0              1              0
                  pids            2              1              1

              The fields in this file are, from left to right:

              [1]  The name of the controller.

              [2]  The unique ID of the cgroup hierarchy on  which  this  con-
                   troller is mounted.  If multiple cgroups v1 controllers are
                   bound to the same hierarchy, then each will show  the  same
                   hierarchy  ID  in this field.  The value in this field will
                   be 0 if:

                   •  the controller is not mounted on a cgroups v1 hierarchy;

                   •  the controller is bound to the cgroups v2 single unified
                      hierarchy; or

                   •  the controller is disabled (see below).

              [3]  The  number  of control groups in this hierarchy using this
                   controller.

              [4]  This field contains the value 1 if this controller  is  en-
                   abled, or 0 if it has been disabled (via the cgroup_disable
                   kernel command-line boot parameter).

       /proc/[pid]/cgroup (since Linux 2.6.24)
              This file describes control groups to which the process with the
              corresponding  PID  belongs.   The displayed information differs
              for cgroups version 1 and version 2 hierarchies.

              For each cgroup hierarchy of which  the  process  is  a  member,
              there is one entry containing three colon-separated fields:

                  hierarchy-ID:controller-list:cgroup-path

              For example:

                  5:cpuacct,cpu,cpuset:/daemons

              The colon-separated fields are, from left to right:

              [1]  For  cgroups  version  1 hierarchies, this field contains a
                   unique hierarchy ID number that can be matched to a hierar-
                   chy ID in /proc/cgroups.  For the cgroups version 2 hierar-
                   chy, this field contains the value 0.

              [2]  For cgroups version 1 hierarchies, this  field  contains  a
                   comma-separated  list of the controllers bound to the hier-
                   archy.  For the cgroups version 2 hierarchy, this field  is
                   empty.

              [3]  This  field  contains  the pathname of the control group in
                   the hierarchy to which the process belongs.  This  pathname
                   is relative to the mount point of the hierarchy.

   /sys/kernel/cgroup files
       /sys/kernel/cgroup/delegate (since Linux 4.15)
              This  file exports a list of the cgroups v2 files (one per line)
              that are delegatable (i.e., whose ownership should be changed to
              the  user ID of the delegatee).  In the future, the set of dele-
              gatable files may change or grow, and this file provides  a  way
              for  the kernel to inform user-space applications of which files
              must be delegated.  As at Linux 4.15,  one  sees  the  following
              when inspecting this file:

                  $ cat /sys/kernel/cgroup/delegate
                  cgroup.procs
                  cgroup.subtree_control
                  cgroup.threads

       /sys/kernel/cgroup/features (since Linux 4.15)
              Over  time,  the set of cgroups v2 features that are provided by
              the kernel may change or grow, or some features may not  be  en-
              abled  by  default.  This file provides a way for user-space ap-
              plications to discover what features the running kernel supports
              and has enabled.  Features are listed one per line:

                  $ cat /sys/kernel/cgroup/features
                  nsdelegate
                  memory_localevents

              The entries that can appear in this file are:

              memory_localevents (since Linux 5.2)
                     The kernel supports the memory_localevents mount option.

              nsdelegate (since Linux 4.15)
                     The kernel supports the nsdelegate mount option.

              memory_recursiveprot (since Linux 5.7)
                     The  kernel  supports  the memory_recursiveprot mount op-
                     tion.

SEE ALSO
       prlimit(1), systemd(1),  systemd-cgls(1),  systemd-cgtop(1),  clone(2),
       ioprio_set(2),  perf_event_open(2), setrlimit(2), cgroup_namespaces(7),
       cpuset(7), namespaces(7), sched(7), user_namespaces(7)

       The kernel source file Documentation/admin-guide/cgroup-v2.rst.

Linux man-pages 6.03              2023-02-05                        cgroups(7)

Generated by dwww version 1.15 on Fri Jun 21 07:23:51 CEST 2024.