CGROUPS(7) Linux Programmer's Manual CGROUPS(7)
NAME
cgroups - Linux control groups
DESCRIPTION
Control groups, usually referred to as cgroups, are a Linux kernel fea-
ture which allow processes to be organized into hierarchical groups
whose usage of various types of resources can then be limited and moni-
tored. The kernel's cgroup interface is provided through a pseudo-
filesystem called cgroupfs. Grouping is implemented in the core cgroup
kernel code, while resource tracking and limits are implemented in a
set of per-resource-type subsystems (memory, CPU, and so on).
Terminology
A cgroup is a collection of processes that are bound to a set of limits
or parameters defined via the cgroup filesystem.
A subsystem is a kernel component that modifies the behavior of the
processes in a cgroup. Various subsystems have been implemented, mak-
ing it possible to do things such as limiting the amount of CPU time
and memory available to a cgroup, accounting for the CPU time used by a
cgroup, and freezing and resuming execution of the processes in a
cgroup. Subsystems are sometimes also known as resource controllers
(or simply, controllers).
The cgroups for a controller are arranged in a hierarchy. This hierar-
chy is defined by creating, removing, and renaming subdirectories
within the cgroup filesystem. At each level of the hierarchy, at-
tributes (e.g., limits) can be defined. The limits, control, and ac-
counting provided by cgroups generally have effect throughout the sub-
hierarchy underneath the cgroup where the attributes are defined.
Thus, for example, the limits placed on a cgroup at a higher level in
the hierarchy cannot be exceeded by descendant cgroups.
Cgroups version 1 and version 2
The initial release of the cgroups implementation was in Linux 2.6.24.
Over time, various cgroup controllers have been added to allow the man-
agement of various types of resources. However, the development of
these controllers was largely uncoordinated, with the result that many
inconsistencies arose between controllers and management of the cgroup
hierarchies became rather complex. (A longer description of these
problems can be found in the kernel source file Documenta-
tion/cgroup-v2.txt.)
Because of the problems with the initial cgroups implementation
(cgroups version 1), starting in Linux 3.10, work began on a new, or-
thogonal implementation to remedy these problems. Initially marked ex-
perimental, and hidden behind the -o __DEVEL__sane_behavior mount op-
tion, the new version (cgroups version 2) was eventually made official
with the release of Linux 4.5. Differences between the two versions
are described in the text below. The file cgroup.sane_behavior,
present in cgroups v1, is a relic of this mount option. The file always
reports "0" and is only retained for backward compatibility.
Although cgroups v2 is intended as a replacement for cgroups v1, the
older system continues to exist (and for compatibility reasons is un-
likely to be removed). Currently, cgroups v2 implements only a subset
of the controllers available in cgroups v1. The two systems are imple-
mented so that both v1 controllers and v2 controllers can be mounted on
the same system. Thus, for example, it is possible to use those con-
trollers that are supported under version 2, while also using version 1
controllers where version 2 does not yet support those controllers.
The only restriction here is that a controller can't be simultaneously
employed in both a cgroups v1 hierarchy and in the cgroups v2 hierar-
chy.
CGROUPS VERSION 1
Under cgroups v1, each controller may be mounted against a separate
cgroup filesystem that provides its own hierarchical organization of
the processes on the system. It is also possible to comount multiple
(or even all) cgroups v1 controllers against the same cgroup filesys-
tem, meaning that the comounted controllers manage the same hierarchi-
cal organization of processes.
For each mounted hierarchy, the directory tree mirrors the control
group hierarchy. Each control group is represented by a directory,
with each of its child control cgroups represented as a child direc-
tory. For instance, /user/joe/1.session represents control group
1.session, which is a child of cgroup joe, which is a child of /user.
Under each cgroup directory is a set of files which can be read or
written to, reflecting resource limits and a few general cgroup proper-
ties.
Tasks (threads) versus processes
In cgroups v1, a distinction is drawn between processes and tasks. In
this view, a process can consist of multiple tasks (more commonly
called threads, from a user-space perspective, and called such in the
remainder of this man page). In cgroups v1, it is possible to indepen-
dently manipulate the cgroup memberships of the threads in a process.
The cgroups v1 ability to split threads across different cgroups caused
problems in some cases. For example, it made no sense for the memory
controller, since all of the threads of a process share a single ad-
dress space. Because of these problems, the ability to independently
manipulate the cgroup memberships of the threads in a process was re-
moved in the initial cgroups v2 implementation, and subsequently re-
stored in a more limited form (see the discussion of "thread mode" be-
low).
Mounting v1 controllers
The use of cgroups requires a kernel built with the CONFIG_CGROUP op-
tion. In addition, each of the v1 controllers has an associated con-
figuration option that must be set in order to employ that controller.
In order to use a v1 controller, it must be mounted against a cgroup
filesystem. The usual place for such mounts is under a tmpfs(5)
filesystem mounted at /sys/fs/cgroup. Thus, one might mount the cpu
controller as follows:
mount -t cgroup -o cpu none /sys/fs/cgroup/cpu
It is possible to comount multiple controllers against the same hierar-
chy. For example, here the cpu and cpuacct controllers are comounted
against a single hierarchy:
mount -t cgroup -o cpu,cpuacct none /sys/fs/cgroup/cpu,cpuacct
Comounting controllers has the effect that a process is in the same
cgroup for all of the comounted controllers. Separately mounting con-
trollers allows a process to be in cgroup /foo1 for one controller
while being in /foo2/foo3 for another.
It is possible to comount all v1 controllers against the same hierar-
chy:
mount -t cgroup -o all cgroup /sys/fs/cgroup
(One can achieve the same result by omitting -o all, since it is the
default if no controllers are explicitly specified.)
It is not possible to mount the same controller against multiple cgroup
hierarchies. For example, it is not possible to mount both the cpu and
cpuacct controllers against one hierarchy, and to mount the cpu con-
troller alone against another hierarchy. It is possible to create mul-
tiple mount points with exactly the same set of comounted controllers.
However, in this case all that results is multiple mount points provid-
ing a view of the same hierarchy.
Note that on many systems, the v1 controllers are automatically mounted
under /sys/fs/cgroup; in particular, systemd(1) automatically creates
such mount points.
Unmounting v1 controllers
A mounted cgroup filesystem can be unmounted using the umount(8) com-
mand, as in the following example:
umount /sys/fs/cgroup/pids
But note well: a cgroup filesystem is unmounted only if it is not busy,
that is, it has no child cgroups. If this is not the case, then the
only effect of the umount(8) is to make the mount invisible. Thus, to
ensure that the mount point is really removed, one must first remove
all child cgroups, which in turn can be done only after all member pro-
cesses have been moved from those cgroups to the root cgroup.
Cgroups version 1 controllers
Each of the cgroups version 1 controllers is governed by a kernel con-
figuration option (listed below). Additionally, the availability of
the cgroups feature is governed by the CONFIG_CGROUPS kernel configura-
tion option.
cpu (since Linux 2.6.24; CONFIG_CGROUP_SCHED)
Cgroups can be guaranteed a minimum number of "CPU shares" when
a system is busy. This does not limit a cgroup's CPU usage if
the CPUs are not busy. For further information, see Documenta-
tion/scheduler/sched-design-CFS.txt.
In Linux 3.2, this controller was extended to provide CPU "band-
width" control. If the kernel is configured with CON-
FIG_CFS_BANDWIDTH, then within each scheduling period (defined
via a file in the cgroup directory), it is possible to define an
upper limit on the CPU time allocated to the processes in a
cgroup. This upper limit applies even if there is no other com-
petition for the CPU. Further information can be found in the
kernel source file Documentation/scheduler/sched-bwc.txt.
cpuacct (since Linux 2.6.24; CONFIG_CGROUP_CPUACCT)
This provides accounting for CPU usage by groups of processes.
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/cpuacct.txt.
cpuset (since Linux 2.6.24; CONFIG_CPUSETS)
This cgroup can be used to bind the processes in a cgroup to a
specified set of CPUs and NUMA nodes.
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/cpusets.txt.
memory (since Linux 2.6.25; CONFIG_MEMCG)
The memory controller supports reporting and limiting of process
memory, kernel memory, and swap used by cgroups.
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/memory.txt.
devices (since Linux 2.6.26; CONFIG_CGROUP_DEVICE)
This supports controlling which processes may create (mknod) de-
vices as well as open them for reading or writing. The policies
may be specified as allow-lists and deny-lists. Hierarchy is
enforced, so new rules must not violate existing rules for the
target or ancestor cgroups.
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/devices.txt.
freezer (since Linux 2.6.28; CONFIG_CGROUP_FREEZER)
The freezer cgroup can suspend and restore (resume) all pro-
cesses in a cgroup. Freezing a cgroup /A also causes its chil-
dren, for example, processes in /A/B, to be frozen.
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/freezer-subsystem.txt.
net_cls (since Linux 2.6.29; CONFIG_CGROUP_NET_CLASSID)
This places a classid, specified for the cgroup, on network
packets created by a cgroup. These classids can then be used in
firewall rules, as well as used to shape traffic using tc(8).
This applies only to packets leaving the cgroup, not to traffic
arriving at the cgroup.
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/net_cls.txt.
blkio (since Linux 2.6.33; CONFIG_BLK_CGROUP)
The blkio cgroup controls and limits access to specified block
devices by applying IO control in the form of throttling and up-
per limits against leaf nodes and intermediate nodes in the
storage hierarchy.
Two policies are available. The first is a proportional-weight
time-based division of disk implemented with CFQ. This is in
effect for leaf nodes using CFQ. The second is a throttling
policy which specifies upper I/O rate limits on a device.
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/blkio-controller.txt.
perf_event (since Linux 2.6.39; CONFIG_CGROUP_PERF)
This controller allows perf monitoring of the set of processes
grouped in a cgroup.
Further information can be found in the kernel source file
tools/perf/Documentation/perf-record.txt.
net_prio (since Linux 3.3; CONFIG_CGROUP_NET_PRIO)
This allows priorities to be specified, per network interface,
for cgroups.
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/net_prio.txt.
hugetlb (since Linux 3.5; CONFIG_CGROUP_HUGETLB)
This supports limiting the use of huge pages by cgroups.
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/hugetlb.txt.
pids (since Linux 4.3; CONFIG_CGROUP_PIDS)
This controller permits limiting the number of process that may
be created in a cgroup (and its descendants).
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/pids.txt.
rdma (since Linux 4.11; CONFIG_CGROUP_RDMA)
The RDMA controller permits limiting the use of RDMA/IB-specific
resources per cgroup.
Further information can be found in the kernel source file Docu-
mentation/cgroup-v1/rdma.txt.
Creating cgroups and moving processes
A cgroup filesystem initially contains a single root cgroup, '/', which
all processes belong to. A new cgroup is created by creating a direc-
tory in the cgroup filesystem:
mkdir /sys/fs/cgroup/cpu/cg1
This creates a new empty cgroup.
A process may be moved to this cgroup by writing its PID into the
cgroup's cgroup.procs file:
echo $$ > /sys/fs/cgroup/cpu/cg1/cgroup.procs
Only one PID at a time should be written to this file.
Writing the value 0 to a cgroup.procs file causes the writing process
to be moved to the corresponding cgroup.
When writing a PID into the cgroup.procs, all threads in the process
are moved into the new cgroup at once.
Within a hierarchy, a process can be a member of exactly one cgroup.
Writing a process's PID to a cgroup.procs file automatically removes it
from the cgroup of which it was previously a member.
The cgroup.procs file can be read to obtain a list of the processes
that are members of a cgroup. The returned list of PIDs is not guaran-
teed to be in order. Nor is it guaranteed to be free of duplicates.
(For example, a PID may be recycled while reading from the list.)
In cgroups v1, an individual thread can be moved to another cgroup by
writing its thread ID (i.e., the kernel thread ID returned by clone(2)
and gettid(2)) to the tasks file in a cgroup directory. This file can
be read to discover the set of threads that are members of the cgroup.
Removing cgroups
To remove a cgroup, it must first have no child cgroups and contain no
(nonzombie) processes. So long as that is the case, one can simply re-
move the corresponding directory pathname. Note that files in a cgroup
directory cannot and need not be removed.
Cgroups v1 release notification
Two files can be used to determine whether the kernel provides notifi-
cations when a cgroup becomes empty. A cgroup is considered to be
empty when it contains no child cgroups and no member processes.
A special file in the root directory of each cgroup hierarchy, re-
lease_agent, can be used to register the pathname of a program that may
be invoked when a cgroup in the hierarchy becomes empty. The pathname
of the newly empty cgroup (relative to the cgroup mount point) is pro-
vided as the sole command-line argument when the release_agent program
is invoked. The release_agent program might remove the cgroup direc-
tory, or perhaps repopulate it with a process.
The default value of the release_agent file is empty, meaning that no
release agent is invoked.
The content of the release_agent file can also be specified via a mount
option when the cgroup filesystem is mounted:
mount -o release_agent=pathname ...
Whether or not the release_agent program is invoked when a particular
cgroup becomes empty is determined by the value in the notify_on_re-
lease file in the corresponding cgroup directory. If this file con-
tains the value 0, then the release_agent program is not invoked. If
it contains the value 1, the release_agent program is invoked. The de-
fault value for this file in the root cgroup is 0. At the time when a
new cgroup is created, the value in this file is inherited from the
corresponding file in the parent cgroup.
Cgroup v1 named hierarchies
In cgroups v1, it is possible to mount a cgroup hierarchy that has no
attached controllers:
mount -t cgroup -o none,name=somename none /some/mount/point
Multiple instances of such hierarchies can be mounted; each hierarchy
must have a unique name. The only purpose of such hierarchies is to
track processes. (See the discussion of release notification below.)
An example of this is the name=systemd cgroup hierarchy that is used by
systemd(1) to track services and user sessions.
Since Linux 5.0, the cgroup_no_v1 kernel boot option (described below)
can be used to disable cgroup v1 named hierarchies, by specifying
cgroup_no_v1=named.
CGROUPS VERSION 2
In cgroups v2, all mounted controllers reside in a single unified hier-
archy. While (different) controllers may be simultaneously mounted un-
der the v1 and v2 hierarchies, it is not possible to mount the same
controller simultaneously under both the v1 and the v2 hierarchies.
The new behaviors in cgroups v2 are summarized here, and in some cases
elaborated in the following subsections.
1. Cgroups v2 provides a unified hierarchy against which all con-
trollers are mounted.
2. "Internal" processes are not permitted. With the exception of the
root cgroup, processes may reside only in leaf nodes (cgroups that
do not themselves contain child cgroups). The details are somewhat
more subtle than this, and are described below.
3. Active cgroups must be specified via the files cgroup.controllers
and cgroup.subtree_control.
4. The tasks file has been removed. In addition, the
cgroup.clone_children file that is employed by the cpuset controller
has been removed.
5. An improved mechanism for notification of empty cgroups is provided
by the cgroup.events file.
For more changes, see the Documentation/cgroup-v2.txt file in the ker-
nel source.
Some of the new behaviors listed above saw subsequent modification with
the addition in Linux 4.14 of "thread mode" (described below).
Cgroups v2 unified hierarchy
In cgroups v1, the ability to mount different controllers against dif-
ferent hierarchies was intended to allow great flexibility for applica-
tion design. In practice, though, the flexibility turned out to be
less useful than expected, and in many cases added complexity. There-
fore, in cgroups v2, all available controllers are mounted against a
single hierarchy. The available controllers are automatically mounted,
meaning that it is not necessary (or possible) to specify the con-
trollers when mounting the cgroup v2 filesystem using a command such as
the following:
mount -t cgroup2 none /mnt/cgroup2
A cgroup v2 controller is available only if it is not currently in use
via a mount against a cgroup v1 hierarchy. Or, to put things another
way, it is not possible to employ the same controller against both a v1
hierarchy and the unified v2 hierarchy. This means that it may be nec-
essary first to unmount a v1 controller (as described above) before
that controller is available in v2. Since systemd(1) makes heavy use
of some v1 controllers by default, it can in some cases be simpler to
boot the system with selected v1 controllers disabled. To do this,
specify the cgroup_no_v1=list option on the kernel boot command line;
list is a comma-separated list of the names of the controllers to dis-
able, or the word all to disable all v1 controllers. (This situation
is correctly handled by systemd(1), which falls back to operating with-
out the specified controllers.)
Note that on many modern systems, systemd(1) automatically mounts the
cgroup2 filesystem at /sys/fs/cgroup/unified during the boot process.
Cgroups v2 mount options
The following options (mount -o) can be specified when mounting the
group v2 filesystem:
nsdelegate (since Linux 4.15)
Treat cgroup namespaces as delegation boundaries. For details,
see below.
memory_localevents (since Linux 5.2)
The memory.events should show statistics only for the cgroup it-
self, and not for any descendant cgroups. This was the behavior
before Linux 5.2. Starting in Linux 5.2, the default behavior
is to include statistics for descendant cgroups in mem-
ory.events, and this mount option can be used to revert to the
legacy behavior. This option is system wide and can be set on
mount or modified through remount only from the initial mount
namespace; it is silently ignored in noninitial namespaces.
Cgroups v2 controllers
The following controllers, documented in the kernel source file Docu-
mentation/cgroup-v2.txt, are supported in cgroups version 2:
cpu (since Linux 4.15)
This is the successor to the version 1 cpu and cpuacct con-
trollers.
cpuset (since Linux 5.0)
This is the successor of the version 1 cpuset controller.
freezer (since Linux 5.2)
This is the successor of the version 1 freezer controller.
hugetlb (since Linux 5.6)
This is the successor of the version 1 hugetlb controller.
io (since Linux 4.5)
This is the successor of the version 1 blkio controller.
memory (since Linux 4.5)
This is the successor of the version 1 memory controller.
perf_event (since Linux 4.11)
This is the same as the version 1 perf_event controller.
pids (since Linux 4.5)
This is the same as the version 1 pids controller.
rdma (since Linux 4.11)
This is the same as the version 1 rdma controller.
There is no direct equivalent of the net_cls and net_prio controllers
from cgroups version 1. Instead, support has been added to iptables(8)
to allow eBPF filters that hook on cgroup v2 pathnames to make deci-
sions about network traffic on a per-cgroup basis.
The v2 devices controller provides no interface files; instead, device
control is gated by attaching an eBPF (BPF_CGROUP_DEVICE) program to a
v2 cgroup.
Cgroups v2 subtree control
Each cgroup in the v2 hierarchy contains the following two files:
cgroup.controllers
This read-only file exposes a list of the controllers that are
available in this cgroup. The contents of this file match the
contents of the cgroup.subtree_control file in the parent
cgroup.
cgroup.subtree_control
This is a list of controllers that are active (enabled) in the
cgroup. The set of controllers in this file is a subset of the
set in the cgroup.controllers of this cgroup. The set of active
controllers is modified by writing strings to this file contain-
ing space-delimited controller names, each preceded by '+' (to
enable a controller) or '-' (to disable a controller), as in the
following example:
echo '+pids -memory' > x/y/cgroup.subtree_control
An attempt to enable a controller that is not present in
cgroup.controllers leads to an ENOENT error when writing to the
cgroup.subtree_control file.
Because the list of controllers in cgroup.subtree_control is a subset
of those cgroup.controllers, a controller that has been disabled in one
cgroup in the hierarchy can never be re-enabled in the subtree below
that cgroup.
A cgroup's cgroup.subtree_control file determines the set of con-
trollers that are exercised in the child cgroups. When a controller
(e.g., pids) is present in the cgroup.subtree_control file of a parent
cgroup, then the corresponding controller-interface files (e.g.,
pids.max) are automatically created in the children of that cgroup and
can be used to exert resource control in the child cgroups.
Cgroups v2 "no internal processes" rule
Cgroups v2 enforces a so-called "no internal processes" rule. Roughly
speaking, this rule means that, with the exception of the root cgroup,
processes may reside only in leaf nodes (cgroups that do not themselves
contain child cgroups). This avoids the need to decide how to parti-
tion resources between processes which are members of cgroup A and pro-
cesses in child cgroups of A.
For instance, if cgroup /cg1/cg2 exists, then a process may reside in
/cg1/cg2, but not in /cg1. This is to avoid an ambiguity in cgroups v1
with respect to the delegation of resources between processes in /cg1
and its child cgroups. The recommended approach in cgroups v2 is to
create a subdirectory called leaf for any nonleaf cgroup which should
contain processes, but no child cgroups. Thus, processes which previ-
ously would have gone into /cg1 would now go into /cg1/leaf. This has
the advantage of making explicit the relationship between processes in
/cg1/leaf and /cg1's other children.
The "no internal processes" rule is in fact more subtle than stated
above. More precisely, the rule is that a (nonroot) cgroup can't both
(1) have member processes, and (2) distribute resources into child
cgroups--that is, have a nonempty cgroup.subtree_control file. Thus,
it is possible for a cgroup to have both member processes and child
cgroups, but before controllers can be enabled for that cgroup, the
member processes must be moved out of the cgroup (e.g., perhaps into
the child cgroups).
With the Linux 4.14 addition of "thread mode" (described below), the
"no internal processes" rule has been relaxed in some cases.
Cgroups v2 cgroup.events file
Each nonroot cgroup in the v2 hierarchy contains a read-only file,
cgroup.events, whose contents are key-value pairs (delimited by newline
characters, with the key and value separated by spaces) providing state
information about the the cgroup:
$ cat mygrp/cgroup.events
populated 1
frozen 0
The following keys may appear in this file:
populated
The value of this key is either 1, if this cgroup or any of its
descendants has member processes, or otherwise 0.
frozen (since Linux 5.2)
The value of this key is 1 if this cgroup is currently frozen,
or 0 if it is not.
The cgroup.events file can be monitored, in order to receive notifica-
tion when the value of one of its keys changes. Such monitoring can be
done using inotify(7), which notifies changes as IN_MODIFY events, or
poll(2), which notifies changes by returning the POLLPRI and POLLERR
bits in the revents field.
Cgroup v2 release notification
Cgroups v2 provides a new mechanism for obtaining notification when a
cgroup becomes empty. The cgroups v1 release_agent and notify_on_re-
lease files are removed, and replaced by the populated key in the
cgroup.events file. This key either has the value 0, meaning that the
cgroup (and its descendants) contain no (nonzombie) member processes,
or 1, meaning that the cgroup (or one of its descendants) contains mem-
ber processes.
The cgroups v2 release-notification mechanism offers the following ad-
vantages over the cgroups v1 release_agent mechanism:
* It allows for cheaper notification, since a single process can moni-
tor multiple cgroup.events files (using the techniques described
earlier). By contrast, the cgroups v1 mechanism requires the ex-
pense of creating a process for each notification.
* Notification for different cgroup subhierarchies can be delegated to
different processes. By contrast, the cgroups v1 mechanism allows
only one release agent for an entire hierarchy.
Cgroups v2 cgroup.stat file
Each cgroup in the v2 hierarchy contains a read-only cgroup.stat file
(first introduced in Linux 4.14) that consists of lines containing key-
value pairs. The following keys currently appear in this file:
nr_descendants
This is the total number of visible (i.e., living) descendant
cgroups underneath this cgroup.
nr_dying_descendants
This is the total number of dying descendant cgroups underneath
this cgroup. A cgroup enters the dying state after being
deleted. It remains in that state for an undefined period
(which will depend on system load) while resources are freed be-
fore the cgroup is destroyed. Note that the presence of some
cgroups in the dying state is normal, and is not indicative of
any problem.
A process can't be made a member of a dying cgroup, and a dying
cgroup can't be brought back to life.
Limiting the number of descendant cgroups
Each cgroup in the v2 hierarchy contains the following files, which can
be used to view and set limits on the number of descendant cgroups un-
der that cgroup:
cgroup.max.depth (since Linux 4.14)
This file defines a limit on the depth of nesting of descendant
cgroups. A value of 0 in this file means that no descendant
cgroups can be created. An attempt to create a descendant whose
nesting level exceeds the limit fails (mkdir(2) fails with the
error EAGAIN).
Writing the string "max" to this file means that no limit is im-
posed. The default value in this file is "max".
cgroup.max.descendants (since Linux 4.14)
This file defines a limit on the number of live descendant
cgroups that this cgroup may have. An attempt to create more
descendants than allowed by the limit fails (mkdir(2) fails with
the error EAGAIN).
Writing the string "max" to this file means that no limit is im-
posed. The default value in this file is "max".
CGROUPS DELEGATION: DELEGATING A HIERARCHY TO A LESS PRIVILEGED USER
In the context of cgroups, delegation means passing management of some
subtree of the cgroup hierarchy to a nonprivileged user. Cgroups v1
provides support for delegation based on file permissions in the cgroup
hierarchy but with less strict containment rules than v2 (as noted be-
low). Cgroups v2 supports delegation with containment by explicit de-
sign. The focus of the discussion in this section is on delegation in
cgroups v2, with some differences for cgroups v1 noted along the way.
Some terminology is required in order to describe delegation. A dele-
gater is a privileged user (i.e., root) who owns a parent cgroup. A
delegatee is a nonprivileged user who will be granted the permissions
needed to manage some subhierarchy under that parent cgroup, known as
the delegated subtree.
To perform delegation, the delegater makes certain directories and
files writable by the delegatee, typically by changing the ownership of
the objects to be the user ID of the delegatee. Assuming that we want
to delegate the hierarchy rooted at (say) /dlgt_grp and that there are
not yet any child cgroups under that cgroup, the ownership of the fol-
lowing is changed to the user ID of the delegatee:
/dlgt_grp
Changing the ownership of the root of the subtree means that any
new cgroups created under the subtree (and the files they con-
tain) will also be owned by the delegatee.
/dlgt_grp/cgroup.procs
Changing the ownership of this file means that the delegatee can
move processes into the root of the delegated subtree.
/dlgt_grp/cgroup.subtree_control (cgroups v2 only)
Changing the ownership of this file means that the delegatee can
enable controllers (that are present in /dlgt_grp/cgroup.con-
trollers) in order to further redistribute resources at lower
levels in the subtree. (As an alternative to changing the own-
ership of this file, the delegater might instead add selected
controllers to this file.)
/dlgt_grp/cgroup.threads (cgroups v2 only)
Changing the ownership of this file is necessary if a threaded
subtree is being delegated (see the description of "thread
mode", below). This permits the delegatee to write thread IDs
to the file. (The ownership of this file can also be changed
when delegating a domain subtree, but currently this serves no
purpose, since, as described below, it is not possible to move a
thread between domain cgroups by writing its thread ID to the
cgroup.threads file.)
In cgroups v1, the corresponding file that should instead be
delegated is the tasks file.
The delegater should not change the ownership of any of the controller
interfaces files (e.g., pids.max, memory.high) in dlgt_grp. Those
files are used from the next level above the delegated subtree in order
to distribute resources into the subtree, and the delegatee should not
have permission to change the resources that are distributed into the
delegated subtree.
See also the discussion of the /sys/kernel/cgroup/delegate file in
NOTES for information about further delegatable files in cgroups v2.
After the aforementioned steps have been performed, the delegatee can
create child cgroups within the delegated subtree (the cgroup subdirec-
tories and the files they contain will be owned by the delegatee) and
move processes between cgroups in the subtree. If some controllers are
present in dlgt_grp/cgroup.subtree_control, or the ownership of that
file was passed to the delegatee, the delegatee can also control the
further redistribution of the corresponding resources into the dele-
gated subtree.
Cgroups v2 delegation: nsdelegate and cgroup namespaces
Starting with Linux 4.13, there is a second way to perform cgroup dele-
gation in the cgroups v2 hierarchy. This is done by mounting or re-
mounting the cgroup v2 filesystem with the nsdelegate mount option.
For example, if the cgroup v2 filesystem has already been mounted, we
can remount it with the nsdelegate option as follows:
mount -t cgroup2 -o remount,nsdelegate \
none /sys/fs/cgroup/unified
The effect of this mount option is to cause cgroup namespaces to auto-
matically become delegation boundaries. More specifically, the follow-
ing restrictions apply for processes inside the cgroup namespace:
* Writes to controller interface files in the root directory of the
namespace will fail with the error EPERM. Processes inside the
cgroup namespace can still write to delegatable files in the root
directory of the cgroup namespace such as cgroup.procs and
cgroup.subtree_control, and can create subhierarchy underneath the
root directory.
* Attempts to migrate processes across the namespace boundary are de-
nied (with the error ENOENT). Processes inside the cgroup namespace
can still (subject to the containment rules described below) move
processes between cgroups within the subhierarchy under the name-
space root.
The ability to define cgroup namespaces as delegation boundaries makes
cgroup namespaces more useful. To understand why, suppose that we al-
ready have one cgroup hierarchy that has been delegated to a nonprivi-
leged user, cecilia, using the older delegation technique described
above. Suppose further that cecilia wanted to further delegate a sub-
hierarchy under the existing delegated hierarchy. (For example, the
delegated hierarchy might be associated with an unprivileged container
run by cecilia.) Even if a cgroup namespace was employed, because both
hierarchies are owned by the unprivileged user cecilia, the following
illegitimate actions could be performed:
* A process in the inferior hierarchy could change the resource con-
troller settings in the root directory of that hierarchy. (These
resource controller settings are intended to allow control to be ex-
ercised from the parent cgroup; a process inside the child cgroup
should not be allowed to modify them.)
* A process inside the inferior hierarchy could move processes into
and out of the inferior hierarchy if the cgroups in the superior hi-
erarchy were somehow visible.
Employing the nsdelegate mount option prevents both of these possibili-
ties.
The nsdelegate mount option only has an effect when performed in the
initial mount namespace; in other mount namespaces, the option is
silently ignored.
Note: On some systems, systemd(1) automatically mounts the cgroup v2
filesystem. In order to experiment with the nsdelegate operation, it
may be useful to boot the kernel with the following command-line op-
tions:
cgroup_no_v1=all systemd.legacy_systemd_cgroup_controller
These options cause the kernel to boot with the cgroups v1 controllers
disabled (meaning that the controllers are available in the v2 hierar-
chy), and tells systemd(1) not to mount and use the cgroup v2 hierar-
chy, so that the v2 hierarchy can be manually mounted with the desired
options after boot-up.
Cgroup delegation containment rules
Some delegation containment rules ensure that the delegatee can move
processes between cgroups within the delegated subtree, but can't move
processes from outside the delegated subtree into the subtree or vice
versa. A nonprivileged process (i.e., the delegatee) can write the PID
of a "target" process into a cgroup.procs file only if all of the fol-
lowing are true:
* The writer has write permission on the cgroup.procs file in the des-
tination cgroup.
* The writer has write permission on the cgroup.procs file in the
nearest common ancestor of the source and destination cgroups. Note
that in some cases, the nearest common ancestor may be the source or
destination cgroup itself. This requirement is not enforced for
cgroups v1 hierarchies, with the consequence that containment in v1
is less strict than in v2. (For example, in cgroups v1 the user
that owns two distinct delegated subhierarchies can move a process
between the hierarchies.)
* If the cgroup v2 filesystem was mounted with the nsdelegate option,
the writer must be able to see the source and destination cgroups
from its cgroup namespace.
* In cgroups v1: the effective UID of the writer (i.e., the delegatee)
matches the real user ID or the saved set-user-ID of the target
process. Before Linux 4.11, this requirement also applied in
cgroups v2 (This was a historical requirement inherited from cgroups
v1 that was later deemed unnecessary, since the other rules suffice
for containment in cgroups v2.)
Note: one consequence of these delegation containment rules is that the
unprivileged delegatee can't place the first process into the delegated
subtree; instead, the delegater must place the first process (a process
owned by the delegatee) into the delegated subtree.
CGROUPS VERSION 2 THREAD MODE
Among the restrictions imposed by cgroups v2 that were not present in
cgroups v1 are the following:
* No thread-granularity control: all of the threads of a process must
be in the same cgroup.
* No internal processes: a cgroup can't both have member processes and
exercise controllers on child cgroups.
Both of these restrictions were added because the lack of these re-
strictions had caused problems in cgroups v1. In particular, the
cgroups v1 ability to allow thread-level granularity for cgroup member-
ship made no sense for some controllers. (A notable example was the
memory controller: since threads share an address space, it made no
sense to split threads across different memory cgroups.)
Notwithstanding the initial design decision in cgroups v2, there were
use cases for certain controllers, notably the cpu controller, for
which thread-level granularity of control was meaningful and useful.
To accommodate such use cases, Linux 4.14 added thread mode for cgroups
v2.
Thread mode allows the following:
* The creation of threaded subtrees in which the threads of a process
may be spread across cgroups inside the tree. (A threaded subtree
may contain multiple multithreaded processes.)
* The concept of threaded controllers, which can distribute resources
across the cgroups in a threaded subtree.
* A relaxation of the "no internal processes rule", so that, within a
threaded subtree, a cgroup can both contain member threads and exer-
cise resource control over child cgroups.
With the addition of thread mode, each nonroot cgroup now contains a
new file, cgroup.type, that exposes, and in some circumstances can be
used to change, the "type" of a cgroup. This file contains one of the
following type values:
domain This is a normal v2 cgroup that provides process-granularity
control. If a process is a member of this cgroup, then all
threads of the process are (by definition) in the same cgroup.
This is the default cgroup type, and provides the same behavior
that was provided for cgroups in the initial cgroups v2 imple-
mentation.
threaded
This cgroup is a member of a threaded subtree. Threads can be
added to this cgroup, and controllers can be enabled for the
cgroup.
domain threaded
This is a domain cgroup that serves as the root of a threaded
subtree. This cgroup type is also known as "threaded root".
domain invalid
This is a cgroup inside a threaded subtree that is in an "in-
valid" state. Processes can't be added to the cgroup, and con-
trollers can't be enabled for the cgroup. The only thing that
can be done with this cgroup (other than deleting it) is to con-
vert it to a threaded cgroup by writing the string "threaded" to
the cgroup.type file.
The rationale for the existence of this "interim" type during
the creation of a threaded subtree (rather than the kernel sim-
ply immediately converting all cgroups under the threaded root
to the type threaded) is to allow for possible future extensions
to the thread mode model
Threaded versus domain controllers
With the addition of threads mode, cgroups v2 now distinguishes two
types of resource controllers:
* Threaded controllers: these controllers support thread-granularity
for resource control and can be enabled inside threaded subtrees,
with the result that the corresponding controller-interface files
appear inside the cgroups in the threaded subtree. As at Linux
4.19, the following controllers are threaded: cpu, perf_event, and
pids.
* Domain controllers: these controllers support only process granular-
ity for resource control. From the perspective of a domain con-
troller, all threads of a process are always in the same cgroup.
Domain controllers can't be enabled inside a threaded subtree.
Creating a threaded subtree
There are two pathways that lead to the creation of a threaded subtree.
The first pathway proceeds as follows:
1. We write the string "threaded" to the cgroup.type file of a cgroup
y/z that currently has the type domain. This has the following ef-
fects:
* The type of the cgroup y/z becomes threaded.
* The type of the parent cgroup, y, becomes domain threaded. The
parent cgroup is the root of a threaded subtree (also known as
the "threaded root").
* All other cgroups under y that were not already of type threaded
(because they were inside already existing threaded subtrees un-
der the new threaded root) are converted to type domain invalid.
Any subsequently created cgroups under y will also have the type
domain invalid.
2. We write the string "threaded" to each of the domain invalid cgroups
under y, in order to convert them to the type threaded. As a conse-
quence of this step, all threads under the threaded root now have
the type threaded and the threaded subtree is now fully usable. The
requirement to write "threaded" to each of these cgroups is somewhat
cumbersome, but allows for possible future extensions to the thread-
mode model.
The second way of creating a threaded subtree is as follows:
1. In an existing cgroup, z, that currently has the type domain, we (1)
enable one or more threaded controllers and (2) make a process a
member of z. (These two steps can be done in either order.) This
has the following consequences:
* The type of z becomes domain threaded.
* All of the descendant cgroups of x that were not already of type
threaded are converted to type domain invalid.
2. As before, we make the threaded subtree usable by writing the string
"threaded" to each of the domain invalid cgroups under y, in order
to convert them to the type threaded.
One of the consequences of the above pathways to creating a threaded
subtree is that the threaded root cgroup can be a parent only to
threaded (and domain invalid) cgroups. The threaded root cgroup can't
be a parent of a domain cgroups, and a threaded cgroup can't have a
sibling that is a domain cgroup.
Using a threaded subtree
Within a threaded subtree, threaded controllers can be enabled in each
subgroup whose type has been changed to threaded; upon doing so, the
corresponding controller interface files appear in the children of that
cgroup.
A process can be moved into a threaded subtree by writing its PID to
the cgroup.procs file in one of the cgroups inside the tree. This has
the effect of making all of the threads in the process members of the
corresponding cgroup and makes the process a member of the threaded
subtree. The threads of the process can then be spread across the
threaded subtree by writing their thread IDs (see gettid(2)) to the
cgroup.threads files in different cgroups inside the subtree. The
threads of a process must all reside in the same threaded subtree.
As with writing to cgroup.procs, some containment rules apply when
writing to the cgroup.threads file:
* The writer must have write permission on the cgroup.threads file in
the destination cgroup.
* The writer must have write permission on the cgroup.procs file in
the common ancestor of the source and destination cgroups. (In some
cases, the common ancestor may be the source or destination cgroup
itself.)
* The source and destination cgroups must be in the same threaded sub-
tree. (Outside a threaded subtree, an attempt to move a thread by
writing its thread ID to the cgroup.threads file in a different do-
main cgroup fails with the error EOPNOTSUPP.)
The cgroup.threads file is present in each cgroup (including domain
cgroups) and can be read in order to discover the set of threads that
is present in the cgroup. The set of thread IDs obtained when reading
this file is not guaranteed to be ordered or free of duplicates.
The cgroup.procs file in the threaded root shows the PIDs of all pro-
cesses that are members of the threaded subtree. The cgroup.procs
files in the other cgroups in the subtree are not readable.
Domain controllers can't be enabled in a threaded subtree; no con-
troller-interface files appear inside the cgroups underneath the
threaded root. From the point of view of a domain controller, threaded
subtrees are invisible: a multithreaded process inside a threaded sub-
tree appears to a domain controller as a process that resides in the
threaded root cgroup.
Within a threaded subtree, the "no internal processes" rule does not
apply: a cgroup can both contain member processes (or thread) and exer-
cise controllers on child cgroups.
Rules for writing to cgroup.type and creating threaded subtrees
A number of rules apply when writing to the cgroup.type file:
* Only the string "threaded" may be written. In other words, the only
explicit transition that is possible is to convert a domain cgroup
to type threaded.
* The effect of writing "threaded" depends on the current value in
cgroup.type, as follows:
o domain or domain threaded: start the creation of a threaded sub-
tree (whose root is the parent of this cgroup) via the first of
the pathways described above;
o domain invalid: convert this cgroup (which is inside a threaded
subtree) to a usable (i.e., threaded) state;
o threaded: no effect (a "no-op").
* We can't write to a cgroup.type file if the parent's type is domain
invalid. In other words, the cgroups of a threaded subtree must be
converted to the threaded state in a top-down manner.
There are also some constraints that must be satisfied in order to cre-
ate a threaded subtree rooted at the cgroup x:
* There can be no member processes in the descendant cgroups of x.
(The cgroup x can itself have member processes.)
* No domain controllers may be enabled in x's cgroup.subtree_control
file.
If any of the above constraints is violated, then an attempt to write
"threaded" to a cgroup.type file fails with the error ENOTSUP.
The "domain threaded" cgroup type
According to the pathways described above, the type of a cgroup can
change to domain threaded in either of the following cases:
* The string "threaded" is written to a child cgroup.
* A threaded controller is enabled inside the cgroup and a process is
made a member of the cgroup.
A domain threaded cgroup, x, can revert to the type domain if the above
conditions no longer hold true--that is, if all threaded child cgroups
of x are removed and either x no longer has threaded controllers en-
abled or no longer has member processes.
When a domain threaded cgroup x reverts to the type domain:
* All domain invalid descendants of x that are not in lower-level
threaded subtrees revert to the type domain.
* The root cgroups in any lower-level threaded subtrees revert to the
type domain threaded.
Exceptions for the root cgroup
The root cgroup of the v2 hierarchy is treated exceptionally: it can be
the parent of both domain and threaded cgroups. If the string
"threaded" is written to the cgroup.type file of one of the children of
the root cgroup, then
* The type of that cgroup becomes threaded.
* The type of any descendants of that cgroup that are not part of
lower-level threaded subtrees changes to domain invalid.
Note that in this case, there is no cgroup whose type becomes domain
threaded. (Notionally, the root cgroup can be considered as the
threaded root for the cgroup whose type was changed to threaded.)
The aim of this exceptional treatment for the root cgroup is to allow a
threaded cgroup that employs the cpu controller to be placed as high as
possible in the hierarchy, so as to minimize the (small) cost of
traversing the cgroup hierarchy.
The cgroups v2 "cpu" controller and realtime threads
As at Linux 4.19, the cgroups v2 cpu controller does not support con-
trol of realtime threads (specifically threads scheduled under any of
the policies SCHED_FIFO, SCHED_RR, described SCHED_DEADLINE; see
sched(7)). Therefore, the cpu controller can be enabled in the root
cgroup only if all realtime threads are in the root cgroup. (If there
are realtime threads in nonroot cgroups, then a write(2) of the string
"+cpu" to the cgroup.subtree_control file fails with the error EINVAL.)
On some systems, systemd(1) places certain realtime threads in nonroot
cgroups in the v2 hierarchy. On such systems, these threads must first
be moved to the root cgroup before the cpu controller can be enabled.
ERRORS
The following errors can occur for mount(2):
EBUSY An attempt to mount a cgroup version 1 filesystem specified nei-
ther the name= option (to mount a named hierarchy) nor a con-
troller name (or all).
NOTES
A child process created via fork(2) inherits its parent's cgroup mem-
berships. A process's cgroup memberships are preserved across ex-
ecve(2).
The clone3(2) CLONE_INTO_CGROUP flag can be used to create a child
process that begins its life in a different version 2 cgroup from the
parent process.
/proc files
/proc/cgroups (since Linux 2.6.24)
This file contains information about the controllers that are
compiled into the kernel. An example of the contents of this
file (reformatted for readability) is the following:
#subsys_name hierarchy num_cgroups enabled
cpuset 4 1 1
cpu 8 1 1
cpuacct 8 1 1
blkio 6 1 1
memory 3 1 1
devices 10 84 1
freezer 7 1 1
net_cls 9 1 1
perf_event 5 1 1
net_prio 9 1 1
hugetlb 0 1 0
pids 2 1 1
The fields in this file are, from left to right:
1. The name of the controller.
2. The unique ID of the cgroup hierarchy on which this con-
troller is mounted. If multiple cgroups v1 controllers are
bound to the same hierarchy, then each will show the same hi-
erarchy ID in this field. The value in this field will be 0
if:
a) the controller is not mounted on a cgroups v1 hierarchy;
b) the controller is bound to the cgroups v2 single unified
hierarchy; or
c) the controller is disabled (see below).
3. The number of control groups in this hierarchy using this
controller.
4. This field contains the value 1 if this controller is en-
abled, or 0 if it has been disabled (via the cgroup_disable
kernel command-line boot parameter).
/proc/[pid]/cgroup (since Linux 2.6.24)
This file describes control groups to which the process with the
corresponding PID belongs. The displayed information differs
for cgroups version 1 and version 2 hierarchies.
For each cgroup hierarchy of which the process is a member,
there is one entry containing three colon-separated fields:
hierarchy-ID:controller-list:cgroup-path
For example:
5:cpuacct,cpu,cpuset:/daemons
The colon-separated fields are, from left to right:
1. For cgroups version 1 hierarchies, this field contains a
unique hierarchy ID number that can be matched to a hierarchy
ID in /proc/cgroups. For the cgroups version 2 hierarchy,
this field contains the value 0.
2. For cgroups version 1 hierarchies, this field contains a
comma-separated list of the controllers bound to the hierar-
chy. For the cgroups version 2 hierarchy, this field is
empty.
3. This field contains the pathname of the control group in the
hierarchy to which the process belongs. This pathname is
relative to the mount point of the hierarchy.
/sys/kernel/cgroup files
/sys/kernel/cgroup/delegate (since Linux 4.15)
This file exports a list of the cgroups v2 files (one per line)
that are delegatable (i.e., whose ownership should be changed to
the user ID of the delegatee). In the future, the set of dele-
gatable files may change or grow, and this file provides a way
for the kernel to inform user-space applications of which files
must be delegated. As at Linux 4.15, one sees the following
when inspecting this file:
$ cat /sys/kernel/cgroup/delegate
cgroup.procs
cgroup.subtree_control
cgroup.threads
/sys/kernel/cgroup/features (since Linux 4.15)
Over time, the set of cgroups v2 features that are provided by
the kernel may change or grow, or some features may not be en-
abled by default. This file provides a way for user-space ap-
plications to discover what features the running kernel supports
and has enabled. Features are listed one per line:
$ cat /sys/kernel/cgroup/features
nsdelegate
memory_localevents
The entries that can appear in this file are:
memory_localevents (since Linux 5.2)
The kernel supports the memory_localevents mount option.
nsdelegate (since Linux 4.15)
The kernel supports the nsdelegate mount option.
SEE ALSO
prlimit(1), systemd(1), systemd-cgls(1), systemd-cgtop(1), clone(2),
ioprio_set(2), perf_event_open(2), setrlimit(2), cgroup_namespaces(7),
cpuset(7), namespaces(7), sched(7), user_namespaces(7)
The kernel source file Documentation/admin-guide/cgroup-v2.rst.
COLOPHON
This page is part of release 5.07 of the Linux man-pages project. A
description of the project, information about reporting bugs, and the
latest version of this page, can be found at
https://www.kernel.org/doc/man-pages/.
Linux 2020-04-11 CGROUPS(7)