-
Notifications
You must be signed in to change notification settings - Fork 122
/
container-security.html.md.erb
306 lines (203 loc) · 18.4 KB
/
container-security.html.md.erb
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
---
title: Container security
owner: Security
---
This topic tells you about how <%= vars.app_runtime_first %> secures the containers
that host app instances on Linux.
For an overview of other <%= vars.app_runtime_abbr %> security features, see [<%= vars.app_runtime_abbr %> Security](security.html).
The sections in this topic provide the following information:
* [Container Mechanics](#mechanics) provides an overview of container isolation.
* [Inbound and Outbound Traffic from <%= vars.app_runtime_abbr %>](#networking) provides an overview of container networking and describes how <%= vars.app_runtime_abbr %> admins customize container network traffic rules for their deployment.
* [Container Security](#security) describes how <%= vars.app_runtime_abbr %> secures containers by running app instances in unprivileged containers and by hardening them.
## <a id='mechanics'></a> Container mechanics
Each instance of an app deployed to <%= vars.app_runtime_abbr %> runs within its own self-contained environment, a Garden container. This container isolates processes, memory, and the filesystem using operating system features and the characteristics of the virtual and physical infrastructure where <%= vars.app_runtime_abbr %> is deployed. For more information about Garden containers, see [Garden](./architecture/garden.html).
<%= vars.app_runtime_abbr %> achieves container isolation by namespacing kernel resources that would otherwise be shared. The intended level of isolation is set to prevent multiple containers that are present on the same host from detecting each other. Every container includes a private root filesystem, which includes a Process ID (PID), namespace, network namespace, and mount namespace.
<%= vars.app_runtime_abbr %> creates container filesystems using the Garden Rootfs (GrootFS) tool. It stacks the following using OverlayFS:
* A **read-only base filesystem**: This filesystem has the minimal set of operating system packages and Garden-specific modifications common to all containers. Containers can share the same read-only base filesystem because all writes are applied to the read-write layer.
* A **container-specific read-write layer**: This layer is unique to each container and its size is limited by XFS project quotas. The quotas prevent the read-write layer from overflowing into unallocated space.
For more information about GrootFS, see [Garden RootFS (GrootFS)](./architecture/garden.html#garden-rootfs) in _Garden_.
Resource control is managed using Linux control groups. Associating each container with its own cgroup or job object limits the amount of memory that the container might use. Linux cgroups also require the container to use a fair share of CPU compared to the relative CPU share of other containers.
<p> <%= vars.app_runtime_abbr %> does not support a RedHat Enterprise Linux OS stemcell. This is due to an
inherent security issue with the way RedHat handles user namespacing and container isolation.</p>
### <a id='cpu'></a> CPU
Each container is placed in its own cgroup. Cgroups make each container use a fair share of CPU relative to the other containers. This prevents oversubscription on the host VM where one or more containers hog the CPU and leave no computing resources to the others.
The way cgroups allocate CPU time is based on shares. CPU shares do not work as direct percentages of total CPU usage. Instead, a share is relative in a given time window to the shares held by the other containers. The total amount of CPU that can be overall divided among the cgroups is what is left by other processes that might run in the host VM.
Generally, cgroups offers two possibilities for limiting the CPU usage: _CPU affinity_ and _CPU bandwidth_, the latter of which is used in <%= vars.app_runtime_abbr %>.
* CPU affinity consists of binding a cgroup to certain CPU cores. The actual amount of CPU cycles that can be consumed by the cgroup is thus limited to what is available on the bound CPU cores.
* CPU bandwidth sets the weight of a cgroup with the process scheduler. The process scheduler divides the available CPU cycles among cgroups depending on the shares held by each cgroup, relative to the shares held by the others.
For example, consider two cgroups, one holding two shares and one holding four. Assuming the process scheduler gets to administer 60 CPU cycles, the first cgroup with two shares will get one third of those available CPU cycles, as it holds one third of the overall shares. Similarly, the second cgroup will get 40 cycles, as it holds two thirds of the collective shares.
The calculation of the CPU usage based on the percentage of the total CPU power available is quite sophisticated and is performed regularly as the CPU demands of the various containers fluctuates. Specifically, the percentage of CPU cycles a cgroup gets can be calculated by dividing the `cpu.shares` it holds by the sum of the `cpu.shares` of all the cgroups that are currently doing CPU activity, as shown in the following calculation:
```
process_cpu_share_percent = cpu.shares / sum_cpu_shares * 100
```
<% if vars.platform_code == 'CF' %>
In <%= vars.app_runtime_abbr %>, cgroup shares range from 10 to 1024. The actual number of shares a cgroup gets can be read from the `cpu.shares` file of the cgroup configurations pseudo-file-system available in the container at `/sys/fs/cgroup/cpu`. The number of shares given to an apps cgroup depends on the amount of memory the app declares to need in the manifest.
#### CPU Shares Algorithm 1
When the [Diego `containers.set_cpu_weight` bosh property](https://github.com/cloudfoundry/diego-release/blob/d051f8f94f652369d02db4cf66712645600d5b4a/jobs/rep/spec#L311-L313) is set to false, then <%= vars.app_runtime_abbr %> scales the number of allocated shares linearly with the amount of memory, until the upper limit of 1024 shares (by default).
<p>In this case, there are a handful of variables that play into the CPU shares calculation.</p>
| variable name | default value | configurable? | link to property |
| ------------- | ------------- | ------------- | ---------------- |
| cc.cpu_weight_min_memory | 128 | yes | [bosh property](https://github.com/cloudfoundry/capi-release/blob/e68034308a57296ac12debf05de012c2a14d4c4c/jobs/cloud_controller_ng/spec#L782-L784) |
| cc.cpu_weight_max_memory | 8192 | yes | [bosh property](https://github.com/cloudfoundry/capi-release/blob/e68034308a57296ac12debf05de012c2a14d4c4c/jobs/cloud_controller_ng/spec#L785-L787) |
| app_memory_in_mb | 1024 | yes | this is set in the app manifest |
| base_weight | 8192 | no | n/a |
| diego.executor.container_max_cpu_shares | 1024 | yes | [bosh property](https://github.com/cloudfoundry/diego-release/blob/d051f8f94f652369d02db4cf66712645600d5b4a/jobs/rep/spec#L122-L124C14)|
These variables are used in the following algorithm to determine the number of CPU shares for an app:
```
app_memory_for_calculation = min( max(cc.cpu_weight_min_memory, app_memory_in_mb), cc.cpu_weight_max_memory)
app_cpu_weight = int( (100 * app_memory_for_calculation) / base_weight)
process_cpu.shares = int( (diego.executor.container_max_cpu_shares * app_cpu_weight) / 100)
```
For example:
```
cc.cpu_weight_min_memory = 128
cc.cpu_weight_max_memory = 8192
app_memory_in_mb = 32
base_weight = 8192
diego.executor.container_max_cpu_shares = 1024
app_memory_for_calculation = min(max(cc.cpu_weight_min_memory, app_memory_in_mb)
app_memory_for_calculation = min(max(128, 32), 8192) = min(128, 8192) = 128
app_cpu_weight = int((100 * app_memory_for_calculation)/base_weight))
app_cpu_weight = int((100 * 128) / 8192) = int(12800 / 8192) = int(1.56) = 1
process_cpu.shares = int( (diego.executor.container_max_cpu_shares * app_cpu_weight) / 100)
process_cpu.shares = int( (1024 * 1) / 100) = int(10.24) = 10
```
#### CPU Shares Algorithm 2
When the [Diego containers.set_cpu_weight BOSH
property](https://github.com/cloudfoundry/diego-release/blob/d051f8f94f652369d02db4cf66712645600d5b4a/jobs/rep/spec#L311-L313)
is set to true, then <%= vars.app_runtime_abbr %> scales the number of
allocated shares linearly with the amount of memory
<p>In this case, there are two variables that play into the CPU shares calculation.</p>
| variable name | default value | configurable? | link to property |
| ------------- | ------------- | ------------- | ---------------- |
| app_memory_in_mb | 1024 | yes | this is set in the app manifest |
| containers.proxy.additional_memory_allocation_mb | 32 | yes | [bosh property](https://github.com/cloudfoundry/diego-release/blob/3496e1be97844d9813a52665239e14ae89c5db3b/jobs/rep/spec#L250-L252) |
In this case, the following algorithm is used to determine the number of CPU shares for an app:
```
process_cpu.shares = app_memory_in_mb + containers.proxy.additional_memory_allocation_mb
```
<% else %>
The actual number of shares a cgroup gets can be read from the `cpu.shares` file of the cgroup configurations pseudo-file-system available in the container at `/sys/fs/cgroup/cpu`.
The number of shares given to an apps cgroup depends on the amount of memory the app declares to need in the manifest.
<%= vars.app_runtime_abbr %> scales the number of allocated shares linearly with the amount of memory.
The following algorithm is used to determine the number of CPU shares for an app:
```
process_cpu.shares = app_memory_in_mb + 32
```
The extra `32` in the algorithm is to account for the extra memory required for the envoy sidecar proxy.
<% end %>
<% if vars.platform_code == 'CF' %>
#### CPU Share Interaction
<% end %>
The next example helps to illustrate this better. Consider three processes: P1, P2 and P3, which are assigned `cpu.shares` of 5, 20 and 30, respectively.
P1 is active, while P2 and P3 require no CPU. Hence, P1 might use the whole CPU. When P2 joins in and is doing some actual work, such as when a request comes in, the CPU share between P1 and P2 is calculated as follows:
* P1 -> 5 / (5+20) = 0.2 = 20%
* P2 -> 20 / (5+20) = 0.8 = 80%
* P3 (idle)
At some point, process P3 joins in. Then the distribution is recalculated again:
* P1 -> 5 / (5+20+30) = 0.0909 = ~9%
* P2 -> 20 / (5+20+30) = 0.363 = ~36%
* P3 -> 30 / (5+20+30) = 0.545 = ~55%
If P1 become idle, the following recalculation between P2 and P3 takes place:
* P1 (idle)
* P2 -> 20 / (20+30) = 0.4 = 40%
* P3 -> 30 / (20+30) = 0.6 = 60%
If P3 finishes or becomes idle, then P2 can consume all the CPU, as another recalculation is performed.
In summary, the cgroup shares are the minimum guaranteed CPU share that the process can get. This limitation becomes effective only when processes on the same host compete for resources.
## <a id='networking'></a> Inbound and outbound traffic from <%= vars.app_runtime_abbr %>
Learn about container networking and how <%= vars.app_runtime_abbr %> admins customize
container network traffic rules for their deployment.
### <a id='net-overview'></a> Networking overview
A host VM has a single IP address. If you configure the deployment with the cluster on a VLAN, as <%= vars.company_name %> recommends, then all traffic goes through the following levels of network address translation, as shown in the following diagram:
* **Inbound** requests flow from the load balancer through the router to the host Diego Cell, then into the app container. The router determines which app instance receives each request.
* **Outbound** traffic flows from the app container to the Diego Cell, then to the gateway on the Diego Cell's virtual network interface. Depending on your IaaS, this gateway might be a NAT to external networks.
![Networking Architecture diagram](images/security/sysbound2.png)
The networking diagram shows the following:
- Inbound requests go to Load Balancer
- Outbound traffic comes from NAT
Cloud Foundry vLAN (on right side):
- Load balancer goes to router and to the app container inside the Diego Cell
- App container response goes to NAT
### <a id='config-traffic'></a> Network traffic rules
Admins configure rules to govern container network traffic. This is how containers send traffic outside of <%= vars.app_runtime_abbr %> and receive traffic from outside, the Internet. These rules can prevent system access from external networks and between internal components and determine if apps can establish connections over the virtual network interface.
Admins configure these rules at two levels:
* Application Security Groups (ASGs) apply network traffic rules at the container level. <%= vars.app_sec_groups_link %>
* Container-to-container networking policies determine app-to-app communication. Within <%= vars.app_runtime_abbr %>, apps can communicate directly with each other, but the containers are isolated from outside <%= vars.app_runtime_abbr %>. <%= vars.container_network_link %>
## <a id='security'></a> Container security
<%= vars.app_runtime_abbr %> secures containers through the following measures:
* Running app instances in unprivileged containers by default. For more information about app instances, see [Types](#types).
* Hardening containers by limiting function and access rights. For more information about hardening containers, see [Hardening](#hardening).
* <%= vars.app_sec_groups_default %> <%= vars.app_sec_groups_link %>
### <a id="types"></a> Types
Garden has the following container types:
* unprivileged
* privileged
Currently, <%= vars.app_runtime_abbr %> runs all app instances and staging tasks in unprivileged containers by default. This measure increases security by eliminating the threat of root escalation inside the container.
<% if vars.platform_code == 'CF' %>
<%= partial 'config_manifest' %>
<% end %>
### <a id='hardening'></a> Hardening
<%= vars.app_runtime_abbr %> mitigates against container breakout and denial of service attacks in
the following ways:
* <%= vars.app_runtime_abbr %> uses the full set of Linux namespaces - IPC, Network, Mount, PID, User, UTS - to provide isolation between containers running on the same host. The User namespace is not used for privileged containers. For more information about Linux namespaces, see [namespaces - overview of Linux namespaces](http://manpages.ubuntu.com/manpages/xenial/en/man7/namespaces.7.html) in the Ubuntu documentation.
* In unprivileged containers, <%= vars.platform_name %> maps UID/GID 0 (root) inside the container user namespace to a different UID/GID on the host to prevent an app from inheriting UID/GID 0 on the host if it breaks out of the container.
* <%= vars.app_runtime_abbr %> uses the same UID/GID for all containers.
* <%= vars.app_runtime_abbr %> maps all UIDs except UID 0 to themselves. <%= vars.app_runtime_abbr %> maps UID 0 inside the container namespace to `MAX_UID-1` outside of the container namespace.
* Container Root does not grant Host Root permissions.
* <%= vars.app_runtime_abbr %> mounts `/proc` and `/sys` as read-only inside containers.
* <%= vars.app_runtime_abbr %> disallows `dmesg` access for unprivileged users and all users in unprivileged containers.
* <%= vars.app_runtime_abbr %> uses `chroot` when importing docker images from docker registries.
* <%= vars.app_runtime_abbr %> establishes a container-specific overlay filesystem mount. <%= vars.app_runtime_abbr %> uses `pivot_root` to move the root filesystem into this overlay, in order to isolate the container from the host system's filesystem. For more information about `pivot_root`, see [pivot_root - change the root filesystem](http://manpages.ubuntu.com/manpages/trusty/man2/pivot_root.2.html) in the Ubuntu documentation.
* <%= vars.app_runtime_abbr %> does not call any binary or script inside the container filesystem, in order to eliminate any dependencies on scripts and binaries inside the root filesystem.
* <%= vars.app_runtime_abbr %> avoids side-loading binaries in the container through bind mounts or other methods. Instead, it runs the same binary again by reading it from `/proc/self/exe` whenever it needs to run a binary in a container.
* <%= vars.app_runtime_abbr %> establishes a virtual ethernet pair for each container for network traffic. For more information, see [Container Network Traffic](#config-traffic). The virtual ethernet pair has the following features:
* One interface in the pair is inside the container's network namespace, and is the only non-loopback interface accessible inside the container.
* The other interface remains in the host network namespace and is bridged to the container-side interface.
* Egress allow list rules are applied to these interfaces according to ASGs configured by the admin.
* First-packet logging rules might also be activated on the TCP allow list rules.
* DNAT rules are established on the host to enable traffic ingress from the host interface to allowed ports on the container-side interface.
* <%= vars.app_runtime_abbr %> applies disk quotas using container-specific XFS quotas with the specified disk-quota capacity.
* <%= vars.app_runtime_abbr %> applies a total memory usage quota through the memory cgroup and destroys the container if the memory usage exceeds the quota.
* <%= vars.app_runtime_abbr %> applies a fair-use limit to CPU usage for processes inside the container through the `cpu.shares` control group.
* <%= vars.app_runtime_abbr %> allows admins to rate limit the maximum bandwidth consumed by single-app containers, configuring `rate` and `burst` properties on the `silk-cni` job.
* <%= vars.app_runtime_abbr %> limits access to devices using cgroups but explicitly allows the following safe device nodes:
* `/dev/full`
* `/dev/fuse`
* `/dev/null`
* `/dev/ptmx`
* `/dev/pts/*`
* `/dev/random`
* `/dev/tty`
* `/dev/tty0`
* `/dev/tty1`
* `/dev/urandom`
* `/dev/zero`
* `/dev/tap`
* `/dev/tun`
* <%= vars.app_runtime_abbr %> drops the following Linux capabilities for all container processes. Every dropped capability limits the actions the root user can perform.
* `CAP_DAC_READ_SEARCH`
* `CAP_LINUX_IMMUTABLE`
* `CAP_NET_BROADCAST`
* `CAP_NET_ADMIN`
* `CAP_IPC_LOCK`
* `CAP_IPC_OWNER`
* `CAP_SYS_MODULE`
* `CAP_SYS_RAWIO`
* `CAP_SYS_PTRACE`
* `CAP_SYS_PACCT`
* `CAP_SYS_BOOT`
* `CAP_SYS_NICE`
* `CAP_SYS_RESOURCE`
* `CAP_SYS_TIME`
* `CAP_SYS_TTY_CONFIG`
* `CAP_LEASE`
* `CAP_AUDIT_CONTROL`
* `CAP_MAC_OVERRIDE`
* `CAP_MAC_ADMIN`
* `CAP_SYSLOG`
* `CAP_WAKE_ALARM`
* `CAP_BLOCK_SUSPEND`
* `CAP_SYS_ADMIN` for unprivileged containers
<br>
<br>
For more information about Linux capabilities, see [capabilities - overview of Linux capabilities](http://manpages.ubuntu.com/manpages/xenial/en/man7/capabilities.7.html) in the Ubuntu documentation.