Okay, so now we've hidden the processes from Eve so Bob and Alice can engage in commerce in privacy and peace. So we're all good, right? They can no longer mess each other, right? Not quite. We're almost there.
So now say it's Black Friday, Boxing Day or Singles' Day (three of the biggest shopping days in the year, pick the one that makes the most sense to you 😄) and Bob and Alice are gearing up for their biggest sales day of the year. Everything is ready to go and at 9:00AM their site suddenly goes down without warning. What happened!? They log on to their chroot'd, unshare'd shell on your server and see that the CPU is pegged at 100% and there's no more memory available to allocate! Oh no! What happened?
The first explanation could be that Eve has her site running on another server and simple logged on and ran a program that ate up all the available resources so that Bob and Alice so that their sites would go down and Eve would be the only site that was up, increasing her sales.
However another, possibly more likely explanation is that both Bob's and Alice's sites got busy at the same time and that in-and-of-itself took all the resources without any malice involved, taking down their sites and everyone else on the server. Or perhaps Bob's site had a memory leak and that was enough to take all the resources available.
Suffice to say, we still have a problem. Every isolated environment has access to all physical resources of the server. There's no isolation of physical components from these environments.
Enter the hero of this story: cgroups, or control groups. Google saw this same problem when building their own infrastructure and wanted to protect runaway processes from taking down entire servers and made this idea of cgroups so you can say "this isolated environment only gets so much CPU, so much memory, etc. and once it's out of those it's out-of-luck, it won't get any more."
This is a bit more difficult to accomplish but let's go ahead and give it a shot.
# outside of unshare'd environment get the tools we'll need here apt-get install -y cgroup-tools htop # create new cgroups cgcreate -g cpu,memory,blkio,devices,freezer:/sandbox # add our unshare'd env to our cgroup ps aux # grab the bash PID that's right after the unshare one cgclassify -g cpu,memory,blkio,devices,freezer:sandbox <PID> # list tasks associated to the sandbox cpu group, we should see the above PID cat /sys/fs/cgroup/cpu/sandbox/tasks # show the cpu share of the sandbox cpu group, this is the number that determines priority between competing resources, higher is is higher priority cat /sys/fs/cgroup/cpu/sandbox/cpu.shares # kill all of sandbox's processes if you need it # kill -9 $(cat /sys/fs/cgroup/cpu/sandbox/tasks) # Limit usage at 5% for a multi core system cgset -r cpu.cfs_period_us=100000 -r cpu.cfs_quota_us=$[ 5000 * $(getconf _NPROCESSORS_ONLN) ] sandbox # Set a limit of 80M cgset -r memory.limit_in_bytes=80M sandbox # Get memory stats used by the cgroup cgget -r memory.stat sandbox # in terminal session #2, outside of the unshare'd env htop # will allow us to see resources being used with a nice visualizer # in terminal session #1, inside unshared'd env yes > /dev/null # this will instantly consume one core's worth of CPU power # notice it's only taking 5% of the CPU, like we set # if you want, run the docker exec from above to get a third session to see the above command take 100% of the available resources # CTRL+C stops the above any time # in terminal session #1, inside unshare'd env yes | tr \\n x | head -c 1048576000 | grep n # this will ramp up to consume ~1GB of RAM # notice in htop it'll keep the memory closer to 80MB due to our cgroup # as above, connect with a third terminal to see it work outside of a cgroup
And now we can call this a container. Using these features together, we allow Bob, Alice, and Eve to run whatever code they want and the only people they can mess with is themselves.
So while this is a container at its most basic sense, we haven't broached more advance topics like networking, deploying, bundling, or anything else that something like Docker takes care of for us. But now you know at its most base level what a container is, what it does, and how you could do this yourself but you'll be grateful that Docker does it for you. On to the next lesson!