TipsAndTricks/BuildingAndRunningClearContainersonTarget
What & Why
Clear containers (CC) offer a hybrid solution that encompasses the advantages of hypervisor security and container deployment. So, we wanted to see if they could be used in a YP environment. This was done for Clear Containers 2.2 based on YP master around the time of 2.4 RC1/2.
Note: this is a Proof of Concept, done by building on target. The eventual goal would be to create a standard recipe to allow the clear containers to be built in the standard way. Hopefully, this guide will help with that by outlining the parts, dependencies, and configuration steps. This guide assumes you already have docker running on your target by having followed Running Docker on your image . The target example is being done with an Intel Nuc. I have successfully run the same code on a Minnowboard Turbot.
Dependencies you need
Layers
The layers I am using:
meta-openembedded/meta-oe meta-openembedded/meta-python meta-openembedded/meta-networking meta-openembedded/meta-filesystems meta-virtualization meta-clear
All of these layers can be found on layer.openembedded.org except the meta-clear. The meta-clear layer was created with the script yocto-layer. It's only purpose is to turn on CONFIG_VHOST_NET=m for the kernel. Here's a tree of the layer:
├── conf │ └── layer.conf ├── COPYING.MIT ├── README └── recipes-kernel └── linux ├── linux-yocto │ ├── clear.cfg │ └── clear.scc ├── linux-yocto_4.10.bbappend └── linux-yocto_4.9.bbappend
I am using the 4.9 kernel. Here's the linux-yocto_4.9.bbappend:
FILESEXTRAPATHS_prepend := "${THISDIR}/${PN}:" SRC_URI += "file://clear.scc \ " KERNEL_MODULE_AUTOLOAD += "vhost-net"
The scc file:
define KFEATURE_DESCRIPTION "Enable clearcon support" define KFEATURE_COMPATIBILITY board kconf non-hardware clear.cfg
And finally the cfg file:
CONFIG_VHOST_NET=m
Conf Changes
This guide presumes you have the setup in your conf file described in Running Docker on your image . In addition, to make on target building easier, I add the following to my conf/local.conf:
EXTRA_IMAGE_FEATURES += " dev-pkgs tools-sdk tools-debug tools-profile "
Additional Dependencies to Bitbake
These are the additional recipes I built in addition to the base I outlined above. They could be added all at once in the local.conf, if you want.
bitbake libcheck mdadm psmisc json-glib libmnl ossp-uuid autoconf-archive python-setuptools libcap-ng tunctl go wget sudo
These are additional packages I built for convenience, but they are not required:
bitbake less zile ntp rsync minicom
Once built these can be installed on the board. Note that we need the dev pkgs as we are mostly completing build requirements for pieces of CC.
dnf install tunctl python-dev python-setuptools-dev libcap-ng-dev libcheck-dev libmnl-dev libjson-glib-1.0-dev autoconf-archive-dev libcap-ng-dev python-setuptools-dev go wget sudo
and the convenient ones:
dnf install zile less ntp rsync minicom
The Image to Build
bitbake core-image--base
The Pieces of CC
Clear Containers are comprised of a set of software and binaries. The main code is a slightly forked (2.9 currently) qemu hypervisor configured to be minimal, a command proxy, a shim, and the oci runtime. The command proxy is written in go. The rest is c/c++. We build the hypervisor itself, but the binaries for the hypervisor are downloaded from the CC site.
The Runtime,Shim & Proxy
This comes from [clear oci runtime]. While getting it to work, I followed the development model outlined in Leveraging Rpm Package Feeds. Here I will list the dependencies to make it shorter.
Which Clear was this?
cc-oci-runtime version: 2.2.0 spec version: 1.0.0-rc1 commit: f92d50ad54003298c139de59777f07588683cdc2
Getting the Source Code
We will pretty much follow the (very good) instructions in the README. Because it is a go project we will follow the go flow...
go get -d github.com/01org/cc-oci-runtime/...
The ... is necessary. If you are behind a proxy, make sure you export http_proxy and https_proxy into your shell.
This will put the src in ~/go/src/github.com/01org/cc-oci-runtime/ by default.
Building It
Again, following their README do
./autogen.sh --disable-functional-tests
We disable functional tests because I was not able to easily find a recipe for BATS (Bash automated testing).
make
The make install is quite clean and show where everything is going:
# make install make[1]: Entering directory '/home/root/go/src/github.com/01org/cc-oci-runtime' /bin/mkdir -p '/usr/bin' /bin/sh ./libtool --mode=install /usr/bin/install -c cc-oci-runtime '/usr/bin' libtool: install: /usr/bin/install -c cc-oci-runtime /usr/bin/cc-oci-runtime /bin/mkdir -p '/usr/bin' /usr/bin/install -c data/cc-oci-runtime.sh '/usr/bin' /bin/mkdir -p '/usr/libexec' /bin/sh ./libtool --mode=install /usr/bin/install -c cc-shim '/usr/libexec' libtool: install: /usr/bin/install -c cc-shim /usr/libexec/cc-shim /bin/mkdir -p '/usr/libexec' /usr/bin/install -c cc-proxy '/usr/libexec' /bin/mkdir -p '/usr/share/defaults/cc-oci-runtime' /usr/bin/install -c -m 644 data/vm.json data/hypervisor.args data/kernel-cmdline '/usr/share/defaults/cc-oci-runtime' /bin/mkdir -p '/lib/systemd/system' /usr/bin/install -c -m 644 proxy/cc-proxy.service proxy/cc-proxy.socket '/lib/systemd/system'
Getting the Artifacts
We need a kernel and an image for the hypervisor.
- The kernel comes from [Clear Linux Releases.] For example, I used release 16050 ~ June 2017. This can be found in the following rpm [kernel rpm 16050]
- The image (rootfs) also comes from [Clear Linux Releases.] I again picked the [16050 release image ]
- Here's the commands to download the above and install it.
wget http://download.clearlinux.org/releases/16050/clear/x86_64/os/Packages/linux-container-4.9.33-74.x86_64.rpm wget https://download.clearlinux.org/releases/16050/clear/clear-16050-containers.img.xz rpm --install ./linux-container-4.9.33-74.x86_64.rpm xz --decompress clear-16050-containers.img.xz cp clear-16050-containers.img /usr/share/clear-containers/ pushd /usr/share/clear-containers/ ln -s clear-16050-containers.img clear-containers.img popd
After this, you should have a /usr/share/clear-containers directory that looks like this:
|-- clear-16050-containers.img |-- clear-containers.img -> clear-16050-containers.img |-- vmlinux-4.9.33-74.container |-- vmlinux.container -> vmlinux-4.9.33-74.container |-- vmlinuz-4.9.33-74.container `-- vmlinuz.container -> vmlinuz-4.9.33-74.container
Configuring It
The configuration is located in /usr/share/defaults/cc-oci-runtime. There are 3 files.
- vm.json defines which hypervisor, kernel and rootfs we use:
{ "vm": { "path": "/usr/bin/qemu-system-x86_64", "image": "/usr/share/clear-containers/clear-containers.img", "kernel": { "path": "/usr/share/clear-containers/vmlinux.container", "parameters": "root=/dev/pmem0p1 rootflags=dax,data=ordered,errors=remount-ro rw rootfstype=ext4 tsc=reliable no_timer_check rcupdate.rcu_expedited=1 i8042.direct=1 i8042.dumbkbd=1 i8042.nopnp=1 i8042.noaux=1 noreplace-smp reboot=k panic=1 console=hvc0 console=hvc1 initcall_debug init=/usr/lib/systemd/systemd systemd.unit=cc-agent.target iommu=off quiet systemd.mask=systemd-networkd.service systemd.mask=systemd-networkd.socket systemd.show_status=false cryptomgr.notests net.ifnames=0" } } }
- kernel-cmdline is well, the kernel command line
root=/dev/pmem0p1 rootflags=dax,data=ordered,errors=remount-ro rw rootfstype=ext4 tsc=reliable no_timer_check rcupdate.rcu_expedited=1 i8042.direct=1 i8042.dumbkbd=1 i8042.nopnp=1 i8042.noaux=1 noreplace-smp reboot=k panic=1 console=hvc0 console=hvc1 initcall_debug init=/usr/lib/systemd/systemd systemd.unit=cc-agent.target iommu=off quiet systemd.mask=systemd-networkd.service systemd.mask=systemd-networkd.socket systemd.show_status=false cryptomgr.notests net.ifnames=0
- hypervisor.args defines how we start up the qemu hypervisor
/usr/bin/qemu-system-x86_64 -name @NAME@ -machine pc-lite,accel=kvm,kernel_irqchip,nvdimm -device nvdimm,memdev=mem0,id=nv0 -object memory-backend-file,id=mem0,mem-path=@IMAGE@,size=@SIZE@ -m 2G,slots=2,maxmem=3G -kernel @KERNEL@ -append @KERNEL_PARAMS@ @KERNEL_NET_PARAMS@ -smp 2,sockets=1,cores=2,threads=1 -cpu host
This is the file we need to change to accommodate our small memory availability on the Minnowboard Turbot (4 G total) . We are also using a newer qemu and machine type than the hypervisor.args file assumes. This helps to make the Atom processor happy. Here's the diff:
# diff hypervisor.args.orig hypervisor.args 5c5 < pc-lite,accel=kvm,kernel_irqchip,nvdimm --- > q35,accel=kvm,kernel_irqchip,nvdimm,nosmm,nosmbus,nosata,nopit,nofw 11c11 < 2G,slots=2,maxmem=3G --- > 256M,slots=2,maxmem=1G
I capped it at 1G for now. This setting would definitely need to be tweaked depending on needs/usage...
The Qemu Hypervisor
Getting the Source Code
git clone https://github.com/clearcontainers/qemu qemu-cc cd qemu-cc git checkout -b qemu-lite-v2.9.0 origin/qemu-lite-v2.9.0 mkdir build cd build ../configure '--disable-tools' '--disable-libssh2' '--disable-tcmalloc' '--disable-glusterfs' '--disable-seccomp' '--disable-bzip2' '--disable-snappy' '--disable-lzo' '--disable-usb-redir' '--disable-libusb' '--disable-libnfs' '--disable-tcg-interpreter' '--disable-debug-tcg' '--disable-libiscsi' '--disable-rbd' '--disable-spice' '--disable-attr' '--disable-cap-ng' '--disable-linux-aio' '--disable-brlapi' '--disable-vnc-jpeg' '--disable-vnc-png' '--disable-vnc-sasl' '--disable-rdma' '--disable-bluez' '--disable-fdt' '--disable-curl' '--disable-curses' '--disable-sdl' '--disable-gtk' '--disable-tpm' '--disable-vte' '--disable-vnc' '--disable-xen' '--disable-opengl' '--disable-slirp' '--enable-trace-backend=nop' '--enable-virtfs' '--enable-attr' '--enable-cap-ng' '--target-list=x86_64-softmmu' "$@"