Skip to content

tiiuae/tii_sel4_build

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

TII seL4 build system

These instructions have been tested with Ubuntu 20.10 desktop and Fedora 33.

Setting up the build environment

Update your computer and install prerequisites

Ubuntu

host% sudo apt-get -y update
host% sudo apt-get -y upgrade
host% sudo apt -y install git repo

Fedora

host% sudo dnf update
host% sudo dnf install -y git
host% mkdir -p ~/.local/bin
host% curl https://storage.googleapis.com/git-repo-downloads/repo > ~/.local/bin/repo
host% chmod u+x ~/.local/bin/repo

Configure git

Only linear git history allowed, no merge commits.

host% git config --global user.email "[email protected]"
host% git config --global user.name "Your Name"
host% git config --global merge.ff only

Set up github access

Unless you have an existing SSH public/private key pair, generate one with ssh-keygen. You may want to supply an empty password or git cloning process will be quite cumbersome. In your github account, go to “Settings” → “SSH and GPG keys” and upload your ${HOME}/.ssh/id_rsa.pub there. Make sure you upload the public key, not the private key. Make sure the latter is secure!

Install and configure docker

Ubuntu

host% sudo apt -y install docker docker.io

Fedora

host% sudo dnf install -y dnf-plugins-core
host% sudo dnf config-manager --add-repo https://download.docker.com/linux/fedora/docker-ce.repo
host% sudo dnf install -y docker-ce docker-ce-cli containerd.io
host% sudo systemctl enable docker
host% sudo systemctl start docker

Add yourself to docker group:

host% sudo usermod -aG docker $USER

In order to supplementary group change take effect, either reboot your computer or log out and back in (the most lazy ones can use the newgrp command). In any case verify with the groups command.

Configure Yocto download directory

host% export YOCTO_SOURCE_MIRROR_DIR=~/yocto-downloads
host% echo 'export YOCTO_SOURCE_MIRROR_DIR='${YOCTO_SOURCE_MIRROR_DIR} >> ~/.bashrc
host% mkdir ${YOCTO_SOURCE_MIRROR_DIR}

Configure build cache directory

Compiling Haskell packages takes quite a long time, several minutes on a power machine. The make equivalent, stack, is able to cache the build results and because the Haskell sources stay the same for most of the time (versions used are configured in capdl.git), it is advisable to export a host directory to container to store the cached build results across containers. Note that for CI purposes you need to judge yourself when to enable the cache.

host% export BUILD_CACHE_DIR=~/.tii_sel4_build
host% echo 'export BUILD_CACHE_DIR='${BUILD_CACHE_DIR} >> ~/.bashrc
host% mkdir -p ${BUILD_CACHE_DIR}/stack

Check out sources

# Choose a working directory, this will be visible in the container at /workspace
# (the WORKSPACE variable will point to /workspace as well inside the container)
host% export WORKSPACE=~/sel4

host% mkdir ${WORKSPACE} && cd ${WORKSPACE}
host% repo init -u [email protected]:tiiuae/tii_sel4_manifest.git -b tii/development
host% repo sync

Build docker images

Prune the Docker caches before proceeding with building the image:

host% docker image prune -a
host% docker builder prune -a

Build the image:

host% make docker

Prepopulate build cache

host% make raspberrypi4-64_defconfig
host% make build_cache

Use it!

host% cd ${WORKSPACE}

# configure for Raspberry Pi 4

host% make raspberrypi4-64_defconfig

# simple seL4 microkernel test

host% make sel4test
host% ls -l rpi4_sel4test/images
-rwxr-xr-x. 1 build build 5832040 Aug 31 08:31 sel4test-driver-image-arm-bcm2711

# More complex examples with VMs (you need to build the Yocto
# images first, see below)

host% make vm_minimal
host% ls -l rpi4_vm_minimal/images
-rwxr-xr-x. 1 build build 37641488 Aug 28 02:50 capdl-loader-image-arm-bcm2711

host% make vm_multi
host% ls -l rpi4_vm_multi/images
-rwxr-xr-x. 1 build build 51656592 Aug 28 02:52 capdl-loader-image-arm-bcm2711

# Enter build container interactively -- you get all tools
# (for example, aarch64-linux-gnu toolchain) without having
# to install them onto host.

host% make shell

container% aarch64-linux-gnu-objdump -D rpi4_sel4test/kernel/kernel.elf

Network booting

This seL4 playground boots from network.

Setting up NFS root

Create a directory on the NFS server:

host% sudo mkdir -p /srv/nfs/rpi4

Add this directory to NFS exports:

host% cat /etc/exports
/srv/nfs/rpi4 192.168.5.0/24(rw,no_root_squash)
host% sudo exportfs -a

Pass NFS server info via DHCP

host% cat /etc/dhcp/dhcpd.conf
option domain-name     "local.domain";
default-lease-time 600;
max-lease-time 7200;
authoritative;

subnet 192.168.5.0 netmask 255.255.255.0 {
    range dynamic-bootp 192.168.5.200 192.168.5.254;
    option broadcast-address 192.168.5.255;
    option routers 192.168.5.1;
    next-server 192.168.5.1;

    option root-path "192.168.5.1:/srv/nfs/rpi4,vers=3,proto=tcp,nolock";
}

It is the most convenient to use TCP since the Fedora firewall will block UDP by default.

Setup board for network booting

See instructions in hardware/${PLATFORM}/README.md:

QEMU, virtio and seL4

The vm_qemu_virtio example demonstrates TII's experimental work to bring QEMU-backed virtio devices into guest VMs. An instance of QEMU is running in driver-VM and it provides virtio-blk, virtio-console, virtio-net and virtfs to the user-VM guest. This demo abandons Buildroot and uses Yocto to build the guest VM images.

host$ repo init -u [email protected]:tiiuae/tii_sel4_manifest.git -b tii/development
host$ repo sync
host$ make docker

# configure for Raspberry Pi 4
host$ make raspberrypi4-64_defconfig

# build VM root filesystems, Linux kernel and initramfs
host$ make linux-image

# build qemu seL4 vm_qemu_virtio
host$ make vm_qemu_virtio

# prepare TFTP directory for booting CAmkES example on target
host$ sudo ./tii_sel4_build/hardware/rpi4/prepare_camkes_boot.sh vm_qemu_virtio

# expose driver-VM image via NFS (update your directory to command)
host$ tar -C /srv/nfs/rpi4 -xjpvf /workspace/projects/camkes-vm-images/rpi4/vm-image-driver.tar.bz2

# create host/guest shared directory
host$ mkdir /srv/nfs/rpi4/host

Using QEMU demo

After the driver-VM has booted, log in (empty root password) and start the user-VM:

driver-vm$ screen -c screenrc-drivervm

This will start a screen session, with one shell for interactive use and the another one is QEMU's stdout, works also as a console for user-VM. You can switch between the windows using ^A. In case you are using minicom, you need to press ^A twice.

When the user-VM has booted, log in. There is screenrc-uservm, which starts an interactive shell and stress tests for virtio-blk, virtio-net and virtio-console. To launch it, just type:

user-vm$ screen -c screenrc-uservm

Within user-VM, the screen control character has been mapped to ^B.

Rebuilding guest VM components

We use Yocto to build guest root filesystems, Linux kernel and initramfs. Buildroot is not supported.

On a first run building everything will take several hours and you need at least 100 GB disk space for the build. After the build finishes, you will find everything /workspace/projects/camkes-vm-images/rpi4.

host% make linux-image

 ...

Here are your images in /workspace/projects/camkes-vm-images/rpi4:
total 12
lrwxrwxrwx. 1 build build  69 Feb 28 14:12 linux -> /workspace/vm-images/build/tmp/deploy/images/vm-raspberrypi4-64/Image
lrwxrwxrwx. 1 build build 104 Feb 28 14:12 rootfs.cpio.gz -> /workspace/vm-images/build/tmp/deploy/images/vm-raspberrypi4-64/vm-image-boot-vm-raspberrypi4-64.cpio.gz
lrwxrwxrwx. 1 build build 106 Feb 28 14:16 vm-image-driver.tar.bz2 -> /workspace/vm-images/build/tmp/deploy/images/vm-raspberrypi4-64/vm-image-driver-vm-raspberrypi4-64.tar.bz2

The driver-VM root filesystem vm-image-driver.tar.bz2 will be mounted over NFS. See instructions below on how to make this root filesystem available to NFS clients.

Customizing Yocto packages

To customize the guest Linux kernel, use facilities Yocto provides:

# Use this to modify just the configs
container% bitbake -c do_menuconfig linux-raspberrypi
container% bitbake linux-raspberrypi

# With devtool you can edit the sources after Yocto has patched them
container% devtool modify kernel-module-sel4-virtio
INFO: Source tree extracted to /workspace/vm-images/build/workspace/sources/kernel-module-sel4-virtio
INFO: Using source tree as build directory since that would be the default for this recipe
INFO: Recipe kernel-module-sel4-virtio now set up to build from /workspace/vm-images/build/workspace/sources/kernel-module-sel4-virtio

The module sources are now visible on the host at ${WORKSPACE}/vm-images/build/workspace/sources/kernel-module-sel4-virtio, use your favorite editor to play with them.

# To rebuild the rootfs with changes made to kernel module:
container% bitbake vm-image-driver

# Commit your changes to temporary git repo devtool made for you
host% cd ${WORKSPACE}/vm-images/build/workspace/sources/kernel-module-sel4-virtio
host% git commit -a --signoff

# Use devtool to embed the change commits into meta-sel4 layer as patches:
container% devtool update-recipe kernel-module-sel4-virtio
INFO: Adding new patch 0001-Say-hello.patch
INFO: Updating recipe kernel-module-sel4-virtio_git.bb

You will find patch you made in ${WORKSPACE}/vm-images/meta-sel4/recipes-kernel/sel4-virtio/kernel-module-sel4-virtio and the patch added to SRC_URI field in the recipe. To get rid of the working copy:

container% devtool reset kernel-module-sel4-virtio

Building Yocto SDK package

TODO: should building/installing the SDK be automated? Yocto system can build an SDK package installer. The package is primarily meant for application developers, so that they can have a SDK toolchain separate from the Yocto build system. We mainly use it to provide GDB debugger, because the Ubuntus' multiarch GDB seems to have some oddities.

host% make shell
container% cd vm-images
# This command will set up your environment for Yocto builds, so remember
# to execute it every time you enter the container.
container% source setup.sh
container% bitbake vm-image-driver -c populate-sdk

Yocto will download and build toolchains. After the build finishes, you will find the SDK at /workspace/vm-images/build/tmp/deploy/sdk/.

Installing the SDK after build:

container% ./tmp/deploy/sdk/poky-glibc-x86_64-vm-image-driver-cortexa72-raspberrypi4-64-toolchain-3.4.sh
Poky (Yocto Project Reference Distro) SDK installer version 3.4
===============================================================
Enter target directory for SDK (default: /opt/poky/3.4): 
You are about to install the SDK to "/opt/poky/3.4". Proceed [Y/n]? y
[sudo] password for build: 
Extracting SDK.....................................................................done
Setting it up...done
SDK has been successfully set up and is ready to be used.
Each time you wish to use the SDK in a new shell session, you need to source the environment setup script e.g.
 $ . /opt/poky/3.4/environment-setup-cortexa72-poky-linux
container% ll /opt/poky/3.4/
total 24
drwxr-xr-x. 3 root root   144 Dec 10 11:50 ./
drwxr-xr-x. 3 root root    17 Dec 10 11:49 ../
-rw-r--r--. 1 root root  3848 Dec 10 11:50 environment-setup-cortexa72-poky-linux
-rw-r--r--. 1 root root 14027 Dec 10 11:50 site-config-cortexa72-poky-linux
drwxr-xr-x. 4 root root    62 Dec 10 11:49 sysroots/
-rw-r--r--. 1 root root   119 Dec 10 11:50 version-cortexa72-poky-linux

All tools and libraries can now be found from /opt/poky/3.4/sysroots/