cloud-hypervisor/docs/fs.md
Sebastien Boeuf 2dc9cfaf01 docs: Update virtio-fs documentation
Update the content of the documentation to clarify how the cache option
from virtiofsd has an impact on the host page cache, while the dax
option for the filesystem in the guest has an impact on the guest page
cache.

The document is also updated to not have lines exceeding 80 characters.

Fixes #3826

Signed-off-by: Sebastien Boeuf <sebastien.boeuf@intel.com>
2022-03-18 11:25:10 +01:00

3.8 KiB

How to use virtio-fs

In the context of virtualization, it is always convenient to be able to share a directory from the host with the guest.

virtio-fs, also known as vhost-user-fs is a virtual device defined by the VIRTIO specification which allows any VMM to perform filesystem sharing.

Pre-requisites

The daemon

This virtual device relies on the vhost-user protocol, which assumes the backend (device emulation) is handled by a dedicated process running on the host. This daemon is called virtiofsd and needs to be present on the host.

Build virtiofsd

git clone https://gitlab.com/virtio-fs/virtiofsd
pushd virtiofsd
cargo build --release
sudo setcap cap_sys_admin+epi target/release/virtiofsd

Create shared directory

mkdir /tmp/shared_dir

Run virtiofsd

./virtiofsd \
    -d \
    --socket-path=/tmp/virtiofs \
    --shared-dir=/tmp/shared_dir \
    --cache=never

The cache=never option is the default when using virtiofsd with Cloud Hypervisor. This prevents from using the host page cache, reducing the overall footprint on host memory. This increases the maximum density of virtual machines that can be launched on a single host.

The cache=always option will allow the host page cache to be used, which can result in better performance for the guest's workload at the cost of increasing the footprint on host memory.

Kernel support

Modern Linux kernels (at least v5.10) have support for virtio-fs. Use of older kernels, with additional patches, are not supported.

How to share directories with cloud-hypervisor

Start the VM

Once the daemon is running, the option --fs from Cloud Hypervisor needs to be used.

Both direct kernel boot and EFI firmware can be used to boot a VM with virtio-fs, given that the cloud image contains a recent enough kernel.

Correct functioning of --fs requires --memory shared=on to facilitate interprocess memory sharing.

Assuming you have focal-server-cloudimg-amd64.raw and vmlinux on your system, here is the Cloud Hypervisor command you need to run:

./cloud-hypervisor \
    --cpus boot=1 \
    --memory size=1G,shared=on \
    --disk path=focal-server-cloudimg-amd64.raw \
    --kernel vmlinux \
    --cmdline "console=hvc0 root=/dev/vda1 rw" \
    --fs tag=myfs,socket=/tmp/virtiofs,num_queues=1,queue_size=512

By default, DAX is enabled with a cache window of 8GiB. You can specify a custom size (let's say 4GiB for this example) for the cache by explicitly setting DAX and the cache size:

--fs tag=myfs,socket=/tmp/virtiofs,num_queues=1,queue_size=512,dax=on,cache_size=4G

In case you don't want to use a shared window of cache to pass the shared files content, this means you will have to explicitly disable DAX with dax=off. Note that in this case, the cache_size parameter will be ignored.

--fs tag=myfs,socket=/tmp/virtiofs,num_queues=1,queue_size=512,dax=off

Mount the shared directory

The last step is to mount the shared directory inside the guest, using the virtiofs filesystem type.

mkdir mount_dir
mount -t virtiofs -o dax myfs mount_dir/

The tag needs to be consistent with what has been provided through the Cloud Hypervisor command line, which happens to be myfs in this example.

The -o dax option must be removed in case the shared cache region is not enabled from the VMM.

The filesystem should be mounted with dax by default as it bypasses the guest page cache, reducing the footprint on host memory. When running multiple virtual machines on the same host, this will let the host deal with page cache, which will increase the density of virtual machines which can be launched.

When the filesystem is not mounted with the dax option, the guest page cache is used, which can result in better performance for the guest's workload at the cost of increasing the footprint on host memory.