vhost-device/README.md
Stefano Garzarella b113dc6b28 sound: promote to main workspace
The current implementation seems ready to be promoted to the
main workspace since the device supports both playback and recording,
and several audio backends.

Also, test coverage is good and aligned with the coverage of crates
in the main workspace.

Signed-off-by: Stefano Garzarella <sgarzare@redhat.com>
2024-01-03 11:16:07 +05:30

115 lines
4.8 KiB
Markdown

# vhost-device
## Design
This repository hosts various 'vhost-user' device backends in their own crates.
See their individual README.md files for specific information about those
crates.
To be included here device backends must:
- be based on a published [VIRTIO specification](https://github.com/oasis-tcs/virtio-spec)
- fulfil basic functionality requirements (in conjunction with a implemented driver)
- meet the [rust-vmm dev requirements](https://github.com/rust-vmm/community#publishing-on-cratesio---requirements-list)
Here is the list of device backends that we support:
- [GPIO](https://github.com/rust-vmm/vhost-device/blob/main/vhost-device-gpio/README.md)
- [I2C](https://github.com/rust-vmm/vhost-device/blob/main/vhost-device-i2c/README.md)
- [Input](https://github.com/rust-vmm/vhost-device/blob/main/vhost-device-input/README.md)
- [RNG](https://github.com/rust-vmm/vhost-device/blob/main/vhost-device-rng/README.md)
- [SCMI](https://github.com/rust-vmm/vhost-device/blob/main/vhost-device-scmi/README.md)
- [SCSI](https://github.com/rust-vmm/vhost-device/blob/main/vhost-device-scsi/README.md)
- [Sound](https://github.com/rust-vmm/vhost-device/blob/main/vhost-device-sound/README.md)
- [VSOCK](https://github.com/rust-vmm/vhost-device/blob/main/vhost-device-vsock/README.md)
The vhost-device workspace also provides a
[template](https://github.com/rust-vmm/vhost-device/blob/main/vhost-device-template/README.md)
to help new developers understand how to write their own vhost-user backend.
### Staging Devices
Implementing a proper VirtIO device requires co-ordination between the
specification, drivers and backend implementations. As these can all
be in flux during development it was decided introducing a staging
workspace which would allow developers to work within the main rust-vmm
project while clearly marking the backends as not production ready.
To be included in the staging workspace there must at least be:
- A public proposal to extend the [VIRTIO specification](https://github.com/oasis-tcs/virtio-spec)
- A public implementation of a device driver
- Documentation pointing to the above
More information may be found in its [README file](./staging/README.md).
Here is the list of device backends in **staging**:
- [Video](https://github.com/rust-vmm/vhost-device/blob/main/staging/vhost-device-video/README.md)
<!--
Template:
- [`_DEVICE_NAME_`](https://github.com/rust-vmm/vhost-device/blob/main/staging/vhost-device-_DEVICE_NAME_/README.md)
-->
## Testing and Code Coverage
Like the wider rust-vmm project we expect new features to come with
comprehensive code coverage. However as a multi-binary repository
there are cases where avoiding a drop in coverage can be hard and an
exception to the approach is allowable. These are:
* adding a new binary target (aim at least 60% overall coverage)
* expanding the main function (a small drop is acceptable)
However any new feature added to an existing binary should not cause a
drop in coverage. The general aim should be to always improve
coverage.
## Separation of Concerns
The binaries built by this repository can be run with any VMM which
can act as a vhost-user frontend. Typically they have been tested with
[QEMU](https://www.qemu.org) although the rust-vmm project does
provide a [vhost-user
frontend](https://github.com/rust-vmm/vhost/tree/main/src/vhost_user)
crate for rust based VMMs.
While it's possible to implement all parts of the backend inside the
vhost-device workspace consideration should be given to separating the
VirtQueue handling and response logic to a crate in [vm-virtio
devices](https://github.com/rust-vmm/vm-virtio/tree/main/crates/devices).
This way a monolithic rust-vmm VMM implementation can reuse the core
logic to service the virtio requests directly in the application.
## Build dependency
The GPIO crate needs a local installation of libgpiod library to be available.
If your distro ships libgpiod >= v2.0, then you should be fine.
Otherwise, you will need to build libgpiod yourself:
git clone --depth 1 --branch v2.0.x https://git.kernel.org/pub/scm/libs/libgpiod/libgpiod.git/
cd libgpiod
./autogen.sh --prefix="$PWD/install/"
make install
In order to inform tools about the build location, you can now set:
export PKG_CONFIG_PATH="<PATH-TO-LIBGPIOD>/install/lib/pkgconfig/"
To prevent setting this in every terminal session, you can also configure
cargo to
[set it automatically](https://doc.rust-lang.org/cargo/reference/config.html#env).
## Xen support
Supporting Xen requires special handling while mapping the guest memory. The
`vm-memory` crate implements xen memory mapping support via a separate feature
`xen`, and this crate uses the same feature name to enable Xen support.
It was decided by the `rust-vmm` maintainers to keep the interface simple and
build the crate for either standard Unix memory mapping or Xen, and not both.