mirror of
				https://github.com/qemu/qemu.git
				synced 2025-10-31 04:06:46 +00:00 
			
		
		
		
	 8aab0d1dbe
			
		
	
	
		8aab0d1dbe
		
	
	
	
	
		
			
			Add functions to support configure interrupt in virtio_net Add the functions to support vhost_net_config_pending and vhost_net_config_mask. Signed-off-by: Cindy Lu <lulu@redhat.com> Message-Id: <20221222070451.936503-9-lulu@redhat.com> Acked-by: Jason Wang <jasowang@redhat.com> Reviewed-by: Michael S. Tsirkin <mst@redhat.com> Signed-off-by: Michael S. Tsirkin <mst@redhat.com>
		
			
				
	
	
		
			631 lines
		
	
	
		
			17 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			631 lines
		
	
	
		
			17 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * vhost-net support
 | |
|  *
 | |
|  * Copyright Red Hat, Inc. 2010
 | |
|  *
 | |
|  * Authors:
 | |
|  *  Michael S. Tsirkin <mst@redhat.com>
 | |
|  *
 | |
|  * This work is licensed under the terms of the GNU GPL, version 2.  See
 | |
|  * the COPYING file in the top-level directory.
 | |
|  *
 | |
|  * Contributions after 2012-01-13 are licensed under the terms of the
 | |
|  * GNU GPL, version 2 or (at your option) any later version.
 | |
|  */
 | |
| 
 | |
| #include "qemu/osdep.h"
 | |
| #include "net/net.h"
 | |
| #include "net/tap.h"
 | |
| #include "net/vhost-user.h"
 | |
| #include "net/vhost-vdpa.h"
 | |
| 
 | |
| #include "standard-headers/linux/vhost_types.h"
 | |
| #include "hw/virtio/virtio-net.h"
 | |
| #include "net/vhost_net.h"
 | |
| #include "qapi/error.h"
 | |
| #include "qemu/error-report.h"
 | |
| #include "qemu/main-loop.h"
 | |
| 
 | |
| #include <sys/socket.h>
 | |
| #include <net/if.h>
 | |
| #include <netinet/in.h>
 | |
| 
 | |
| 
 | |
| #include "standard-headers/linux/virtio_ring.h"
 | |
| #include "hw/virtio/vhost.h"
 | |
| #include "hw/virtio/virtio-bus.h"
 | |
| #include "linux-headers/linux/vhost.h"
 | |
| 
 | |
| 
 | |
| /* Features supported by host kernel. */
 | |
| static const int kernel_feature_bits[] = {
 | |
|     VIRTIO_F_NOTIFY_ON_EMPTY,
 | |
|     VIRTIO_RING_F_INDIRECT_DESC,
 | |
|     VIRTIO_RING_F_EVENT_IDX,
 | |
|     VIRTIO_NET_F_MRG_RXBUF,
 | |
|     VIRTIO_F_VERSION_1,
 | |
|     VIRTIO_NET_F_MTU,
 | |
|     VIRTIO_F_IOMMU_PLATFORM,
 | |
|     VIRTIO_F_RING_PACKED,
 | |
|     VIRTIO_F_RING_RESET,
 | |
|     VIRTIO_NET_F_HASH_REPORT,
 | |
|     VHOST_INVALID_FEATURE_BIT
 | |
| };
 | |
| 
 | |
| /* Features supported by others. */
 | |
| static const int user_feature_bits[] = {
 | |
|     VIRTIO_F_NOTIFY_ON_EMPTY,
 | |
|     VIRTIO_RING_F_INDIRECT_DESC,
 | |
|     VIRTIO_RING_F_EVENT_IDX,
 | |
| 
 | |
|     VIRTIO_F_ANY_LAYOUT,
 | |
|     VIRTIO_F_VERSION_1,
 | |
|     VIRTIO_NET_F_CSUM,
 | |
|     VIRTIO_NET_F_GUEST_CSUM,
 | |
|     VIRTIO_NET_F_GSO,
 | |
|     VIRTIO_NET_F_GUEST_TSO4,
 | |
|     VIRTIO_NET_F_GUEST_TSO6,
 | |
|     VIRTIO_NET_F_GUEST_ECN,
 | |
|     VIRTIO_NET_F_GUEST_UFO,
 | |
|     VIRTIO_NET_F_HOST_TSO4,
 | |
|     VIRTIO_NET_F_HOST_TSO6,
 | |
|     VIRTIO_NET_F_HOST_ECN,
 | |
|     VIRTIO_NET_F_HOST_UFO,
 | |
|     VIRTIO_NET_F_MRG_RXBUF,
 | |
|     VIRTIO_NET_F_MTU,
 | |
|     VIRTIO_F_IOMMU_PLATFORM,
 | |
|     VIRTIO_F_RING_PACKED,
 | |
|     VIRTIO_F_RING_RESET,
 | |
|     VIRTIO_NET_F_RSS,
 | |
|     VIRTIO_NET_F_HASH_REPORT,
 | |
| 
 | |
|     /* This bit implies RARP isn't sent by QEMU out of band */
 | |
|     VIRTIO_NET_F_GUEST_ANNOUNCE,
 | |
| 
 | |
|     VIRTIO_NET_F_MQ,
 | |
| 
 | |
|     VHOST_INVALID_FEATURE_BIT
 | |
| };
 | |
| 
 | |
| static const int *vhost_net_get_feature_bits(struct vhost_net *net)
 | |
| {
 | |
|     const int *feature_bits = 0;
 | |
| 
 | |
|     switch (net->nc->info->type) {
 | |
|     case NET_CLIENT_DRIVER_TAP:
 | |
|         feature_bits = kernel_feature_bits;
 | |
|         break;
 | |
|     case NET_CLIENT_DRIVER_VHOST_USER:
 | |
|         feature_bits = user_feature_bits;
 | |
|         break;
 | |
| #ifdef CONFIG_VHOST_NET_VDPA
 | |
|     case NET_CLIENT_DRIVER_VHOST_VDPA:
 | |
|         feature_bits = vdpa_feature_bits;
 | |
|         break;
 | |
| #endif
 | |
|     default:
 | |
|         error_report("Feature bits not defined for this type: %d",
 | |
|                 net->nc->info->type);
 | |
|         break;
 | |
|     }
 | |
| 
 | |
|     return feature_bits;
 | |
| }
 | |
| 
 | |
| uint64_t vhost_net_get_features(struct vhost_net *net, uint64_t features)
 | |
| {
 | |
|     return vhost_get_features(&net->dev, vhost_net_get_feature_bits(net),
 | |
|             features);
 | |
| }
 | |
| int vhost_net_get_config(struct vhost_net *net,  uint8_t *config,
 | |
|                          uint32_t config_len)
 | |
| {
 | |
|     return vhost_dev_get_config(&net->dev, config, config_len, NULL);
 | |
| }
 | |
| int vhost_net_set_config(struct vhost_net *net, const uint8_t *data,
 | |
|                          uint32_t offset, uint32_t size, uint32_t flags)
 | |
| {
 | |
|     return vhost_dev_set_config(&net->dev, data, offset, size, flags);
 | |
| }
 | |
| 
 | |
| void vhost_net_ack_features(struct vhost_net *net, uint64_t features)
 | |
| {
 | |
|     net->dev.acked_features = net->dev.backend_features;
 | |
|     vhost_ack_features(&net->dev, vhost_net_get_feature_bits(net), features);
 | |
| }
 | |
| 
 | |
| uint64_t vhost_net_get_max_queues(VHostNetState *net)
 | |
| {
 | |
|     return net->dev.max_queues;
 | |
| }
 | |
| 
 | |
| uint64_t vhost_net_get_acked_features(VHostNetState *net)
 | |
| {
 | |
|     return net->dev.acked_features;
 | |
| }
 | |
| 
 | |
| void vhost_net_save_acked_features(NetClientState *nc)
 | |
| {
 | |
| #ifdef CONFIG_VHOST_NET_USER
 | |
|     if (nc->info->type == NET_CLIENT_DRIVER_VHOST_USER) {
 | |
|         vhost_user_save_acked_features(nc);
 | |
|     }
 | |
| #endif
 | |
| }
 | |
| 
 | |
| static int vhost_net_get_fd(NetClientState *backend)
 | |
| {
 | |
|     switch (backend->info->type) {
 | |
|     case NET_CLIENT_DRIVER_TAP:
 | |
|         return tap_get_fd(backend);
 | |
|     default:
 | |
|         fprintf(stderr, "vhost-net requires tap backend\n");
 | |
|         return -ENOSYS;
 | |
|     }
 | |
| }
 | |
| 
 | |
| struct vhost_net *vhost_net_init(VhostNetOptions *options)
 | |
| {
 | |
|     int r;
 | |
|     bool backend_kernel = options->backend_type == VHOST_BACKEND_TYPE_KERNEL;
 | |
|     struct vhost_net *net = g_new0(struct vhost_net, 1);
 | |
|     uint64_t features = 0;
 | |
|     Error *local_err = NULL;
 | |
| 
 | |
|     if (!options->net_backend) {
 | |
|         fprintf(stderr, "vhost-net requires net backend to be setup\n");
 | |
|         goto fail;
 | |
|     }
 | |
|     net->nc = options->net_backend;
 | |
|     net->dev.nvqs = options->nvqs;
 | |
| 
 | |
|     net->dev.max_queues = 1;
 | |
|     net->dev.vqs = net->vqs;
 | |
| 
 | |
|     if (backend_kernel) {
 | |
|         r = vhost_net_get_fd(options->net_backend);
 | |
|         if (r < 0) {
 | |
|             goto fail;
 | |
|         }
 | |
|         net->dev.backend_features = qemu_has_vnet_hdr(options->net_backend)
 | |
|             ? 0 : (1ULL << VHOST_NET_F_VIRTIO_NET_HDR);
 | |
|         net->backend = r;
 | |
|         net->dev.protocol_features = 0;
 | |
|     } else {
 | |
|         net->dev.backend_features = 0;
 | |
|         net->dev.protocol_features = 0;
 | |
|         net->backend = -1;
 | |
| 
 | |
|         /* vhost-user needs vq_index to initiate a specific queue pair */
 | |
|         net->dev.vq_index = net->nc->queue_index * net->dev.nvqs;
 | |
|     }
 | |
| 
 | |
|     r = vhost_dev_init(&net->dev, options->opaque,
 | |
|                        options->backend_type, options->busyloop_timeout,
 | |
|                        &local_err);
 | |
|     if (r < 0) {
 | |
|         error_report_err(local_err);
 | |
|         goto fail;
 | |
|     }
 | |
|     if (backend_kernel) {
 | |
|         if (!qemu_has_vnet_hdr_len(options->net_backend,
 | |
|                                sizeof(struct virtio_net_hdr_mrg_rxbuf))) {
 | |
|             net->dev.features &= ~(1ULL << VIRTIO_NET_F_MRG_RXBUF);
 | |
|         }
 | |
|         if (~net->dev.features & net->dev.backend_features) {
 | |
|             fprintf(stderr, "vhost lacks feature mask 0x%" PRIx64
 | |
|                    " for backend\n",
 | |
|                    (uint64_t)(~net->dev.features & net->dev.backend_features));
 | |
|             goto fail;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     /* Set sane init value. Override when guest acks. */
 | |
| #ifdef CONFIG_VHOST_NET_USER
 | |
|     if (net->nc->info->type == NET_CLIENT_DRIVER_VHOST_USER) {
 | |
|         features = vhost_user_get_acked_features(net->nc);
 | |
|         if (~net->dev.features & features) {
 | |
|             fprintf(stderr, "vhost lacks feature mask 0x%" PRIx64
 | |
|                     " for backend\n",
 | |
|                     (uint64_t)(~net->dev.features & features));
 | |
|             goto fail;
 | |
|         }
 | |
|     }
 | |
| #endif
 | |
| 
 | |
|     vhost_net_ack_features(net, features);
 | |
| 
 | |
|     return net;
 | |
| 
 | |
| fail:
 | |
|     vhost_dev_cleanup(&net->dev);
 | |
|     g_free(net);
 | |
|     return NULL;
 | |
| }
 | |
| 
 | |
| static void vhost_net_set_vq_index(struct vhost_net *net, int vq_index,
 | |
|                                    int vq_index_end)
 | |
| {
 | |
|     net->dev.vq_index = vq_index;
 | |
|     net->dev.vq_index_end = vq_index_end;
 | |
| }
 | |
| 
 | |
| static int vhost_net_start_one(struct vhost_net *net,
 | |
|                                VirtIODevice *dev)
 | |
| {
 | |
|     struct vhost_vring_file file = { };
 | |
|     int r;
 | |
| 
 | |
|     if (net->nc->info->start) {
 | |
|         r = net->nc->info->start(net->nc);
 | |
|         if (r < 0) {
 | |
|             return r;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     r = vhost_dev_enable_notifiers(&net->dev, dev);
 | |
|     if (r < 0) {
 | |
|         goto fail_notifiers;
 | |
|     }
 | |
| 
 | |
|     r = vhost_dev_start(&net->dev, dev, false);
 | |
|     if (r < 0) {
 | |
|         goto fail_start;
 | |
|     }
 | |
| 
 | |
|     if (net->nc->info->poll) {
 | |
|         net->nc->info->poll(net->nc, false);
 | |
|     }
 | |
| 
 | |
|     if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) {
 | |
|         qemu_set_fd_handler(net->backend, NULL, NULL, NULL);
 | |
|         file.fd = net->backend;
 | |
|         for (file.index = 0; file.index < net->dev.nvqs; ++file.index) {
 | |
|             if (!virtio_queue_enabled(dev, net->dev.vq_index +
 | |
|                                       file.index)) {
 | |
|                 /* Queue might not be ready for start */
 | |
|                 continue;
 | |
|             }
 | |
|             r = vhost_net_set_backend(&net->dev, &file);
 | |
|             if (r < 0) {
 | |
|                 r = -errno;
 | |
|                 goto fail;
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (net->nc->info->load) {
 | |
|         r = net->nc->info->load(net->nc);
 | |
|         if (r < 0) {
 | |
|             goto fail;
 | |
|         }
 | |
|     }
 | |
|     return 0;
 | |
| fail:
 | |
|     file.fd = -1;
 | |
|     if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) {
 | |
|         while (file.index-- > 0) {
 | |
|             if (!virtio_queue_enabled(dev, net->dev.vq_index +
 | |
|                                       file.index)) {
 | |
|                 /* Queue might not be ready for start */
 | |
|                 continue;
 | |
|             }
 | |
|             int r = vhost_net_set_backend(&net->dev, &file);
 | |
|             assert(r >= 0);
 | |
|         }
 | |
|     }
 | |
|     if (net->nc->info->poll) {
 | |
|         net->nc->info->poll(net->nc, true);
 | |
|     }
 | |
|     vhost_dev_stop(&net->dev, dev, false);
 | |
| fail_start:
 | |
|     vhost_dev_disable_notifiers(&net->dev, dev);
 | |
| fail_notifiers:
 | |
|     return r;
 | |
| }
 | |
| 
 | |
| static void vhost_net_stop_one(struct vhost_net *net,
 | |
|                                VirtIODevice *dev)
 | |
| {
 | |
|     struct vhost_vring_file file = { .fd = -1 };
 | |
| 
 | |
|     if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) {
 | |
|         for (file.index = 0; file.index < net->dev.nvqs; ++file.index) {
 | |
|             int r = vhost_net_set_backend(&net->dev, &file);
 | |
|             assert(r >= 0);
 | |
|         }
 | |
|     }
 | |
|     if (net->nc->info->poll) {
 | |
|         net->nc->info->poll(net->nc, true);
 | |
|     }
 | |
|     vhost_dev_stop(&net->dev, dev, false);
 | |
|     if (net->nc->info->stop) {
 | |
|         net->nc->info->stop(net->nc);
 | |
|     }
 | |
|     vhost_dev_disable_notifiers(&net->dev, dev);
 | |
| }
 | |
| 
 | |
| int vhost_net_start(VirtIODevice *dev, NetClientState *ncs,
 | |
|                     int data_queue_pairs, int cvq)
 | |
| {
 | |
|     BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(dev)));
 | |
|     VirtioBusState *vbus = VIRTIO_BUS(qbus);
 | |
|     VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(vbus);
 | |
|     int total_notifiers = data_queue_pairs * 2 + cvq;
 | |
|     VirtIONet *n = VIRTIO_NET(dev);
 | |
|     int nvhosts = data_queue_pairs + cvq;
 | |
|     struct vhost_net *net;
 | |
|     int r, e, i, index_end = data_queue_pairs * 2;
 | |
|     NetClientState *peer;
 | |
| 
 | |
|     if (cvq) {
 | |
|         index_end += 1;
 | |
|     }
 | |
| 
 | |
|     if (!k->set_guest_notifiers) {
 | |
|         error_report("binding does not support guest notifiers");
 | |
|         return -ENOSYS;
 | |
|     }
 | |
| 
 | |
|     for (i = 0; i < nvhosts; i++) {
 | |
| 
 | |
|         if (i < data_queue_pairs) {
 | |
|             peer = qemu_get_peer(ncs, i);
 | |
|         } else { /* Control Virtqueue */
 | |
|             peer = qemu_get_peer(ncs, n->max_queue_pairs);
 | |
|         }
 | |
| 
 | |
|         net = get_vhost_net(peer);
 | |
|         vhost_net_set_vq_index(net, i * 2, index_end);
 | |
| 
 | |
|         /* Suppress the masking guest notifiers on vhost user
 | |
|          * because vhost user doesn't interrupt masking/unmasking
 | |
|          * properly.
 | |
|          */
 | |
|         if (net->nc->info->type == NET_CLIENT_DRIVER_VHOST_USER) {
 | |
|             dev->use_guest_notifier_mask = false;
 | |
|         }
 | |
|      }
 | |
| 
 | |
|     r = k->set_guest_notifiers(qbus->parent, total_notifiers, true);
 | |
|     if (r < 0) {
 | |
|         error_report("Error binding guest notifier: %d", -r);
 | |
|         goto err;
 | |
|     }
 | |
| 
 | |
|     for (i = 0; i < nvhosts; i++) {
 | |
|         if (i < data_queue_pairs) {
 | |
|             peer = qemu_get_peer(ncs, i);
 | |
|         } else {
 | |
|             peer = qemu_get_peer(ncs, n->max_queue_pairs);
 | |
|         }
 | |
| 
 | |
|         if (peer->vring_enable) {
 | |
|             /* restore vring enable state */
 | |
|             r = vhost_set_vring_enable(peer, peer->vring_enable);
 | |
| 
 | |
|             if (r < 0) {
 | |
|                 goto err_start;
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         r = vhost_net_start_one(get_vhost_net(peer), dev);
 | |
|         if (r < 0) {
 | |
|             goto err_start;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| 
 | |
| err_start:
 | |
|     while (--i >= 0) {
 | |
|         peer = qemu_get_peer(ncs, i < data_queue_pairs ?
 | |
|                                   i : n->max_queue_pairs);
 | |
|         vhost_net_stop_one(get_vhost_net(peer), dev);
 | |
|     }
 | |
|     e = k->set_guest_notifiers(qbus->parent, total_notifiers, false);
 | |
|     if (e < 0) {
 | |
|         fprintf(stderr, "vhost guest notifier cleanup failed: %d\n", e);
 | |
|         fflush(stderr);
 | |
|     }
 | |
| err:
 | |
|     return r;
 | |
| }
 | |
| 
 | |
| void vhost_net_stop(VirtIODevice *dev, NetClientState *ncs,
 | |
|                     int data_queue_pairs, int cvq)
 | |
| {
 | |
|     BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(dev)));
 | |
|     VirtioBusState *vbus = VIRTIO_BUS(qbus);
 | |
|     VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(vbus);
 | |
|     VirtIONet *n = VIRTIO_NET(dev);
 | |
|     NetClientState *peer;
 | |
|     int total_notifiers = data_queue_pairs * 2 + cvq;
 | |
|     int nvhosts = data_queue_pairs + cvq;
 | |
|     int i, r;
 | |
| 
 | |
|     for (i = 0; i < nvhosts; i++) {
 | |
|         if (i < data_queue_pairs) {
 | |
|             peer = qemu_get_peer(ncs, i);
 | |
|         } else {
 | |
|             peer = qemu_get_peer(ncs, n->max_queue_pairs);
 | |
|         }
 | |
|         vhost_net_stop_one(get_vhost_net(peer), dev);
 | |
|     }
 | |
| 
 | |
|     r = k->set_guest_notifiers(qbus->parent, total_notifiers, false);
 | |
|     if (r < 0) {
 | |
|         fprintf(stderr, "vhost guest notifier cleanup failed: %d\n", r);
 | |
|         fflush(stderr);
 | |
|     }
 | |
|     assert(r >= 0);
 | |
| }
 | |
| 
 | |
| void vhost_net_cleanup(struct vhost_net *net)
 | |
| {
 | |
|     vhost_dev_cleanup(&net->dev);
 | |
| }
 | |
| 
 | |
| int vhost_net_notify_migration_done(struct vhost_net *net, char* mac_addr)
 | |
| {
 | |
|     const VhostOps *vhost_ops = net->dev.vhost_ops;
 | |
| 
 | |
|     assert(vhost_ops->backend_type == VHOST_BACKEND_TYPE_USER);
 | |
|     assert(vhost_ops->vhost_migration_done);
 | |
| 
 | |
|     return vhost_ops->vhost_migration_done(&net->dev, mac_addr);
 | |
| }
 | |
| 
 | |
| bool vhost_net_virtqueue_pending(VHostNetState *net, int idx)
 | |
| {
 | |
|     return vhost_virtqueue_pending(&net->dev, idx);
 | |
| }
 | |
| 
 | |
| void vhost_net_virtqueue_mask(VHostNetState *net, VirtIODevice *dev,
 | |
|                               int idx, bool mask)
 | |
| {
 | |
|     vhost_virtqueue_mask(&net->dev, dev, idx, mask);
 | |
| }
 | |
| 
 | |
| bool vhost_net_config_pending(VHostNetState *net)
 | |
| {
 | |
|     return vhost_config_pending(&net->dev);
 | |
| }
 | |
| 
 | |
| void vhost_net_config_mask(VHostNetState *net, VirtIODevice *dev, bool mask)
 | |
| {
 | |
|     vhost_config_mask(&net->dev, dev, mask);
 | |
| }
 | |
| VHostNetState *get_vhost_net(NetClientState *nc)
 | |
| {
 | |
|     VHostNetState *vhost_net = 0;
 | |
| 
 | |
|     if (!nc) {
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     switch (nc->info->type) {
 | |
|     case NET_CLIENT_DRIVER_TAP:
 | |
|         vhost_net = tap_get_vhost_net(nc);
 | |
|         break;
 | |
| #ifdef CONFIG_VHOST_NET_USER
 | |
|     case NET_CLIENT_DRIVER_VHOST_USER:
 | |
|         vhost_net = vhost_user_get_vhost_net(nc);
 | |
|         assert(vhost_net);
 | |
|         break;
 | |
| #endif
 | |
| #ifdef CONFIG_VHOST_NET_VDPA
 | |
|     case NET_CLIENT_DRIVER_VHOST_VDPA:
 | |
|         vhost_net = vhost_vdpa_get_vhost_net(nc);
 | |
|         assert(vhost_net);
 | |
|         break;
 | |
| #endif
 | |
|     default:
 | |
|         break;
 | |
|     }
 | |
| 
 | |
|     return vhost_net;
 | |
| }
 | |
| 
 | |
| int vhost_set_vring_enable(NetClientState *nc, int enable)
 | |
| {
 | |
|     VHostNetState *net = get_vhost_net(nc);
 | |
|     const VhostOps *vhost_ops = net->dev.vhost_ops;
 | |
| 
 | |
|     nc->vring_enable = enable;
 | |
| 
 | |
|     if (vhost_ops && vhost_ops->vhost_set_vring_enable) {
 | |
|         return vhost_ops->vhost_set_vring_enable(&net->dev, enable);
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| int vhost_net_set_mtu(struct vhost_net *net, uint16_t mtu)
 | |
| {
 | |
|     const VhostOps *vhost_ops = net->dev.vhost_ops;
 | |
| 
 | |
|     if (!vhost_ops->vhost_net_set_mtu) {
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     return vhost_ops->vhost_net_set_mtu(&net->dev, mtu);
 | |
| }
 | |
| 
 | |
| void vhost_net_virtqueue_reset(VirtIODevice *vdev, NetClientState *nc,
 | |
|                                int vq_index)
 | |
| {
 | |
|     VHostNetState *net = get_vhost_net(nc->peer);
 | |
|     const VhostOps *vhost_ops = net->dev.vhost_ops;
 | |
|     struct vhost_vring_file file = { .fd = -1 };
 | |
|     int idx;
 | |
| 
 | |
|     /* should only be called after backend is connected */
 | |
|     assert(vhost_ops);
 | |
| 
 | |
|     idx = vhost_ops->vhost_get_vq_index(&net->dev, vq_index);
 | |
| 
 | |
|     if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) {
 | |
|         file.index = idx;
 | |
|         int r = vhost_net_set_backend(&net->dev, &file);
 | |
|         assert(r >= 0);
 | |
|     }
 | |
| 
 | |
|     vhost_virtqueue_stop(&net->dev,
 | |
|                          vdev,
 | |
|                          net->dev.vqs + idx,
 | |
|                          net->dev.vq_index + idx);
 | |
| }
 | |
| 
 | |
| int vhost_net_virtqueue_restart(VirtIODevice *vdev, NetClientState *nc,
 | |
|                                 int vq_index)
 | |
| {
 | |
|     VHostNetState *net = get_vhost_net(nc->peer);
 | |
|     const VhostOps *vhost_ops = net->dev.vhost_ops;
 | |
|     struct vhost_vring_file file = { };
 | |
|     int idx, r;
 | |
| 
 | |
|     if (!net->dev.started) {
 | |
|         return -EBUSY;
 | |
|     }
 | |
| 
 | |
|     /* should only be called after backend is connected */
 | |
|     assert(vhost_ops);
 | |
| 
 | |
|     idx = vhost_ops->vhost_get_vq_index(&net->dev, vq_index);
 | |
| 
 | |
|     r = vhost_virtqueue_start(&net->dev,
 | |
|                               vdev,
 | |
|                               net->dev.vqs + idx,
 | |
|                               net->dev.vq_index + idx);
 | |
|     if (r < 0) {
 | |
|         goto err_start;
 | |
|     }
 | |
| 
 | |
|     if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) {
 | |
|         file.index = idx;
 | |
|         file.fd = net->backend;
 | |
|         r = vhost_net_set_backend(&net->dev, &file);
 | |
|         if (r < 0) {
 | |
|             r = -errno;
 | |
|             goto err_start;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| 
 | |
| err_start:
 | |
|     error_report("Error when restarting the queue.");
 | |
| 
 | |
|     if (net->nc->info->type == NET_CLIENT_DRIVER_TAP) {
 | |
|         file.fd = VHOST_FILE_UNBIND;
 | |
|         file.index = idx;
 | |
|         int r = vhost_net_set_backend(&net->dev, &file);
 | |
|         assert(r >= 0);
 | |
|     }
 | |
| 
 | |
|     vhost_dev_stop(&net->dev, vdev, false);
 | |
| 
 | |
|     return r;
 | |
| }
 |