#include "kvm/virtio-pci-dev.h" #include "kvm/virtio-net.h" #include "kvm/virtio.h" #include "kvm/mutex.h" #include "kvm/util.h" #include "kvm/kvm.h" #include "kvm/uip.h" #include "kvm/guest_compat.h" #include "kvm/iovec.h" #include "kvm/strbuf.h" #include #include #include #include #include #include #include #include #include #include #include #include #include #define VIRTIO_NET_QUEUE_SIZE 256 #define VIRTIO_NET_NUM_QUEUES 8 struct net_dev; struct net_dev_operations { int (*rx)(struct iovec *iov, u16 in, struct net_dev *ndev); int (*tx)(struct iovec *iov, u16 in, struct net_dev *ndev); }; struct net_dev_queue { int id; struct net_dev *ndev; struct virt_queue vq; pthread_t thread; struct mutex lock; pthread_cond_t cond; }; struct net_dev { struct mutex mutex; struct virtio_device vdev; struct list_head list; struct net_dev_queue queues[VIRTIO_NET_NUM_QUEUES * 2 + 1]; struct virtio_net_config config; u32 queue_pairs; int vhost_fd; int tap_fd; char tap_name[IFNAMSIZ]; bool tap_ufo; int mode; struct uip_info info; struct net_dev_operations *ops; struct kvm *kvm; struct virtio_net_params *params; }; static LIST_HEAD(ndevs); static int compat_id = -1; #define MAX_PACKET_SIZE 65550 static bool has_virtio_feature(struct net_dev *ndev, u32 feature) { return ndev->vdev.features & (1 << feature); } static int virtio_net_hdr_len(struct net_dev *ndev) { if (has_virtio_feature(ndev, VIRTIO_NET_F_MRG_RXBUF) || !ndev->vdev.legacy) return sizeof(struct virtio_net_hdr_mrg_rxbuf); return sizeof(struct virtio_net_hdr); } static void *virtio_net_rx_thread(void *p) { struct iovec iov[VIRTIO_NET_QUEUE_SIZE]; struct net_dev_queue *queue = p; struct virt_queue *vq = &queue->vq; struct net_dev *ndev = queue->ndev; struct kvm *kvm; u16 out, in; u16 head; int len, copied; kvm__set_thread_name("virtio-net-rx"); kvm = ndev->kvm; while (1) { mutex_lock(&queue->lock); if (!virt_queue__available(vq)) pthread_cond_wait(&queue->cond, &queue->lock.mutex); mutex_unlock(&queue->lock); while (virt_queue__available(vq)) { unsigned char buffer[MAX_PACKET_SIZE + sizeof(struct virtio_net_hdr_mrg_rxbuf)]; struct iovec dummy_iov = { .iov_base = buffer, .iov_len = sizeof(buffer), }; struct virtio_net_hdr_mrg_rxbuf *hdr; u16 num_buffers; len = ndev->ops->rx(&dummy_iov, 1, ndev); if (len < 0) { pr_warning("%s: rx on vq %u failed (%d), exiting thread\n", __func__, queue->id, len); goto out_err; } copied = num_buffers = 0; head = virt_queue__get_iov(vq, iov, &out, &in, kvm); hdr = iov[0].iov_base; while (copied < len) { size_t iovsize = min_t(size_t, len - copied, iov_size(iov, in)); memcpy_toiovec(iov, buffer + copied, iovsize); copied += iovsize; virt_queue__set_used_elem_no_update(vq, head, iovsize, num_buffers++); if (copied == len) break; while (!virt_queue__available(vq)) sleep(0); head = virt_queue__get_iov(vq, iov, &out, &in, kvm); } /* * The device MUST set num_buffers, except in the case * where the legacy driver did not negotiate * VIRTIO_NET_F_MRG_RXBUF and the field does not exist. */ if (has_virtio_feature(ndev, VIRTIO_NET_F_MRG_RXBUF) || !ndev->vdev.legacy) hdr->num_buffers = virtio_host_to_guest_u16(vq->endian, num_buffers); virt_queue__used_idx_advance(vq, num_buffers); /* We should interrupt guest right now, otherwise latency is huge. */ if (virtio_queue__should_signal(vq)) ndev->vdev.ops->signal_vq(kvm, &ndev->vdev, queue->id); } } out_err: pthread_exit(NULL); return NULL; } static void *virtio_net_tx_thread(void *p) { struct iovec iov[VIRTIO_NET_QUEUE_SIZE]; struct net_dev_queue *queue = p; struct virt_queue *vq = &queue->vq; struct net_dev *ndev = queue->ndev; struct kvm *kvm; u16 out, in; u16 head; int len; kvm__set_thread_name("virtio-net-tx"); kvm = ndev->kvm; while (1) { mutex_lock(&queue->lock); if (!virt_queue__available(vq)) pthread_cond_wait(&queue->cond, &queue->lock.mutex); mutex_unlock(&queue->lock); while (virt_queue__available(vq)) { head = virt_queue__get_iov(vq, iov, &out, &in, kvm); len = ndev->ops->tx(iov, out, ndev); if (len < 0) { pr_warning("%s: tx on vq %u failed (%d)\n", __func__, queue->id, errno); goto out_err; } virt_queue__set_used_elem(vq, head, len); } if (virtio_queue__should_signal(vq)) ndev->vdev.ops->signal_vq(kvm, &ndev->vdev, queue->id); } out_err: pthread_exit(NULL); return NULL; } static virtio_net_ctrl_ack virtio_net_handle_mq(struct kvm* kvm, struct net_dev *ndev, struct virtio_net_ctrl_hdr *ctrl) { /* Not much to do here */ return VIRTIO_NET_OK; } static void *virtio_net_ctrl_thread(void *p) { struct iovec iov[VIRTIO_NET_QUEUE_SIZE]; struct net_dev_queue *queue = p; struct virt_queue *vq = &queue->vq; struct net_dev *ndev = queue->ndev; u16 out, in, head; struct kvm *kvm = ndev->kvm; struct virtio_net_ctrl_hdr ctrl; virtio_net_ctrl_ack ack; size_t len; kvm__set_thread_name("virtio-net-ctrl"); while (1) { mutex_lock(&queue->lock); if (!virt_queue__available(vq)) pthread_cond_wait(&queue->cond, &queue->lock.mutex); mutex_unlock(&queue->lock); while (virt_queue__available(vq)) { head = virt_queue__get_iov(vq, iov, &out, &in, kvm); len = min(iov_size(iov, in), sizeof(ctrl)); memcpy_fromiovec((void *)&ctrl, iov, len); switch (ctrl.class) { case VIRTIO_NET_CTRL_MQ: ack = virtio_net_handle_mq(kvm, ndev, &ctrl); break; default: ack = VIRTIO_NET_ERR; break; } memcpy_toiovec(iov + in, &ack, sizeof(ack)); virt_queue__set_used_elem(vq, head, sizeof(ack)); } if (virtio_queue__should_signal(vq)) ndev->vdev.ops->signal_vq(kvm, &ndev->vdev, queue->id); } pthread_exit(NULL); return NULL; } static void virtio_net_handle_callback(struct kvm *kvm, struct net_dev *ndev, int queue) { struct net_dev_queue *net_queue = &ndev->queues[queue]; if ((u32)queue >= (ndev->queue_pairs * 2 + 1)) { pr_warning("Unknown queue index %u", queue); return; } mutex_lock(&net_queue->lock); pthread_cond_signal(&net_queue->cond); mutex_unlock(&net_queue->lock); } static int virtio_net_request_tap(struct net_dev *ndev, struct ifreq *ifr, const char *tapname) { int ret; memset(ifr, 0, sizeof(*ifr)); ifr->ifr_flags = IFF_TAP | IFF_NO_PI | IFF_VNET_HDR; if (tapname) strlcpy(ifr->ifr_name, tapname, sizeof(ifr->ifr_name)); ret = ioctl(ndev->tap_fd, TUNSETIFF, ifr); if (ret >= 0) strlcpy(ndev->tap_name, ifr->ifr_name, sizeof(ndev->tap_name)); return ret; } static int virtio_net_exec_script(const char* script, const char *tap_name) { pid_t pid; int status; pid = vfork(); if (pid == 0) { execl(script, script, tap_name, NULL); _exit(1); } else { waitpid(pid, &status, 0); if (WIFEXITED(status) && WEXITSTATUS(status) != 0) { pr_warning("Fail to setup tap by %s", script); return -1; } } return 0; } static bool virtio_net__tap_init(struct net_dev *ndev) { int sock = socket(AF_INET, SOCK_STREAM, 0); int hdr_len; struct sockaddr_in sin = {0}; struct ifreq ifr; const struct virtio_net_params *params = ndev->params; bool skipconf = !!params->tapif; hdr_len = virtio_net_hdr_len(ndev); if (ioctl(ndev->tap_fd, TUNSETVNETHDRSZ, &hdr_len) < 0) pr_warning("Config tap device TUNSETVNETHDRSZ error"); if (strcmp(params->script, "none")) { if (virtio_net_exec_script(params->script, ndev->tap_name) < 0) goto fail; } else if (!skipconf) { memset(&ifr, 0, sizeof(ifr)); strncpy(ifr.ifr_name, ndev->tap_name, sizeof(ifr.ifr_name)); sin.sin_addr.s_addr = inet_addr(params->host_ip); memcpy(&(ifr.ifr_addr), &sin, sizeof(ifr.ifr_addr)); ifr.ifr_addr.sa_family = AF_INET; if (ioctl(sock, SIOCSIFADDR, &ifr) < 0) { pr_warning("Could not set ip address on tap device"); goto fail; } } if (!skipconf) { memset(&ifr, 0, sizeof(ifr)); strncpy(ifr.ifr_name, ndev->tap_name, sizeof(ifr.ifr_name)); ioctl(sock, SIOCGIFFLAGS, &ifr); ifr.ifr_flags |= IFF_UP | IFF_RUNNING; if (ioctl(sock, SIOCSIFFLAGS, &ifr) < 0) pr_warning("Could not bring tap device up"); } close(sock); return 1; fail: if (sock >= 0) close(sock); if (ndev->tap_fd >= 0) close(ndev->tap_fd); return 0; } static void virtio_net__tap_exit(struct net_dev *ndev) { int sock; struct ifreq ifr; if (ndev->params->tapif) return; sock = socket(AF_INET, SOCK_STREAM, 0); strncpy(ifr.ifr_name, ndev->tap_name, sizeof(ifr.ifr_name)); ioctl(sock, SIOCGIFFLAGS, &ifr); ifr.ifr_flags &= ~(IFF_UP | IFF_RUNNING); if (ioctl(sock, SIOCGIFFLAGS, &ifr) < 0) pr_warning("Count not bring tap device down"); close(sock); } static bool virtio_net__tap_create(struct net_dev *ndev) { int offload; struct ifreq ifr; const struct virtio_net_params *params = ndev->params; bool macvtap = (!!params->tapif) && (params->tapif[0] == '/'); /* Did the user already gave us the FD? */ if (params->fd) ndev->tap_fd = params->fd; else { const char *tap_file = "/dev/net/tun"; /* Did the user ask us to use macvtap? */ if (macvtap) tap_file = params->tapif; ndev->tap_fd = open(tap_file, O_RDWR); if (ndev->tap_fd < 0) { pr_warning("Unable to open %s", tap_file); return 0; } } if (!macvtap && virtio_net_request_tap(ndev, &ifr, params->tapif) < 0) { pr_warning("Config tap device error. Are you root?"); goto fail; } /* * The UFO support had been removed from kernel in commit: * ID: fb652fdfe83710da0ca13448a41b7ed027d0a984 * https://www.spinics.net/lists/netdev/msg443562.html * In oder to support the older kernels without this commit, * we set the TUN_F_UFO to offload by default to test the status of * UFO kernel support. */ ndev->tap_ufo = true; offload = TUN_F_CSUM | TUN_F_TSO4 | TUN_F_TSO6 | TUN_F_UFO; if (ioctl(ndev->tap_fd, TUNSETOFFLOAD, offload) < 0) { /* * Is this failure caused by kernel remove the UFO support? * Try TUNSETOFFLOAD without TUN_F_UFO. */ offload &= ~TUN_F_UFO; if (ioctl(ndev->tap_fd, TUNSETOFFLOAD, offload) < 0) { pr_warning("Config tap device TUNSETOFFLOAD error"); goto fail; } ndev->tap_ufo = false; } return 1; fail: if ((ndev->tap_fd >= 0) || (!params->fd) ) close(ndev->tap_fd); return 0; } static inline int tap_ops_tx(struct iovec *iov, u16 out, struct net_dev *ndev) { return writev(ndev->tap_fd, iov, out); } static inline int tap_ops_rx(struct iovec *iov, u16 in, struct net_dev *ndev) { return readv(ndev->tap_fd, iov, in); } static inline int uip_ops_tx(struct iovec *iov, u16 out, struct net_dev *ndev) { return uip_tx(iov, out, &ndev->info); } static inline int uip_ops_rx(struct iovec *iov, u16 in, struct net_dev *ndev) { return uip_rx(iov, in, &ndev->info); } static struct net_dev_operations tap_ops = { .rx = tap_ops_rx, .tx = tap_ops_tx, }; static struct net_dev_operations uip_ops = { .rx = uip_ops_rx, .tx = uip_ops_tx, }; static u8 *get_config(struct kvm *kvm, void *dev) { struct net_dev *ndev = dev; return ((u8 *)(&ndev->config)); } static size_t get_config_size(struct kvm *kvm, void *dev) { struct net_dev *ndev = dev; return sizeof(ndev->config); } static u64 get_host_features(struct kvm *kvm, void *dev) { u64 features; struct net_dev *ndev = dev; features = 1UL << VIRTIO_NET_F_MAC | 1UL << VIRTIO_NET_F_CSUM | 1UL << VIRTIO_NET_F_HOST_TSO4 | 1UL << VIRTIO_NET_F_HOST_TSO6 | 1UL << VIRTIO_NET_F_GUEST_TSO4 | 1UL << VIRTIO_NET_F_GUEST_TSO6 | 1UL << VIRTIO_RING_F_EVENT_IDX | 1UL << VIRTIO_RING_F_INDIRECT_DESC | 1UL << VIRTIO_NET_F_CTRL_VQ | 1UL << VIRTIO_NET_F_MRG_RXBUF | 1UL << (ndev->queue_pairs > 1 ? VIRTIO_NET_F_MQ : 0) | 1UL << VIRTIO_F_ANY_LAYOUT; /* * The UFO feature for host and guest only can be enabled when the * kernel has TAP UFO support. */ if (ndev->tap_ufo) features |= (1UL << VIRTIO_NET_F_HOST_UFO | 1UL << VIRTIO_NET_F_GUEST_UFO); if (ndev->vhost_fd) { u64 vhost_features; if (ioctl(ndev->vhost_fd, VHOST_GET_FEATURES, &vhost_features) != 0) die_perror("VHOST_GET_FEATURES failed"); features &= vhost_features; } return features; } static void virtio_net_start(struct net_dev *ndev) { /* VHOST_NET_F_VIRTIO_NET_HDR clashes with VIRTIO_F_ANY_LAYOUT! */ u64 features = ndev->vdev.features & ~(1UL << VHOST_NET_F_VIRTIO_NET_HDR); if (ndev->mode == NET_MODE_TAP) { if (!virtio_net__tap_init(ndev)) die_perror("TAP device initialized failed because"); if (ndev->vhost_fd && virtio_vhost_set_features(ndev->vhost_fd, features)) die_perror("VHOST_SET_FEATURES failed"); } else { ndev->info.vnet_hdr_len = virtio_net_hdr_len(ndev); uip_init(&ndev->info); } } static void virtio_net_stop(struct net_dev *ndev) { /* Undo whatever start() did */ if (ndev->mode == NET_MODE_TAP) virtio_net__tap_exit(ndev); else uip_exit(&ndev->info); } static void virtio_net_update_endian(struct net_dev *ndev) { struct virtio_net_config *conf = &ndev->config; conf->status = virtio_host_to_guest_u16(ndev->vdev.endian, VIRTIO_NET_S_LINK_UP); conf->max_virtqueue_pairs = virtio_host_to_guest_u16(ndev->vdev.endian, ndev->queue_pairs); /* Let TAP know about vnet header endianness */ if (ndev->mode == NET_MODE_TAP && ndev->vdev.endian != VIRTIO_ENDIAN_HOST) { int enable_val = 1, disable_val = 0; int enable_req, disable_req; if (ndev->vdev.endian == VIRTIO_ENDIAN_LE) { enable_req = TUNSETVNETLE; disable_req = TUNSETVNETBE; } else { enable_req = TUNSETVNETBE; disable_req = TUNSETVNETLE; } ioctl(ndev->tap_fd, disable_req, &disable_val); if (ioctl(ndev->tap_fd, enable_req, &enable_val) < 0) pr_err("Config tap device TUNSETVNETLE/BE error"); } } static void notify_status(struct kvm *kvm, void *dev, u32 status) { struct net_dev *ndev = dev; if (status & VIRTIO__STATUS_CONFIG) virtio_net_update_endian(ndev); if (status & VIRTIO__STATUS_START) virtio_net_start(dev); else if (status & VIRTIO__STATUS_STOP) virtio_net_stop(dev); } static bool is_ctrl_vq(struct net_dev *ndev, u32 vq) { return vq == (u32)(ndev->queue_pairs * 2); } static int init_vq(struct kvm *kvm, void *dev, u32 vq) { struct vhost_vring_file file = { .index = vq }; struct net_dev_queue *net_queue; struct net_dev *ndev = dev; struct virt_queue *queue; int r; compat__remove_message(compat_id); net_queue = &ndev->queues[vq]; net_queue->id = vq; net_queue->ndev = ndev; queue = &net_queue->vq; virtio_init_device_vq(kvm, &ndev->vdev, queue, VIRTIO_NET_QUEUE_SIZE); mutex_init(&net_queue->lock); pthread_cond_init(&net_queue->cond, NULL); if (is_ctrl_vq(ndev, vq)) { pthread_create(&net_queue->thread, NULL, virtio_net_ctrl_thread, net_queue); return 0; } else if (ndev->vhost_fd == 0 ) { if (vq & 1) pthread_create(&net_queue->thread, NULL, virtio_net_tx_thread, net_queue); else pthread_create(&net_queue->thread, NULL, virtio_net_rx_thread, net_queue); return 0; } virtio_vhost_set_vring(kvm, ndev->vhost_fd, vq, queue); file.fd = ndev->tap_fd; r = ioctl(ndev->vhost_fd, VHOST_NET_SET_BACKEND, &file); if (r < 0) die_perror("VHOST_NET_SET_BACKEND failed"); return 0; } static void exit_vq(struct kvm *kvm, void *dev, u32 vq) { struct net_dev *ndev = dev; struct net_dev_queue *queue = &ndev->queues[vq]; virtio_vhost_reset_vring(kvm, ndev->vhost_fd, vq, &queue->vq); /* * TODO: vhost reset owner. It's the only way to cleanly stop vhost, but * we can't restart it at the moment. */ if (ndev->vhost_fd && !is_ctrl_vq(ndev, vq)) { pr_warning("Cannot reset VHOST queue"); ioctl(ndev->vhost_fd, VHOST_RESET_OWNER); return; } /* * Threads are waiting on cancellation points (readv or * pthread_cond_wait) and should stop gracefully. */ pthread_cancel(queue->thread); pthread_join(queue->thread, NULL); } static void notify_vq_gsi(struct kvm *kvm, void *dev, u32 vq, u32 gsi) { struct net_dev *ndev = dev; struct net_dev_queue *queue = &ndev->queues[vq]; if (ndev->vhost_fd == 0 || is_ctrl_vq(ndev, vq)) return; virtio_vhost_set_vring_irqfd(kvm, gsi, &queue->vq); } static void notify_vq_eventfd(struct kvm *kvm, void *dev, u32 vq, u32 efd) { struct net_dev *ndev = dev; if (ndev->vhost_fd == 0 || is_ctrl_vq(ndev, vq)) return; virtio_vhost_set_vring_kick(kvm, ndev->vhost_fd, vq, efd); } static int notify_vq(struct kvm *kvm, void *dev, u32 vq) { struct net_dev *ndev = dev; virtio_net_handle_callback(kvm, ndev, vq); return 0; } static struct virt_queue *get_vq(struct kvm *kvm, void *dev, u32 vq) { struct net_dev *ndev = dev; return &ndev->queues[vq].vq; } static int get_size_vq(struct kvm *kvm, void *dev, u32 vq) { /* FIXME: dynamic */ return VIRTIO_NET_QUEUE_SIZE; } static int set_size_vq(struct kvm *kvm, void *dev, u32 vq, int size) { /* FIXME: dynamic */ return size; } static unsigned int get_vq_count(struct kvm *kvm, void *dev) { struct net_dev *ndev = dev; return ndev->queue_pairs * 2 + 1; } static struct virtio_ops net_dev_virtio_ops = { .get_config = get_config, .get_config_size = get_config_size, .get_host_features = get_host_features, .get_vq_count = get_vq_count, .init_vq = init_vq, .exit_vq = exit_vq, .get_vq = get_vq, .get_size_vq = get_size_vq, .set_size_vq = set_size_vq, .notify_vq = notify_vq, .notify_vq_gsi = notify_vq_gsi, .notify_vq_eventfd = notify_vq_eventfd, .notify_status = notify_status, }; static void virtio_net__vhost_init(struct kvm *kvm, struct net_dev *ndev) { if (ndev->queue_pairs > 1) { pr_warning("multiqueue is not supported with vhost yet"); return; } ndev->vhost_fd = open("/dev/vhost-net", O_RDWR); if (ndev->vhost_fd < 0) die_perror("Failed openning vhost-net device"); virtio_vhost_init(kvm, ndev->vhost_fd); ndev->vdev.use_vhost = true; } static inline void str_to_mac(const char *str, char *mac) { sscanf(str, "%hhx:%hhx:%hhx:%hhx:%hhx:%hhx", mac, mac+1, mac+2, mac+3, mac+4, mac+5); } static int set_net_param(struct kvm *kvm, struct virtio_net_params *p, const char *param, const char *val) { if (strcmp(param, "guest_mac") == 0) { str_to_mac(val, p->guest_mac); } else if (strcmp(param, "mode") == 0) { if (!strncmp(val, "user", 4)) { int i; for (i = 0; i < kvm->cfg.num_net_devices; i++) if (kvm->cfg.net_params[i].mode == NET_MODE_USER) die("Only one usermode network device allowed at a time"); p->mode = NET_MODE_USER; } else if (!strncmp(val, "tap", 3)) { p->mode = NET_MODE_TAP; } else if (!strncmp(val, "none", 4)) { kvm->cfg.no_net = 1; return -1; } else die("Unknown network mode %s, please use user, tap or none", kvm->cfg.network); } else if (strcmp(param, "script") == 0) { p->script = strdup(val); } else if (strcmp(param, "downscript") == 0) { p->downscript = strdup(val); } else if (strcmp(param, "guest_ip") == 0) { p->guest_ip = strdup(val); } else if (strcmp(param, "host_ip") == 0) { p->host_ip = strdup(val); } else if (strcmp(param, "trans") == 0) { p->trans = strdup(val); } else if (strcmp(param, "tapif") == 0) { p->tapif = strdup(val); } else if (strcmp(param, "vhost") == 0) { p->vhost = atoi(val); } else if (strcmp(param, "fd") == 0) { p->fd = atoi(val); } else if (strcmp(param, "mq") == 0) { p->mq = atoi(val); } else die("Unknown network parameter %s", param); return 0; } int netdev_parser(const struct option *opt, const char *arg, int unset) { struct virtio_net_params p; char *buf = NULL, *cmd = NULL, *cur = NULL; bool on_cmd = true; struct kvm *kvm = opt->ptr; if (arg) { buf = strdup(arg); if (buf == NULL) die("Failed allocating new net buffer"); cur = strtok(buf, ",="); } p = (struct virtio_net_params) { .guest_ip = DEFAULT_GUEST_ADDR, .host_ip = DEFAULT_HOST_ADDR, .script = DEFAULT_SCRIPT, .downscript = DEFAULT_SCRIPT, .mode = NET_MODE_TAP, }; str_to_mac(DEFAULT_GUEST_MAC, p.guest_mac); p.guest_mac[5] += kvm->cfg.num_net_devices; while (cur) { if (on_cmd) { cmd = cur; } else { if (set_net_param(kvm, &p, cmd, cur) < 0) goto done; } on_cmd = !on_cmd; cur = strtok(NULL, ",="); }; kvm->cfg.num_net_devices++; kvm->cfg.net_params = realloc(kvm->cfg.net_params, kvm->cfg.num_net_devices * sizeof(*kvm->cfg.net_params)); if (kvm->cfg.net_params == NULL) die("Failed adding new network device"); kvm->cfg.net_params[kvm->cfg.num_net_devices - 1] = p; done: free(buf); return 0; } static int virtio_net__init_one(struct virtio_net_params *params) { enum virtio_trans trans = params->kvm->cfg.virtio_transport; struct net_dev *ndev; struct virtio_ops *ops; int i, r; ndev = calloc(1, sizeof(struct net_dev)); if (ndev == NULL) return -ENOMEM; list_add_tail(&ndev->list, &ndevs); ops = malloc(sizeof(*ops)); if (ops == NULL) return -ENOMEM; ndev->kvm = params->kvm; ndev->params = params; mutex_init(&ndev->mutex); ndev->queue_pairs = max(1, min(VIRTIO_NET_NUM_QUEUES, params->mq)); for (i = 0 ; i < 6 ; i++) { ndev->config.mac[i] = params->guest_mac[i]; ndev->info.guest_mac.addr[i] = params->guest_mac[i]; ndev->info.host_mac.addr[i] = params->host_mac[i]; } ndev->mode = params->mode; if (ndev->mode == NET_MODE_TAP) { ndev->ops = &tap_ops; if (!virtio_net__tap_create(ndev)) die_perror("You have requested a TAP device, but creation of one has failed because"); } else { ndev->info.host_ip = ntohl(inet_addr(params->host_ip)); ndev->info.guest_ip = ntohl(inet_addr(params->guest_ip)); ndev->info.guest_netmask = ntohl(inet_addr("255.255.255.0")); ndev->info.buf_nr = 20, ndev->ops = &uip_ops; uip_static_init(&ndev->info); } *ops = net_dev_virtio_ops; if (params->trans) { if (strcmp(params->trans, "mmio") == 0) trans = VIRTIO_MMIO; else if (strcmp(params->trans, "pci") == 0) trans = VIRTIO_PCI; else pr_warning("virtio-net: Unknown transport method : %s, " "falling back to %s.", params->trans, virtio_trans_name(trans)); } r = virtio_init(params->kvm, ndev, &ndev->vdev, ops, trans, PCI_DEVICE_ID_VIRTIO_NET, VIRTIO_ID_NET, PCI_CLASS_NET); if (r < 0) { free(ops); return r; } if (params->vhost) virtio_net__vhost_init(params->kvm, ndev); if (compat_id == -1) compat_id = virtio_compat_add_message("virtio-net", "CONFIG_VIRTIO_NET"); return 0; } int virtio_net__init(struct kvm *kvm) { int i, r; for (i = 0; i < kvm->cfg.num_net_devices; i++) { kvm->cfg.net_params[i].kvm = kvm; r = virtio_net__init_one(&kvm->cfg.net_params[i]); if (r < 0) goto cleanup; } if (kvm->cfg.num_net_devices == 0 && kvm->cfg.no_net == 0) { static struct virtio_net_params net_params; net_params = (struct virtio_net_params) { .guest_ip = kvm->cfg.guest_ip, .host_ip = kvm->cfg.host_ip, .kvm = kvm, .script = kvm->cfg.script, .mode = NET_MODE_USER, }; str_to_mac(kvm->cfg.guest_mac, net_params.guest_mac); str_to_mac(kvm->cfg.host_mac, net_params.host_mac); r = virtio_net__init_one(&net_params); if (r < 0) goto cleanup; } return 0; cleanup: virtio_net__exit(kvm); return r; } virtio_dev_init(virtio_net__init); int virtio_net__exit(struct kvm *kvm) { struct virtio_net_params *params; struct net_dev *ndev; struct list_head *ptr, *n; list_for_each_safe(ptr, n, &ndevs) { ndev = list_entry(ptr, struct net_dev, list); params = ndev->params; /* Cleanup any tap device which attached to bridge */ if (ndev->mode == NET_MODE_TAP && strcmp(params->downscript, "none")) virtio_net_exec_script(params->downscript, ndev->tap_name); virtio_net_stop(ndev); list_del(&ndev->list); virtio_exit(kvm, &ndev->vdev); free(ndev); } return 0; } virtio_dev_exit(virtio_net__exit);