On Sun, Apr 16, 2023 at 01:14:00PM -0400, Dave Voutila wrote:
> Moving vmd to use zero-copy semantics for virtqueues introduced a bug in
> the vm send/receive functionality. The host va is potentially invalid on
> restore if vmd has restarted and re-randomized the address space of the
> vmm process that forks vm's.
>
> This NULL's out the hva to and resets it on restore.
>
> This fix is also required for my upcoming "+exec" diff because each vm
> will get a new address space every execution, so the hva has practically
> no chance of being valid on restore.
>
> ok?
>
ok mlarkin, sorry for the delay reviewing this.
> Index: virtio.c
> ===================================================================
> RCS file: /cvs/src/usr.sbin/vmd/virtio.c,v
> retrieving revision 1.99
> diff -u -p -r1.99 virtio.c
> --- virtio.c 28 Dec 2022 21:30:19 -0000 1.99
> +++ virtio.c 16 Apr 2023 17:11:29 -0000
> @@ -2015,6 +2015,8 @@ vmmci_restore(int fd, uint32_t vm_id)
> int
> viornd_restore(int fd, struct vm_create_params *vcp)
> {
> + void *hva = NULL;
> +
> log_debug("%s: receiving viornd", __func__);
> if (atomicio(read, fd, &viornd, sizeof(viornd)) != sizeof(viornd)) {
> log_warnx("%s: error reading viornd from fd", __func__);
> @@ -2028,6 +2030,11 @@ viornd_restore(int fd, struct vm_create_
> viornd.vm_id = vcp->vcp_id;
> viornd.irq = pci_get_dev_irq(viornd.pci_id);
>
> + hva = hvaddr_mem(viornd.vq[0].q_gpa, vring_size(VIORND_QUEUE_SIZE));
> + if (hva == NULL)
> + fatal("failed to restore viornd virtqueue");
> + viornd.vq[0].q_hva = hva;
> +
> return (0);
> }
>
> @@ -2038,6 +2045,7 @@ vionet_restore(int fd, struct vmd_vm *vm
> struct vm_create_params *vcp = &vmc->vmc_params;
> uint8_t i;
> int ret;
> + void *hva = NULL;
>
> nr_vionet = vcp->vcp_nnics;
> if (vcp->vcp_nnics > 0) {
> @@ -2079,6 +2087,18 @@ vionet_restore(int fd, struct vmd_vm *vm
> vionet[i].vm_vmid = vm->vm_vmid;
> vionet[i].irq = pci_get_dev_irq(vionet[i].pci_id);
>
> + hva = hvaddr_mem(vionet[i].vq[RXQ].q_gpa,
> + vring_size(VIONET_QUEUE_SIZE));
> + if (hva == NULL)
> + fatal("failed to restore vionet RX virtqueue");
> + vionet[i].vq[RXQ].q_hva = hva;
> +
> + hva = hvaddr_mem(vionet[i].vq[TXQ].q_gpa,
> + vring_size(VIONET_QUEUE_SIZE));
> + if (hva == NULL)
> + fatal("failed to restore vionet TX virtqueue");
> + vionet[i].vq[TXQ].q_hva = hva;
> +
> memset(&vionet[i].event, 0, sizeof(struct event));
> event_set(&vionet[i].event, vionet[i].fd,
> EV_READ | EV_PERSIST, vionet_rx_event, &vionet[i]);
> @@ -2093,6 +2113,7 @@ vioblk_restore(int fd, struct vmop_creat
> {
> struct vm_create_params *vcp = &vmc->vmc_params;
> uint8_t i;
> + void *hva = NULL;
>
> nr_vioblk = vcp->vcp_ndisks;
> vioblk = calloc(vcp->vcp_ndisks, sizeof(struct vioblk_dev));
> @@ -2123,6 +2144,12 @@ vioblk_restore(int fd, struct vmop_creat
> }
> vioblk[i].vm_id = vcp->vcp_id;
> vioblk[i].irq = pci_get_dev_irq(vioblk[i].pci_id);
> +
> + hva = hvaddr_mem(vioblk[i].vq[0].q_gpa,
> + vring_size(VIOBLK_QUEUE_SIZE));
> + if (hva == NULL)
> + fatal("failed to restore vioblk virtqueue");
> + vioblk[i].vq[0].q_hva = hva;
> }
> return (0);
> }
> @@ -2130,6 +2157,9 @@ vioblk_restore(int fd, struct vmop_creat
> int
> vioscsi_restore(int fd, struct vm_create_params *vcp, int child_cdrom)
> {
> + void *hva = NULL;
> + unsigned int i;
> +
> if (!strlen(vcp->vcp_cdrom))
> return (0);
>
> @@ -2161,6 +2191,15 @@ vioscsi_restore(int fd, struct vm_create
> vioscsi->vm_id = vcp->vcp_id;
> vioscsi->irq = pci_get_dev_irq(vioscsi->pci_id);
>
> + /* vioscsi uses 3 virtqueues. */
> + for (i = 0; i < 3; i++) {
> + hva = hvaddr_mem(vioscsi->vq[i].q_gpa,
> + vring_size(VIOSCSI_QUEUE_SIZE));
> + if (hva == NULL)
> + fatal("failed to restore vioscsi virtqueue");
> + vioscsi->vq[i].q_hva = hva;
> + }
> +
> return (0);
> }
>
> @@ -2194,6 +2233,9 @@ int
> viornd_dump(int fd)
> {
> log_debug("%s: sending viornd", __func__);
> +
> + viornd.vq[0].q_hva = NULL;
> +
> if (atomicio(vwrite, fd, &viornd, sizeof(viornd)) != sizeof(viornd)) {
> log_warnx("%s: error writing viornd to fd", __func__);
> return (-1);
> @@ -2205,6 +2247,7 @@ int
> vmmci_dump(int fd)
> {
> log_debug("%s: sending vmmci", __func__);
> +
> if (atomicio(vwrite, fd, &vmmci, sizeof(vmmci)) != sizeof(vmmci)) {
> log_warnx("%s: error writing vmmci to fd", __func__);
> return (-1);
> @@ -2215,7 +2258,15 @@ vmmci_dump(int fd)
> int
> vionet_dump(int fd)
> {
> + int i;
> +
> log_debug("%s: sending vionet", __func__);
> +
> + for (i = 0; i < nr_vionet; i++) {
> + vionet[i].vq[RXQ].q_hva = NULL;
> + vionet[i].vq[TXQ].q_hva = NULL;
> + }
> +
> if (atomicio(vwrite, fd, vionet,
> nr_vionet * sizeof(struct vionet_dev)) !=
> nr_vionet * sizeof(struct vionet_dev)) {
> @@ -2228,7 +2279,13 @@ vionet_dump(int fd)
> int
> vioblk_dump(int fd)
> {
> + int i;
> +
> log_debug("%s: sending vioblk", __func__);
> +
> + for (i = 0; i < nr_vioblk; i++)
> + vioblk[i].vq[0].q_hva = NULL;
> +
> if (atomicio(vwrite, fd, vioblk,
> nr_vioblk * sizeof(struct vioblk_dev)) !=
> nr_vioblk * sizeof(struct vioblk_dev)) {
> @@ -2241,10 +2298,16 @@ vioblk_dump(int fd)
> int
> vioscsi_dump(int fd)
> {
> + unsigned int i;
> +
> if (vioscsi == NULL)
> return (0);
>
> log_debug("%s: sending vioscsi", __func__);
> +
> + for (i = 0; i < 3; i++)
> + vioscsi->vq[i].q_hva = NULL;
> +
> if (atomicio(vwrite, fd, vioscsi, sizeof(struct vioscsi_dev)) !=
> sizeof(struct vioscsi_dev)) {
> log_warnx("%s: error writing vioscsi to fd", __func__);