Hi Liu, Thank you so much for your reviewing. I will fix them, then submit again in this week.
Thanks, Tetsuya On 2015/11/09 15:21, Yuanhan Liu wrote: > Hi Tetsuya, > > Here I just got some minor nits after a very rough glimpse. > > On Mon, Nov 09, 2015 at 02:17:01PM +0900, Tetsuya Mukawa wrote: > ... >> +static uint16_t >> +eth_vhost_rx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs) >> +{ >> + struct vhost_queue *r = q; >> + uint16_t nb_rx = 0; >> + >> + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0)) >> + return 0; >> + >> + rte_atomic32_set(&r->while_queuing, 1); >> + >> + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0)) >> + goto out; >> + >> + /* Dequeue packets from guest TX queue */ >> + nb_rx = (uint16_t)rte_vhost_dequeue_burst(r->device, >> + r->virtqueue_id, r->mb_pool, bufs, nb_bufs); > Unnecessary cast, as rte_vhost_enqueue_burst is defined with uint16_t > return type. > >> + >> + r->rx_pkts += nb_rx; >> + >> +out: >> + rte_atomic32_set(&r->while_queuing, 0); >> + >> + return nb_rx; >> +} >> + >> +static uint16_t >> +eth_vhost_tx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs) >> +{ >> + struct vhost_queue *r = q; >> + uint16_t i, nb_tx = 0; >> + >> + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0)) >> + return 0; >> + >> + rte_atomic32_set(&r->while_queuing, 1); >> + >> + if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0)) >> + goto out; >> + >> + /* Enqueue packets to guest RX queue */ >> + nb_tx = (uint16_t)rte_vhost_enqueue_burst(r->device, >> + r->virtqueue_id, bufs, nb_bufs); > Ditto. > >> + >> + r->tx_pkts += nb_tx; >> + r->err_pkts += nb_bufs - nb_tx; >> + >> + for (i = 0; likely(i < nb_tx); i++) >> + rte_pktmbuf_free(bufs[i]); >> + >> +out: >> + rte_atomic32_set(&r->while_queuing, 0); >> + >> + return nb_tx; >> +} >> + >> +static int >> +eth_dev_configure(struct rte_eth_dev *dev __rte_unused) { return 0; } > I personally would not prefer to saving few lines of code to sacrifice > the readability. > >> + >> +static int >> +eth_rx_queue_setup(struct rte_eth_dev *dev, uint16_t rx_queue_id, >> + uint16_t nb_rx_desc __rte_unused, >> + unsigned int socket_id, >> + const struct rte_eth_rxconf *rx_conf __rte_unused, >> + struct rte_mempool *mb_pool) >> +{ >> + struct pmd_internal *internal = dev->data->dev_private; >> + struct vhost_queue *vq; >> + >> + if (internal->rx_vhost_queues[rx_queue_id] != NULL) >> + rte_free(internal->rx_vhost_queues[rx_queue_id]); > Such NULL check is unnecessary; rte_free will handle it. > >> + >> + vq = rte_zmalloc_socket(NULL, sizeof(struct vhost_queue), >> + RTE_CACHE_LINE_SIZE, socket_id); >> + if (vq == NULL) { >> + RTE_LOG(ERR, PMD, "Failed to allocate memory for rx queue\n"); >> + return -ENOMEM; >> + } >> + >> + vq->mb_pool = mb_pool; >> + vq->virtqueue_id = rx_queue_id * VIRTIO_QNUM + VIRTIO_TXQ; >> + internal->rx_vhost_queues[rx_queue_id] = vq; >> + dev->data->rx_queues[rx_queue_id] = vq; >> + return 0; >> +} >> + >> +static int >> +eth_tx_queue_setup(struct rte_eth_dev *dev, uint16_t tx_queue_id, >> + uint16_t nb_tx_desc __rte_unused, >> + unsigned int socket_id, >> + const struct rte_eth_txconf *tx_conf __rte_unused) >> +{ >> + struct pmd_internal *internal = dev->data->dev_private; >> + struct vhost_queue *vq; >> + >> + if (internal->tx_vhost_queues[tx_queue_id] != NULL) >> + rte_free(internal->tx_vhost_queues[tx_queue_id]); > Ditto. > >> + >> + vq = rte_zmalloc_socket(NULL, sizeof(struct vhost_queue), >> + RTE_CACHE_LINE_SIZE, socket_id); >> + if (vq == NULL) { >> + RTE_LOG(ERR, PMD, "Failed to allocate memory for tx queue\n"); >> + return -ENOMEM; >> + } >> + >> + vq->virtqueue_id = tx_queue_id * VIRTIO_QNUM + VIRTIO_RXQ; >> + internal->tx_vhost_queues[tx_queue_id] = vq; >> + dev->data->tx_queues[tx_queue_id] = vq; >> + return 0; >> +} >> + >> + >> +static void >> +eth_dev_info(struct rte_eth_dev *dev, >> + struct rte_eth_dev_info *dev_info) >> +{ >> + struct pmd_internal *internal = dev->data->dev_private; >> + >> + dev_info->driver_name = drivername; >> + dev_info->max_mac_addrs = 1; >> + dev_info->max_rx_pktlen = (uint32_t)-1; >> + dev_info->max_rx_queues = (uint16_t)internal->nb_rx_queues; >> + dev_info->max_tx_queues = (uint16_t)internal->nb_tx_queues; >> + dev_info->min_rx_bufsize = 0; >> +} >> + >> +static void >> +eth_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *igb_stats) >> +{ >> + unsigned i; >> + unsigned long rx_total = 0, tx_total = 0, tx_err_total = 0; >> + const struct pmd_internal *internal = dev->data->dev_private; >> + >> + for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS && >> + i < internal->nb_rx_queues; i++) { >> + if (internal->rx_vhost_queues[i] == NULL) >> + continue; >> + igb_stats->q_ipackets[i] = >> internal->rx_vhost_queues[i]->rx_pkts; >> + rx_total += igb_stats->q_ipackets[i]; >> + } >> + >> + for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS && >> + i < internal->nb_tx_queues; i++) { >> + if (internal->tx_vhost_queues[i] == NULL) >> + continue; >> + igb_stats->q_opackets[i] = >> internal->tx_vhost_queues[i]->tx_pkts; >> + igb_stats->q_errors[i] = internal->tx_vhost_queues[i]->err_pkts; >> + tx_total += igb_stats->q_opackets[i]; >> + tx_err_total += igb_stats->q_errors[i]; >> + } >> + >> + igb_stats->ipackets = rx_total; >> + igb_stats->opackets = tx_total; >> + igb_stats->oerrors = tx_err_total; >> +} >> + >> +static void >> +eth_stats_reset(struct rte_eth_dev *dev) >> +{ >> + unsigned i; >> + struct pmd_internal *internal = dev->data->dev_private; >> + >> + for (i = 0; i < internal->nb_rx_queues; i++) { >> + if (internal->rx_vhost_queues[i] == NULL) >> + continue; >> + internal->rx_vhost_queues[i]->rx_pkts = 0; >> + } >> + for (i = 0; i < internal->nb_tx_queues; i++) { >> + if (internal->tx_vhost_queues[i] == NULL) >> + continue; >> + internal->tx_vhost_queues[i]->tx_pkts = 0; >> + internal->tx_vhost_queues[i]->err_pkts = 0; >> + } >> +} >> + >> +static void >> +eth_queue_release(void *q __rte_unused) { ; } >> +static int >> +eth_link_update(struct rte_eth_dev *dev __rte_unused, >> + int wait_to_complete __rte_unused) { return 0; } > Ditto. > >> + >> +static const struct eth_dev_ops ops = { >> + .dev_start = eth_dev_start, >> + .dev_stop = eth_dev_stop, >> + .dev_configure = eth_dev_configure, >> + .dev_infos_get = eth_dev_info, >> + .rx_queue_setup = eth_rx_queue_setup, >> + .tx_queue_setup = eth_tx_queue_setup, >> + .rx_queue_release = eth_queue_release, >> + .tx_queue_release = eth_queue_release, >> + .link_update = eth_link_update, >> + .stats_get = eth_stats_get, >> + .stats_reset = eth_stats_reset, >> +}; >> + >> +static int >> +eth_dev_vhost_create(const char *name, int index, >> + char *iface_name, >> + int16_t queues, >> + const unsigned numa_node) >> +{ >> + struct rte_eth_dev_data *data = NULL; >> + struct pmd_internal *internal = NULL; >> + struct rte_eth_dev *eth_dev = NULL; >> + struct ether_addr *eth_addr = NULL; >> + >> + RTE_LOG(INFO, PMD, "Creating VHOST-USER backend on numa socket %u\n", >> + numa_node); >> + >> + /* now do all data allocation - for eth_dev structure, dummy pci driver >> + * and internal (private) data >> + */ >> + data = rte_zmalloc_socket(name, sizeof(*data), 0, numa_node); >> + if (data == NULL) >> + goto error; >> + >> + internal = rte_zmalloc_socket(name, sizeof(*internal), 0, numa_node); >> + if (internal == NULL) >> + goto error; >> + >> + eth_addr = rte_zmalloc_socket(name, sizeof(*eth_addr), 0, numa_node); >> + if (eth_addr == NULL) >> + goto error; >> + *eth_addr = base_eth_addr; >> + eth_addr->addr_bytes[5] = index; >> + >> + /* reserve an ethdev entry */ >> + eth_dev = rte_eth_dev_allocate(name, RTE_ETH_DEV_VIRTUAL); >> + if (eth_dev == NULL) >> + goto error; >> + >> + /* now put it all together >> + * - store queue data in internal, >> + * - store numa_node info in ethdev data >> + * - point eth_dev_data to internals >> + * - and point eth_dev structure to new eth_dev_data structure >> + */ >> + internal->nb_rx_queues = queues; >> + internal->nb_tx_queues = queues; >> + internal->dev_name = strdup(name); >> + if (internal->dev_name == NULL) >> + goto error; >> + internal->iface_name = strdup(iface_name); >> + if (internal->iface_name == NULL) >> + goto error; > If allocation failed here, you will find that internal->dev_name is not > freed. > >> + >> + pthread_mutex_lock(&internal_list_lock); >> + TAILQ_INSERT_TAIL(&internals_list, internal, next); >> + pthread_mutex_unlock(&internal_list_lock); >> + >> + data->dev_private = internal; >> + data->port_id = eth_dev->data->port_id; >> + memmove(data->name, eth_dev->data->name, sizeof(data->name)); >> + data->nb_rx_queues = queues; >> + data->nb_tx_queues = queues; >> + data->dev_link = pmd_link; >> + data->mac_addrs = eth_addr; >> + >> + /* We'll replace the 'data' originally allocated by eth_dev. So the >> + * vhost PMD resources won't be shared between multi processes. >> + */ >> + eth_dev->data = data; >> + eth_dev->dev_ops = &ops; >> + eth_dev->driver = NULL; >> + eth_dev->data->dev_flags = RTE_ETH_DEV_DETACHABLE; >> + eth_dev->data->kdrv = RTE_KDRV_NONE; >> + eth_dev->data->drv_name = internal->dev_name; >> + eth_dev->data->numa_node = numa_node; >> + >> + /* finally assign rx and tx ops */ >> + eth_dev->rx_pkt_burst = eth_vhost_rx; >> + eth_dev->tx_pkt_burst = eth_vhost_tx; >> + >> + return data->port_id; >> + >> +error: >> + rte_free(data); >> + rte_free(internal); >> + rte_free(eth_addr); >> + >> + return -1; >> +} > ... > ... >> + >> + if ((internal) && (internal->dev_name)) >> + free(internal->dev_name); >> + if ((internal) && (internal->iface_name)) >> + free(internal->iface_name); >> + >> + rte_free(eth_dev->data->mac_addrs); >> + rte_free(eth_dev->data); >> + >> + for (i = 0; i < internal->nb_rx_queues; i++) { >> + if (internal->rx_vhost_queues[i] != NULL) >> + rte_free(internal->rx_vhost_queues[i]); >> + } >> + for (i = 0; i < internal->nb_tx_queues; i++) { >> + if (internal->tx_vhost_queues[i] != NULL) >> + rte_free(internal->tx_vhost_queues[i]); > Ditto. > > (Hopefully I could have a detailed review later, say next week). > > --yliu