net/virtio: fix resuming port with Rx vector path
Since commit efc83a1e7fc3 ("net/virtio: fix queue setup consistency"), when resuming a virtio port, the rx rings are refilled with new mbufs until they are full (vq->vq_free_cnt == 0). This is done without ensuring that the descriptor index remains a multiple of RTE_VIRTIO_VPMD_RX_REARM_THRESH, which is a prerequisite when using the vector mode. This can cause an out of bound access in the rx ring. This commit changes the vector refill method from virtqueue_enqueue_recv_refill_simple() to virtio_rxq_rearm_vec(), which properly checks that the refill is done by batch of RTE_VIRTIO_VPMD_RX_REARM_THRESH. As virtqueue_enqueue_recv_refill_simple() is no more used, this patch also removes the function. Fixes: efc83a1e7fc3 ("net/virtio: fix queue setup consistency") Cc: stable@dpdk.org Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com> Signed-off-by: Olivier Matz <olivier.matz@6wind.com> Signed-off-by: Tiwei Bie <tiwei.bie@intel.com> Reviewed-by: Jianfeng Tan <jianfeng.tan@intel.com>
This commit is contained in:
parent
9fedc2da38
commit
4785747066
@ -30,6 +30,7 @@
|
|||||||
#include "virtio_pci.h"
|
#include "virtio_pci.h"
|
||||||
#include "virtqueue.h"
|
#include "virtqueue.h"
|
||||||
#include "virtio_rxtx.h"
|
#include "virtio_rxtx.h"
|
||||||
|
#include "virtio_rxtx_simple.h"
|
||||||
|
|
||||||
#ifdef RTE_LIBRTE_VIRTIO_DEBUG_DUMP
|
#ifdef RTE_LIBRTE_VIRTIO_DEBUG_DUMP
|
||||||
#define VIRTIO_DUMP_PACKET(m, len) rte_pktmbuf_dump(stdout, m, len)
|
#define VIRTIO_DUMP_PACKET(m, len) rte_pktmbuf_dump(stdout, m, len)
|
||||||
@ -437,6 +438,8 @@ virtio_dev_rx_queue_setup_finish(struct rte_eth_dev *dev, uint16_t queue_idx)
|
|||||||
vq->vq_ring.desc[desc_idx].flags =
|
vq->vq_ring.desc[desc_idx].flags =
|
||||||
VRING_DESC_F_WRITE;
|
VRING_DESC_F_WRITE;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
virtio_rxq_vec_setup(rxvq);
|
||||||
}
|
}
|
||||||
|
|
||||||
memset(&rxvq->fake_mbuf, 0, sizeof(rxvq->fake_mbuf));
|
memset(&rxvq->fake_mbuf, 0, sizeof(rxvq->fake_mbuf));
|
||||||
@ -446,30 +449,31 @@ virtio_dev_rx_queue_setup_finish(struct rte_eth_dev *dev, uint16_t queue_idx)
|
|||||||
&rxvq->fake_mbuf;
|
&rxvq->fake_mbuf;
|
||||||
}
|
}
|
||||||
|
|
||||||
while (!virtqueue_full(vq)) {
|
if (hw->use_simple_rx) {
|
||||||
m = rte_mbuf_raw_alloc(rxvq->mpool);
|
while (vq->vq_free_cnt >= RTE_VIRTIO_VPMD_RX_REARM_THRESH) {
|
||||||
if (m == NULL)
|
virtio_rxq_rearm_vec(rxvq);
|
||||||
break;
|
nbufs += RTE_VIRTIO_VPMD_RX_REARM_THRESH;
|
||||||
|
|
||||||
/* Enqueue allocated buffers */
|
|
||||||
if (hw->use_simple_rx)
|
|
||||||
error = virtqueue_enqueue_recv_refill_simple(vq, m);
|
|
||||||
else
|
|
||||||
error = virtqueue_enqueue_recv_refill(vq, m);
|
|
||||||
|
|
||||||
if (error) {
|
|
||||||
rte_pktmbuf_free(m);
|
|
||||||
break;
|
|
||||||
}
|
}
|
||||||
nbufs++;
|
} else {
|
||||||
|
while (!virtqueue_full(vq)) {
|
||||||
|
m = rte_mbuf_raw_alloc(rxvq->mpool);
|
||||||
|
if (m == NULL)
|
||||||
|
break;
|
||||||
|
|
||||||
|
/* Enqueue allocated buffers */
|
||||||
|
error = virtqueue_enqueue_recv_refill(vq, m);
|
||||||
|
if (error) {
|
||||||
|
rte_pktmbuf_free(m);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
nbufs++;
|
||||||
|
}
|
||||||
|
|
||||||
|
vq_update_avail_idx(vq);
|
||||||
}
|
}
|
||||||
|
|
||||||
vq_update_avail_idx(vq);
|
|
||||||
|
|
||||||
PMD_INIT_LOG(DEBUG, "Allocated %d bufs", nbufs);
|
PMD_INIT_LOG(DEBUG, "Allocated %d bufs", nbufs);
|
||||||
|
|
||||||
virtio_rxq_vec_setup(rxvq);
|
|
||||||
|
|
||||||
VIRTQUEUE_DUMP(vq);
|
VIRTQUEUE_DUMP(vq);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -60,7 +60,4 @@ struct virtnet_ctl {
|
|||||||
|
|
||||||
int virtio_rxq_vec_setup(struct virtnet_rx *rxvq);
|
int virtio_rxq_vec_setup(struct virtnet_rx *rxvq);
|
||||||
|
|
||||||
int virtqueue_enqueue_recv_refill_simple(struct virtqueue *vq,
|
|
||||||
struct rte_mbuf *m);
|
|
||||||
|
|
||||||
#endif /* _VIRTIO_RXTX_H_ */
|
#endif /* _VIRTIO_RXTX_H_ */
|
||||||
|
@ -27,35 +27,6 @@
|
|||||||
#pragma GCC diagnostic ignored "-Wcast-qual"
|
#pragma GCC diagnostic ignored "-Wcast-qual"
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
int __attribute__((cold))
|
|
||||||
virtqueue_enqueue_recv_refill_simple(struct virtqueue *vq,
|
|
||||||
struct rte_mbuf *cookie)
|
|
||||||
{
|
|
||||||
struct vq_desc_extra *dxp;
|
|
||||||
struct vring_desc *start_dp;
|
|
||||||
uint16_t desc_idx;
|
|
||||||
|
|
||||||
cookie->port = vq->rxq.port_id;
|
|
||||||
cookie->data_off = RTE_PKTMBUF_HEADROOM;
|
|
||||||
|
|
||||||
desc_idx = vq->vq_avail_idx & (vq->vq_nentries - 1);
|
|
||||||
dxp = &vq->vq_descx[desc_idx];
|
|
||||||
dxp->cookie = (void *)cookie;
|
|
||||||
vq->sw_ring[desc_idx] = cookie;
|
|
||||||
|
|
||||||
start_dp = vq->vq_ring.desc;
|
|
||||||
start_dp[desc_idx].addr =
|
|
||||||
VIRTIO_MBUF_ADDR(cookie, vq) +
|
|
||||||
RTE_PKTMBUF_HEADROOM - vq->hw->vtnet_hdr_size;
|
|
||||||
start_dp[desc_idx].len = cookie->buf_len -
|
|
||||||
RTE_PKTMBUF_HEADROOM + vq->hw->vtnet_hdr_size;
|
|
||||||
|
|
||||||
vq->vq_free_cnt--;
|
|
||||||
vq->vq_avail_idx++;
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
uint16_t
|
uint16_t
|
||||||
virtio_xmit_pkts_simple(void *tx_queue, struct rte_mbuf **tx_pkts,
|
virtio_xmit_pkts_simple(void *tx_queue, struct rte_mbuf **tx_pkts,
|
||||||
uint16_t nb_pkts)
|
uint16_t nb_pkts)
|
||||||
@ -78,7 +49,7 @@ virtio_xmit_pkts_simple(void *tx_queue, struct rte_mbuf **tx_pkts,
|
|||||||
rte_compiler_barrier();
|
rte_compiler_barrier();
|
||||||
|
|
||||||
if (nb_used >= VIRTIO_TX_FREE_THRESH)
|
if (nb_used >= VIRTIO_TX_FREE_THRESH)
|
||||||
virtio_xmit_cleanup(vq);
|
virtio_xmit_cleanup_simple(vq);
|
||||||
|
|
||||||
nb_commit = nb_pkts = RTE_MIN((vq->vq_free_cnt >> 1), nb_pkts);
|
nb_commit = nb_pkts = RTE_MIN((vq->vq_free_cnt >> 1), nb_pkts);
|
||||||
desc_idx = (uint16_t)(vq->vq_avail_idx & desc_idx_max);
|
desc_idx = (uint16_t)(vq->vq_avail_idx & desc_idx_max);
|
||||||
|
@ -60,7 +60,7 @@ virtio_rxq_rearm_vec(struct virtnet_rx *rxvq)
|
|||||||
#define VIRTIO_TX_FREE_NR 32
|
#define VIRTIO_TX_FREE_NR 32
|
||||||
/* TODO: vq->tx_free_cnt could mean num of free slots so we could avoid shift */
|
/* TODO: vq->tx_free_cnt could mean num of free slots so we could avoid shift */
|
||||||
static inline void
|
static inline void
|
||||||
virtio_xmit_cleanup(struct virtqueue *vq)
|
virtio_xmit_cleanup_simple(struct virtqueue *vq)
|
||||||
{
|
{
|
||||||
uint16_t i, desc_idx;
|
uint16_t i, desc_idx;
|
||||||
uint32_t nb_free = 0;
|
uint32_t nb_free = 0;
|
||||||
|
Loading…
x
Reference in New Issue
Block a user