net/virtio: add vectorized packed ring Tx

Optimize packed ring Tx path like Rx path. Split Tx path into batch and
single Tx functions. Batch function is further optimized by AVX512
instructions.

Signed-off-by: Marvin Liu <yong.liu@intel.com>
Reviewed-by: Maxime Coquelin <maxime.coquelin@redhat.com>
This commit is contained in:
Marvin Liu 2020-04-29 15:28:20 +08:00 committed by Ferruh Yigit
parent 77d66da838
commit 6494b9a9b2
4 changed files with 246 additions and 0 deletions

View File

@ -505,6 +505,7 @@ are shown in below table:
Packed virtqueue in-order mergeable path virtio_recv_mergeable_pkts_packed virtio_xmit_pkts_packed
Packed virtqueue in-order non-mergeable path virtio_recv_pkts_packed virtio_xmit_pkts_packed
Packed virtqueue vectorized Rx path virtio_recv_pkts_packed_vec virtio_xmit_pkts_packed
Packed virtqueue vectorized Tx path virtio_recv_pkts_packed virtio_xmit_pkts_packed_vec
============================================ ================================= ========================
Virtio paths Support Status from Release to Release
@ -536,6 +537,7 @@ All virtio paths support status are shown in below table:
Packed virtqueue in-order mergeable path Y Y
Packed virtqueue in-order non-mergeable path Y Y
Packed virtqueue vectorized Rx path Y
Packed virtqueue vectorized Tx path Y
============================================ ============= ============= ============= =======
QEMU Support Status

View File

@ -108,6 +108,9 @@ uint16_t virtio_recv_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts,
uint16_t virtio_recv_pkts_packed_vec(void *rx_queue, struct rte_mbuf **rx_pkts,
uint16_t nb_pkts);
uint16_t virtio_xmit_pkts_packed_vec(void *tx_queue, struct rte_mbuf **tx_pkts,
uint16_t nb_pkts);
int eth_virtio_dev_init(struct rte_eth_dev *eth_dev);
void virtio_interrupt_handler(void *param);

View File

@ -2039,4 +2039,12 @@ virtio_recv_pkts_packed_vec(void *rx_queue __rte_unused,
{
return 0;
}
uint16_t
virtio_xmit_pkts_packed_vec(void *tx_queue __rte_unused,
struct rte_mbuf **tx_pkts __rte_unused,
uint16_t nb_pkts __rte_unused)
{
return 0;
}
#endif /* ifndef CC_AVX512_SUPPORT */

View File

@ -23,6 +23,24 @@
#define PACKED_FLAGS_MASK ((0ULL | VRING_PACKED_DESC_F_AVAIL_USED) << \
FLAGS_BITS_OFFSET)
/* reference count offset in mbuf rearm data */
#define REFCNT_BITS_OFFSET ((offsetof(struct rte_mbuf, refcnt) - \
offsetof(struct rte_mbuf, rearm_data)) * BYTE_SIZE)
/* segment number offset in mbuf rearm data */
#define SEG_NUM_BITS_OFFSET ((offsetof(struct rte_mbuf, nb_segs) - \
offsetof(struct rte_mbuf, rearm_data)) * BYTE_SIZE)
/* default rearm data */
#define DEFAULT_REARM_DATA (1ULL << SEG_NUM_BITS_OFFSET | \
1ULL << REFCNT_BITS_OFFSET)
/* id bits offset in packed ring desc higher 64bits */
#define ID_BITS_OFFSET ((offsetof(struct vring_packed_desc, id) - \
offsetof(struct vring_packed_desc, len)) * BYTE_SIZE)
/* net hdr short size mask */
#define NET_HDR_MASK 0x3F
#define PACKED_BATCH_SIZE (RTE_CACHE_LINE_SIZE / \
sizeof(struct vring_packed_desc))
#define PACKED_BATCH_MASK (PACKED_BATCH_SIZE - 1)
@ -60,6 +78,221 @@ virtio_update_batch_stats(struct virtnet_stats *stats,
stats->bytes += pkt_len4;
}
static inline int
virtqueue_enqueue_batch_packed_vec(struct virtnet_tx *txvq,
struct rte_mbuf **tx_pkts)
{
struct virtqueue *vq = txvq->vq;
uint16_t head_size = vq->hw->vtnet_hdr_size;
uint16_t idx = vq->vq_avail_idx;
struct virtio_net_hdr *hdr;
struct vq_desc_extra *dxp;
uint16_t i, cmp;
if (vq->vq_avail_idx & PACKED_BATCH_MASK)
return -1;
if (unlikely((idx + PACKED_BATCH_SIZE) > vq->vq_nentries))
return -1;
/* Load four mbufs rearm data */
RTE_BUILD_BUG_ON(REFCNT_BITS_OFFSET >= 64);
RTE_BUILD_BUG_ON(SEG_NUM_BITS_OFFSET >= 64);
__m256i mbufs = _mm256_set_epi64x(*tx_pkts[3]->rearm_data,
*tx_pkts[2]->rearm_data,
*tx_pkts[1]->rearm_data,
*tx_pkts[0]->rearm_data);
/* refcnt=1 and nb_segs=1 */
__m256i mbuf_ref = _mm256_set1_epi64x(DEFAULT_REARM_DATA);
__m256i head_rooms = _mm256_set1_epi16(head_size);
/* Check refcnt and nb_segs */
const __mmask16 mask = 0x6 | 0x6 << 4 | 0x6 << 8 | 0x6 << 12;
cmp = _mm256_mask_cmpneq_epu16_mask(mask, mbufs, mbuf_ref);
if (unlikely(cmp))
return -1;
/* Check headroom is enough */
const __mmask16 data_mask = 0x1 | 0x1 << 4 | 0x1 << 8 | 0x1 << 12;
RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, data_off) !=
offsetof(struct rte_mbuf, rearm_data));
cmp = _mm256_mask_cmplt_epu16_mask(data_mask, mbufs, head_rooms);
if (unlikely(cmp))
return -1;
virtio_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
dxp = &vq->vq_descx[idx + i];
dxp->ndescs = 1;
dxp->cookie = tx_pkts[i];
}
virtio_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
tx_pkts[i]->data_off -= head_size;
tx_pkts[i]->data_len += head_size;
}
__m512i descs_base = _mm512_set_epi64(tx_pkts[3]->data_len,
VIRTIO_MBUF_ADDR(tx_pkts[3], vq),
tx_pkts[2]->data_len,
VIRTIO_MBUF_ADDR(tx_pkts[2], vq),
tx_pkts[1]->data_len,
VIRTIO_MBUF_ADDR(tx_pkts[1], vq),
tx_pkts[0]->data_len,
VIRTIO_MBUF_ADDR(tx_pkts[0], vq));
/* id offset and data offset */
__m512i data_offsets = _mm512_set_epi64((uint64_t)3 << ID_BITS_OFFSET,
tx_pkts[3]->data_off,
(uint64_t)2 << ID_BITS_OFFSET,
tx_pkts[2]->data_off,
(uint64_t)1 << ID_BITS_OFFSET,
tx_pkts[1]->data_off,
0, tx_pkts[0]->data_off);
__m512i new_descs = _mm512_add_epi64(descs_base, data_offsets);
uint64_t flags_temp = (uint64_t)idx << ID_BITS_OFFSET |
(uint64_t)vq->vq_packed.cached_flags << FLAGS_BITS_OFFSET;
/* flags offset and guest virtual address offset */
__m128i flag_offset = _mm_set_epi64x(flags_temp, 0);
__m512i v_offset = _mm512_broadcast_i32x4(flag_offset);
__m512i v_desc = _mm512_add_epi64(new_descs, v_offset);
if (!vq->hw->has_tx_offload) {
__m128i all_mask = _mm_set1_epi16(0xFFFF);
virtio_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
hdr = rte_pktmbuf_mtod_offset(tx_pkts[i],
struct virtio_net_hdr *, -head_size);
__m128i v_hdr = _mm_loadu_si128((void *)hdr);
if (unlikely(_mm_mask_test_epi16_mask(NET_HDR_MASK,
v_hdr, all_mask))) {
__m128i all_zero = _mm_setzero_si128();
_mm_mask_storeu_epi16((void *)hdr,
NET_HDR_MASK, all_zero);
}
}
} else {
virtio_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
hdr = rte_pktmbuf_mtod_offset(tx_pkts[i],
struct virtio_net_hdr *, -head_size);
virtqueue_xmit_offload(hdr, tx_pkts[i], true);
}
}
/* Enqueue Packet buffers */
_mm512_storeu_si512((void *)&vq->vq_packed.ring.desc[idx], v_desc);
virtio_update_batch_stats(&txvq->stats, tx_pkts[0]->pkt_len,
tx_pkts[1]->pkt_len, tx_pkts[2]->pkt_len,
tx_pkts[3]->pkt_len);
vq->vq_avail_idx += PACKED_BATCH_SIZE;
vq->vq_free_cnt -= PACKED_BATCH_SIZE;
if (vq->vq_avail_idx >= vq->vq_nentries) {
vq->vq_avail_idx -= vq->vq_nentries;
vq->vq_packed.cached_flags ^=
VRING_PACKED_DESC_F_AVAIL_USED;
}
return 0;
}
static inline int
virtqueue_enqueue_single_packed_vec(struct virtnet_tx *txvq,
struct rte_mbuf *txm)
{
struct virtqueue *vq = txvq->vq;
struct virtio_hw *hw = vq->hw;
uint16_t hdr_size = hw->vtnet_hdr_size;
uint16_t slots, can_push;
int16_t need;
/* How many main ring entries are needed to this Tx?
* any_layout => number of segments
* default => number of segments + 1
*/
can_push = rte_mbuf_refcnt_read(txm) == 1 &&
RTE_MBUF_DIRECT(txm) &&
txm->nb_segs == 1 &&
rte_pktmbuf_headroom(txm) >= hdr_size;
slots = txm->nb_segs + !can_push;
need = slots - vq->vq_free_cnt;
/* Positive value indicates it need free vring descriptors */
if (unlikely(need > 0)) {
virtio_xmit_cleanup_inorder_packed(vq, need);
need = slots - vq->vq_free_cnt;
if (unlikely(need > 0)) {
PMD_TX_LOG(ERR,
"No free tx descriptors to transmit");
return -1;
}
}
/* Enqueue Packet buffers */
virtqueue_enqueue_xmit_packed(txvq, txm, slots, can_push, 1);
txvq->stats.bytes += txm->pkt_len;
return 0;
}
uint16_t
virtio_xmit_pkts_packed_vec(void *tx_queue, struct rte_mbuf **tx_pkts,
uint16_t nb_pkts)
{
struct virtnet_tx *txvq = tx_queue;
struct virtqueue *vq = txvq->vq;
struct virtio_hw *hw = vq->hw;
uint16_t nb_tx = 0;
uint16_t remained;
if (unlikely(hw->started == 0 && tx_pkts != hw->inject_pkts))
return nb_tx;
if (unlikely(nb_pkts < 1))
return nb_pkts;
PMD_TX_LOG(DEBUG, "%d packets to xmit", nb_pkts);
if (vq->vq_free_cnt <= vq->vq_nentries - vq->vq_free_thresh)
virtio_xmit_cleanup_inorder_packed(vq, vq->vq_free_thresh);
remained = RTE_MIN(nb_pkts, vq->vq_free_cnt);
while (remained) {
if (remained >= PACKED_BATCH_SIZE) {
if (!virtqueue_enqueue_batch_packed_vec(txvq,
&tx_pkts[nb_tx])) {
nb_tx += PACKED_BATCH_SIZE;
remained -= PACKED_BATCH_SIZE;
continue;
}
}
if (!virtqueue_enqueue_single_packed_vec(txvq,
tx_pkts[nb_tx])) {
nb_tx++;
remained--;
continue;
}
break;
};
txvq->stats.packets += nb_tx;
if (likely(nb_tx)) {
if (unlikely(virtqueue_kick_prepare_packed(vq))) {
virtqueue_notify(vq);
PMD_TX_LOG(DEBUG, "Notified backend after xmit");
}
}
return nb_tx;
}
/* Optionally fill offload information in structure */
static inline int
virtio_vec_rx_offload(struct rte_mbuf *m, struct virtio_net_hdr *hdr)