numam-dpdk/drivers/net/netvsc/hn_rxtx.c
Stephen Hemminger 530af95a78 bus/vmbus: avoid signalling host on read
Don't signal host that receive ring has been read until all events
have been processed. This reduces the number of guest exits and
therefore improves performance.

Signed-off-by: Stephen Hemminger <sthemmin@microsoft.com>
2018-08-05 11:03:18 +02:00

1326 lines
31 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright(c) 2016-2018 Microsoft Corporation
* Copyright(c) 2013-2016 Brocade Communications Systems, Inc.
* All rights reserved.
*/
#include <stdint.h>
#include <string.h>
#include <stdio.h>
#include <errno.h>
#include <unistd.h>
#include <strings.h>
#include <rte_ethdev.h>
#include <rte_memcpy.h>
#include <rte_string_fns.h>
#include <rte_memzone.h>
#include <rte_malloc.h>
#include <rte_atomic.h>
#include <rte_branch_prediction.h>
#include <rte_ether.h>
#include <rte_common.h>
#include <rte_errno.h>
#include <rte_memory.h>
#include <rte_eal.h>
#include <rte_dev.h>
#include <rte_bus_vmbus.h>
#include <rte_spinlock.h>
#include "hn_logs.h"
#include "hn_var.h"
#include "hn_rndis.h"
#include "hn_nvs.h"
#include "ndis.h"
#define HN_NVS_SEND_MSG_SIZE \
(sizeof(struct vmbus_chanpkt_hdr) + sizeof(struct hn_nvs_rndis))
#define HN_TXD_CACHE_SIZE 32 /* per cpu tx_descriptor pool cache */
#define HN_TXCOPY_THRESHOLD 512
#define HN_RXCOPY_THRESHOLD 256
#define HN_RXQ_EVENT_DEFAULT 2048
struct hn_rxinfo {
uint32_t vlan_info;
uint32_t csum_info;
uint32_t hash_info;
uint32_t hash_value;
};
#define HN_RXINFO_VLAN 0x0001
#define HN_RXINFO_CSUM 0x0002
#define HN_RXINFO_HASHINF 0x0004
#define HN_RXINFO_HASHVAL 0x0008
#define HN_RXINFO_ALL \
(HN_RXINFO_VLAN | \
HN_RXINFO_CSUM | \
HN_RXINFO_HASHINF | \
HN_RXINFO_HASHVAL)
#define HN_NDIS_VLAN_INFO_INVALID 0xffffffff
#define HN_NDIS_RXCSUM_INFO_INVALID 0
#define HN_NDIS_HASH_INFO_INVALID 0
/*
* Per-transmit book keeping.
* A slot in transmit ring (chim_index) is reserved for each transmit.
*
* There are two types of transmit:
* - buffered transmit where chimney buffer is used and RNDIS header
* is in the buffer. mbuf == NULL for this case.
*
* - direct transmit where RNDIS header is in the in rndis_pkt
* mbuf is freed after transmit.
*
* Descriptors come from per-port pool which is used
* to limit number of outstanding requests per device.
*/
struct hn_txdesc {
struct rte_mbuf *m;
uint16_t queue_id;
uint16_t chim_index;
uint32_t chim_size;
uint32_t data_size;
uint32_t packets;
struct rndis_packet_msg *rndis_pkt;
};
#define HN_RNDIS_PKT_LEN \
(sizeof(struct rndis_packet_msg) + \
RNDIS_PKTINFO_SIZE(NDIS_HASH_VALUE_SIZE) + \
RNDIS_PKTINFO_SIZE(NDIS_VLAN_INFO_SIZE) + \
RNDIS_PKTINFO_SIZE(NDIS_LSO2_INFO_SIZE) + \
RNDIS_PKTINFO_SIZE(NDIS_TXCSUM_INFO_SIZE))
/* Minimum space required for a packet */
#define HN_PKTSIZE_MIN(align) \
RTE_ALIGN(ETHER_MIN_LEN + HN_RNDIS_PKT_LEN, align)
#define DEFAULT_TX_FREE_THRESH 32U
static void
hn_update_packet_stats(struct hn_stats *stats, const struct rte_mbuf *m)
{
uint32_t s = m->pkt_len;
const struct ether_addr *ea;
if (s == 64) {
stats->size_bins[1]++;
} else if (s > 64 && s < 1024) {
uint32_t bin;
/* count zeros, and offset into correct bin */
bin = (sizeof(s) * 8) - __builtin_clz(s) - 5;
stats->size_bins[bin]++;
} else {
if (s < 64)
stats->size_bins[0]++;
else if (s < 1519)
stats->size_bins[6]++;
else if (s >= 1519)
stats->size_bins[7]++;
}
ea = rte_pktmbuf_mtod(m, const struct ether_addr *);
if (is_multicast_ether_addr(ea)) {
if (is_broadcast_ether_addr(ea))
stats->broadcast++;
else
stats->multicast++;
}
}
static inline unsigned int hn_rndis_pktlen(const struct rndis_packet_msg *pkt)
{
return pkt->pktinfooffset + pkt->pktinfolen;
}
static inline uint32_t
hn_rndis_pktmsg_offset(uint32_t ofs)
{
return ofs - offsetof(struct rndis_packet_msg, dataoffset);
}
static void hn_txd_init(struct rte_mempool *mp __rte_unused,
void *opaque, void *obj, unsigned int idx)
{
struct hn_txdesc *txd = obj;
struct rte_eth_dev *dev = opaque;
struct rndis_packet_msg *pkt;
memset(txd, 0, sizeof(*txd));
txd->chim_index = idx;
pkt = rte_malloc_socket("RNDIS_TX", HN_RNDIS_PKT_LEN,
rte_align32pow2(HN_RNDIS_PKT_LEN),
dev->device->numa_node);
if (!pkt)
rte_exit(EXIT_FAILURE, "can not allocate RNDIS header");
txd->rndis_pkt = pkt;
}
/*
* Unlike Linux and FreeBSD, this driver uses a mempool
* to limit outstanding transmits and reserve buffers
*/
int
hn_tx_pool_init(struct rte_eth_dev *dev)
{
struct hn_data *hv = dev->data->dev_private;
char name[RTE_MEMPOOL_NAMESIZE];
struct rte_mempool *mp;
snprintf(name, sizeof(name),
"hn_txd_%u", dev->data->port_id);
PMD_INIT_LOG(DEBUG, "create a TX send pool %s n=%u size=%zu socket=%d",
name, hv->chim_cnt, sizeof(struct hn_txdesc),
dev->device->numa_node);
mp = rte_mempool_create(name, hv->chim_cnt, sizeof(struct hn_txdesc),
HN_TXD_CACHE_SIZE, 0,
NULL, NULL,
hn_txd_init, dev,
dev->device->numa_node, 0);
if (!mp) {
PMD_DRV_LOG(ERR,
"mempool %s create failed: %d", name, rte_errno);
return -rte_errno;
}
hv->tx_pool = mp;
return 0;
}
static void hn_reset_txagg(struct hn_tx_queue *txq)
{
txq->agg_szleft = txq->agg_szmax;
txq->agg_pktleft = txq->agg_pktmax;
txq->agg_txd = NULL;
txq->agg_prevpkt = NULL;
}
int
hn_dev_tx_queue_setup(struct rte_eth_dev *dev,
uint16_t queue_idx, uint16_t nb_desc __rte_unused,
unsigned int socket_id,
const struct rte_eth_txconf *tx_conf)
{
struct hn_data *hv = dev->data->dev_private;
struct hn_tx_queue *txq;
uint32_t tx_free_thresh;
PMD_INIT_FUNC_TRACE();
txq = rte_zmalloc_socket("HN_TXQ", sizeof(*txq), RTE_CACHE_LINE_SIZE,
socket_id);
if (!txq)
return -ENOMEM;
txq->hv = hv;
txq->chan = hv->channels[queue_idx];
txq->port_id = dev->data->port_id;
txq->queue_id = queue_idx;
tx_free_thresh = tx_conf->tx_free_thresh;
if (tx_free_thresh == 0)
tx_free_thresh = RTE_MIN(hv->chim_cnt / 4,
DEFAULT_TX_FREE_THRESH);
if (tx_free_thresh >= hv->chim_cnt - 3)
tx_free_thresh = hv->chim_cnt - 3;
txq->free_thresh = tx_free_thresh;
txq->agg_szmax = RTE_MIN(hv->chim_szmax, hv->rndis_agg_size);
txq->agg_pktmax = hv->rndis_agg_pkts;
txq->agg_align = hv->rndis_agg_align;
hn_reset_txagg(txq);
dev->data->tx_queues[queue_idx] = txq;
return 0;
}
void
hn_dev_tx_queue_release(void *arg)
{
struct hn_tx_queue *txq = arg;
struct hn_txdesc *txd;
PMD_INIT_FUNC_TRACE();
if (!txq)
return;
/* If any pending data is still present just drop it */
txd = txq->agg_txd;
if (txd)
rte_mempool_put(txq->hv->tx_pool, txd);
rte_free(txq);
}
void
hn_dev_tx_queue_info(struct rte_eth_dev *dev, uint16_t queue_idx,
struct rte_eth_txq_info *qinfo)
{
struct hn_data *hv = dev->data->dev_private;
struct hn_tx_queue *txq = dev->data->rx_queues[queue_idx];
qinfo->conf.tx_free_thresh = txq->free_thresh;
qinfo->nb_desc = hv->tx_pool->size;
}
static void
hn_nvs_send_completed(struct rte_eth_dev *dev, uint16_t queue_id,
unsigned long xactid, const struct hn_nvs_rndis_ack *ack)
{
struct hn_txdesc *txd = (struct hn_txdesc *)xactid;
struct hn_tx_queue *txq;
/* Control packets are sent with xacid == 0 */
if (!txd)
return;
txq = dev->data->tx_queues[queue_id];
if (likely(ack->status == NVS_STATUS_OK)) {
PMD_TX_LOG(DEBUG, "port %u:%u complete tx %u packets %u bytes %u",
txq->port_id, txq->queue_id, txd->chim_index,
txd->packets, txd->data_size);
txq->stats.bytes += txd->data_size;
txq->stats.packets += txd->packets;
} else {
PMD_TX_LOG(NOTICE, "port %u:%u complete tx %u failed status %u",
txq->port_id, txq->queue_id, txd->chim_index, ack->status);
++txq->stats.errors;
}
rte_pktmbuf_free(txd->m);
rte_mempool_put(txq->hv->tx_pool, txd);
}
/* Handle transmit completion events */
static void
hn_nvs_handle_comp(struct rte_eth_dev *dev, uint16_t queue_id,
const struct vmbus_chanpkt_hdr *pkt,
const void *data)
{
const struct hn_nvs_hdr *hdr = data;
switch (hdr->type) {
case NVS_TYPE_RNDIS_ACK:
hn_nvs_send_completed(dev, queue_id, pkt->xactid, data);
break;
default:
PMD_TX_LOG(NOTICE,
"unexpected send completion type %u",
hdr->type);
}
}
/* Parse per-packet info (meta data) */
static int
hn_rndis_rxinfo(const void *info_data, unsigned int info_dlen,
struct hn_rxinfo *info)
{
const struct rndis_pktinfo *pi = info_data;
uint32_t mask = 0;
while (info_dlen != 0) {
const void *data;
uint32_t dlen;
if (unlikely(info_dlen < sizeof(*pi)))
return -EINVAL;
if (unlikely(info_dlen < pi->size))
return -EINVAL;
info_dlen -= pi->size;
if (unlikely(pi->size & RNDIS_PKTINFO_SIZE_ALIGNMASK))
return -EINVAL;
if (unlikely(pi->size < pi->offset))
return -EINVAL;
dlen = pi->size - pi->offset;
data = pi->data;
switch (pi->type) {
case NDIS_PKTINFO_TYPE_VLAN:
if (unlikely(dlen < NDIS_VLAN_INFO_SIZE))
return -EINVAL;
info->vlan_info = *((const uint32_t *)data);
mask |= HN_RXINFO_VLAN;
break;
case NDIS_PKTINFO_TYPE_CSUM:
if (unlikely(dlen < NDIS_RXCSUM_INFO_SIZE))
return -EINVAL;
info->csum_info = *((const uint32_t *)data);
mask |= HN_RXINFO_CSUM;
break;
case NDIS_PKTINFO_TYPE_HASHVAL:
if (unlikely(dlen < NDIS_HASH_VALUE_SIZE))
return -EINVAL;
info->hash_value = *((const uint32_t *)data);
mask |= HN_RXINFO_HASHVAL;
break;
case NDIS_PKTINFO_TYPE_HASHINF:
if (unlikely(dlen < NDIS_HASH_INFO_SIZE))
return -EINVAL;
info->hash_info = *((const uint32_t *)data);
mask |= HN_RXINFO_HASHINF;
break;
default:
goto next;
}
if (mask == HN_RXINFO_ALL)
break; /* All found; done */
next:
pi = (const struct rndis_pktinfo *)
((const uint8_t *)pi + pi->size);
}
/*
* Final fixup.
* - If there is no hash value, invalidate the hash info.
*/
if (!(mask & HN_RXINFO_HASHVAL))
info->hash_info = HN_NDIS_HASH_INFO_INVALID;
return 0;
}
/*
* Ack the consumed RXBUF associated w/ this channel packet,
* so that this RXBUF can be recycled by the hypervisor.
*/
static void hn_rx_buf_release(struct hn_rx_bufinfo *rxb)
{
struct rte_mbuf_ext_shared_info *shinfo = &rxb->shinfo;
struct hn_data *hv = rxb->hv;
if (rte_mbuf_ext_refcnt_update(shinfo, -1) == 0) {
hn_nvs_ack_rxbuf(rxb->chan, rxb->xactid);
--hv->rxbuf_outstanding;
}
}
static void hn_rx_buf_free_cb(void *buf __rte_unused, void *opaque)
{
hn_rx_buf_release(opaque);
}
static struct hn_rx_bufinfo *hn_rx_buf_init(const struct hn_rx_queue *rxq,
const struct vmbus_chanpkt_rxbuf *pkt)
{
struct hn_rx_bufinfo *rxb;
rxb = rxq->hv->rxbuf_info + pkt->hdr.xactid;
rxb->chan = rxq->chan;
rxb->xactid = pkt->hdr.xactid;
rxb->hv = rxq->hv;
rxb->shinfo.free_cb = hn_rx_buf_free_cb;
rxb->shinfo.fcb_opaque = rxb;
rte_mbuf_ext_refcnt_set(&rxb->shinfo, 1);
return rxb;
}
static void hn_rxpkt(struct hn_rx_queue *rxq, struct hn_rx_bufinfo *rxb,
uint8_t *data, unsigned int headroom, unsigned int dlen,
const struct hn_rxinfo *info)
{
struct hn_data *hv = rxq->hv;
struct rte_mbuf *m;
m = rte_pktmbuf_alloc(rxq->mb_pool);
if (unlikely(!m)) {
struct rte_eth_dev *dev =
&rte_eth_devices[rxq->port_id];
dev->data->rx_mbuf_alloc_failed++;
return;
}
/*
* For large packets, avoid copy if possible but need to keep
* some space available in receive area for later packets.
*/
if (dlen >= HN_RXCOPY_THRESHOLD &&
hv->rxbuf_outstanding < hv->rxbuf_section_cnt / 2) {
struct rte_mbuf_ext_shared_info *shinfo;
const void *rxbuf;
rte_iova_t iova;
/*
* Build an external mbuf that points to recveive area.
* Use refcount to handle multiple packets in same
* receive buffer section.
*/
rxbuf = hv->rxbuf_res->addr;
iova = rte_mem_virt2iova(rxbuf) + RTE_PTR_DIFF(data, rxbuf);
shinfo = &rxb->shinfo;
if (rte_mbuf_ext_refcnt_update(shinfo, 1) == 1)
++hv->rxbuf_outstanding;
rte_pktmbuf_attach_extbuf(m, data, iova,
dlen + headroom, shinfo);
m->data_off = headroom;
} else {
/* Mbuf's in pool must be large enough to hold small packets */
if (unlikely(rte_pktmbuf_tailroom(m) < dlen)) {
rte_pktmbuf_free_seg(m);
++rxq->stats.errors;
return;
}
rte_memcpy(rte_pktmbuf_mtod(m, void *),
data + headroom, dlen);
}
m->port = rxq->port_id;
m->pkt_len = dlen;
m->data_len = dlen;
if (info->vlan_info != HN_NDIS_VLAN_INFO_INVALID) {
m->vlan_tci = info->vlan_info;
m->ol_flags |= PKT_RX_VLAN_STRIPPED | PKT_RX_VLAN;
}
if (info->csum_info != HN_NDIS_RXCSUM_INFO_INVALID) {
if (info->csum_info & NDIS_RXCSUM_INFO_IPCS_OK)
m->ol_flags |= PKT_RX_IP_CKSUM_GOOD;
if (info->csum_info & (NDIS_RXCSUM_INFO_UDPCS_OK
| NDIS_RXCSUM_INFO_TCPCS_OK))
m->ol_flags |= PKT_RX_L4_CKSUM_GOOD;
}
if (info->hash_info != HN_NDIS_HASH_INFO_INVALID) {
m->ol_flags |= PKT_RX_RSS_HASH;
m->hash.rss = info->hash_value;
}
PMD_RX_LOG(DEBUG, "port %u:%u RX id %" PRIu64 " size %u ol_flags %#" PRIx64,
rxq->port_id, rxq->queue_id, rxb->xactid,
m->pkt_len, m->ol_flags);
++rxq->stats.packets;
rxq->stats.bytes += m->pkt_len;
hn_update_packet_stats(&rxq->stats, m);
if (unlikely(rte_ring_sp_enqueue(rxq->rx_ring, m) != 0)) {
++rxq->ring_full;
rte_pktmbuf_free(m);
}
}
static void hn_rndis_rx_data(struct hn_rx_queue *rxq,
struct hn_rx_bufinfo *rxb,
void *data, uint32_t dlen)
{
unsigned int data_off, data_len, pktinfo_off, pktinfo_len;
const struct rndis_packet_msg *pkt = data;
struct hn_rxinfo info = {
.vlan_info = HN_NDIS_VLAN_INFO_INVALID,
.csum_info = HN_NDIS_RXCSUM_INFO_INVALID,
.hash_info = HN_NDIS_HASH_INFO_INVALID,
};
int err;
hn_rndis_dump(pkt);
if (unlikely(dlen < sizeof(*pkt)))
goto error;
if (unlikely(dlen < pkt->len))
goto error; /* truncated RNDIS from host */
if (unlikely(pkt->len < pkt->datalen
+ pkt->oobdatalen + pkt->pktinfolen))
goto error;
if (unlikely(pkt->datalen == 0))
goto error;
/* Check offsets. */
if (unlikely(pkt->dataoffset < RNDIS_PACKET_MSG_OFFSET_MIN))
goto error;
if (likely(pkt->pktinfooffset > 0) &&
unlikely(pkt->pktinfooffset < RNDIS_PACKET_MSG_OFFSET_MIN ||
(pkt->pktinfooffset & RNDIS_PACKET_MSG_OFFSET_ALIGNMASK)))
goto error;
data_off = RNDIS_PACKET_MSG_OFFSET_ABS(pkt->dataoffset);
data_len = pkt->datalen;
pktinfo_off = RNDIS_PACKET_MSG_OFFSET_ABS(pkt->pktinfooffset);
pktinfo_len = pkt->pktinfolen;
if (likely(pktinfo_len > 0)) {
err = hn_rndis_rxinfo((const uint8_t *)pkt + pktinfo_off,
pktinfo_len, &info);
if (err)
goto error;
}
if (unlikely(data_off + data_len > pkt->len))
goto error;
if (unlikely(data_len < ETHER_HDR_LEN))
goto error;
hn_rxpkt(rxq, rxb, data, data_off, data_len, &info);
return;
error:
++rxq->stats.errors;
}
static void
hn_rndis_receive(const struct rte_eth_dev *dev, struct hn_rx_queue *rxq,
struct hn_rx_bufinfo *rxb, void *buf, uint32_t len)
{
const struct rndis_msghdr *hdr = buf;
switch (hdr->type) {
case RNDIS_PACKET_MSG:
if (dev->data->dev_started)
hn_rndis_rx_data(rxq, rxb, buf, len);
break;
case RNDIS_INDICATE_STATUS_MSG:
hn_rndis_link_status(rxq->hv, buf);
break;
case RNDIS_INITIALIZE_CMPLT:
case RNDIS_QUERY_CMPLT:
case RNDIS_SET_CMPLT:
hn_rndis_receive_response(rxq->hv, buf, len);
break;
default:
PMD_DRV_LOG(NOTICE,
"unexpected RNDIS message (type %#x len %u)",
hdr->type, len);
break;
}
}
static void
hn_nvs_handle_rxbuf(struct rte_eth_dev *dev,
struct hn_data *hv,
struct hn_rx_queue *rxq,
const struct vmbus_chanpkt_hdr *hdr,
const void *buf)
{
const struct vmbus_chanpkt_rxbuf *pkt;
const struct hn_nvs_hdr *nvs_hdr = buf;
uint32_t rxbuf_sz = hv->rxbuf_res->len;
char *rxbuf = hv->rxbuf_res->addr;
unsigned int i, hlen, count;
struct hn_rx_bufinfo *rxb;
/* At minimum we need type header */
if (unlikely(vmbus_chanpkt_datalen(hdr) < sizeof(*nvs_hdr))) {
PMD_RX_LOG(ERR, "invalid receive nvs RNDIS");
return;
}
/* Make sure that this is a RNDIS message. */
if (unlikely(nvs_hdr->type != NVS_TYPE_RNDIS)) {
PMD_RX_LOG(ERR, "nvs type %u, not RNDIS",
nvs_hdr->type);
return;
}
hlen = vmbus_chanpkt_getlen(hdr->hlen);
if (unlikely(hlen < sizeof(*pkt))) {
PMD_RX_LOG(ERR, "invalid rxbuf chanpkt");
return;
}
pkt = container_of(hdr, const struct vmbus_chanpkt_rxbuf, hdr);
if (unlikely(pkt->rxbuf_id != NVS_RXBUF_SIG)) {
PMD_RX_LOG(ERR, "invalid rxbuf_id 0x%08x",
pkt->rxbuf_id);
return;
}
count = pkt->rxbuf_cnt;
if (unlikely(hlen < offsetof(struct vmbus_chanpkt_rxbuf,
rxbuf[count]))) {
PMD_RX_LOG(ERR, "invalid rxbuf_cnt %u", count);
return;
}
if (pkt->hdr.xactid > hv->rxbuf_section_cnt) {
PMD_RX_LOG(ERR, "invalid rxbuf section id %" PRIx64,
pkt->hdr.xactid);
return;
}
/* Setup receive buffer info to allow for callback */
rxb = hn_rx_buf_init(rxq, pkt);
/* Each range represents 1 RNDIS pkt that contains 1 Ethernet frame */
for (i = 0; i < count; ++i) {
unsigned int ofs, len;
ofs = pkt->rxbuf[i].ofs;
len = pkt->rxbuf[i].len;
if (unlikely(ofs + len > rxbuf_sz)) {
PMD_RX_LOG(ERR,
"%uth RNDIS msg overflow ofs %u, len %u",
i, ofs, len);
continue;
}
if (unlikely(len == 0)) {
PMD_RX_LOG(ERR, "%uth RNDIS msg len %u", i, len);
continue;
}
hn_rndis_receive(dev, rxq, rxb,
rxbuf + ofs, len);
}
/* Send ACK now if external mbuf not used */
hn_rx_buf_release(rxb);
}
struct hn_rx_queue *hn_rx_queue_alloc(struct hn_data *hv,
uint16_t queue_id,
unsigned int socket_id)
{
struct hn_rx_queue *rxq;
rxq = rte_zmalloc_socket("HN_RXQ",
sizeof(*rxq) + HN_RXQ_EVENT_DEFAULT,
RTE_CACHE_LINE_SIZE, socket_id);
if (rxq) {
rxq->hv = hv;
rxq->chan = hv->channels[queue_id];
rte_spinlock_init(&rxq->ring_lock);
rxq->port_id = hv->port_id;
rxq->queue_id = queue_id;
}
return rxq;
}
int
hn_dev_rx_queue_setup(struct rte_eth_dev *dev,
uint16_t queue_idx, uint16_t nb_desc,
unsigned int socket_id,
const struct rte_eth_rxconf *rx_conf __rte_unused,
struct rte_mempool *mp)
{
struct hn_data *hv = dev->data->dev_private;
char ring_name[RTE_RING_NAMESIZE];
struct hn_rx_queue *rxq;
unsigned int count;
PMD_INIT_FUNC_TRACE();
if (queue_idx == 0) {
rxq = hv->primary;
} else {
rxq = hn_rx_queue_alloc(hv, queue_idx, socket_id);
if (!rxq)
return -ENOMEM;
}
rxq->mb_pool = mp;
count = rte_mempool_avail_count(mp) / dev->data->nb_rx_queues;
if (nb_desc == 0 || nb_desc > count)
nb_desc = count;
/*
* Staging ring from receive event logic to rx_pkts.
* rx_pkts assumes caller is handling multi-thread issue.
* event logic has locking.
*/
snprintf(ring_name, sizeof(ring_name),
"hn_rx_%u_%u", dev->data->port_id, queue_idx);
rxq->rx_ring = rte_ring_create(ring_name,
rte_align32pow2(nb_desc),
socket_id, 0);
if (!rxq->rx_ring)
goto fail;
dev->data->rx_queues[queue_idx] = rxq;
return 0;
fail:
rte_ring_free(rxq->rx_ring);
rte_free(rxq->event_buf);
rte_free(rxq);
return -ENOMEM;
}
void
hn_dev_rx_queue_release(void *arg)
{
struct hn_rx_queue *rxq = arg;
PMD_INIT_FUNC_TRACE();
if (!rxq)
return;
rte_ring_free(rxq->rx_ring);
rxq->rx_ring = NULL;
rxq->mb_pool = NULL;
if (rxq != rxq->hv->primary) {
rte_free(rxq->event_buf);
rte_free(rxq);
}
}
void
hn_dev_rx_queue_info(struct rte_eth_dev *dev, uint16_t queue_idx,
struct rte_eth_rxq_info *qinfo)
{
struct hn_rx_queue *rxq = dev->data->rx_queues[queue_idx];
qinfo->mp = rxq->mb_pool;
qinfo->scattered_rx = 1;
qinfo->nb_desc = rte_ring_get_capacity(rxq->rx_ring);
}
static void
hn_nvs_handle_notify(const struct vmbus_chanpkt_hdr *pkthdr,
const void *data)
{
const struct hn_nvs_hdr *hdr = data;
if (unlikely(vmbus_chanpkt_datalen(pkthdr) < sizeof(*hdr))) {
PMD_DRV_LOG(ERR, "invalid nvs notify");
return;
}
PMD_DRV_LOG(INFO,
"got notify, nvs type %u", hdr->type);
}
/*
* Process pending events on the channel.
* Called from both Rx queue poll and Tx cleanup
*/
void hn_process_events(struct hn_data *hv, uint16_t queue_id)
{
struct rte_eth_dev *dev = &rte_eth_devices[hv->port_id];
struct hn_rx_queue *rxq;
uint32_t bytes_read = 0;
int ret = 0;
rxq = queue_id == 0 ? hv->primary : dev->data->rx_queues[queue_id];
/* If no pending data then nothing to do */
if (rte_vmbus_chan_rx_empty(rxq->chan))
return;
/*
* Since channel is shared between Rx and TX queue need to have a lock
* since DPDK does not force same CPU to be used for Rx/Tx.
*/
if (unlikely(!rte_spinlock_trylock(&rxq->ring_lock)))
return;
for (;;) {
const struct vmbus_chanpkt_hdr *pkt;
uint32_t len = HN_RXQ_EVENT_DEFAULT;
const void *data;
ret = rte_vmbus_chan_recv_raw(rxq->chan, rxq->event_buf, &len);
if (ret == -EAGAIN)
break; /* ring is empty */
else if (ret == -ENOBUFS)
rte_exit(EXIT_FAILURE, "event buffer not big enough (%u < %u)",
HN_RXQ_EVENT_DEFAULT, len);
else if (ret <= 0)
rte_exit(EXIT_FAILURE,
"vmbus ring buffer error: %d", ret);
bytes_read += ret;
pkt = (const struct vmbus_chanpkt_hdr *)rxq->event_buf;
data = (char *)rxq->event_buf + vmbus_chanpkt_getlen(pkt->hlen);
switch (pkt->type) {
case VMBUS_CHANPKT_TYPE_COMP:
hn_nvs_handle_comp(dev, queue_id, pkt, data);
break;
case VMBUS_CHANPKT_TYPE_RXBUF:
hn_nvs_handle_rxbuf(dev, hv, rxq, pkt, data);
break;
case VMBUS_CHANPKT_TYPE_INBAND:
hn_nvs_handle_notify(pkt, data);
break;
default:
PMD_DRV_LOG(ERR, "unknown chan pkt %u", pkt->type);
break;
}
if (rxq->rx_ring && rte_ring_full(rxq->rx_ring))
break;
}
if (bytes_read > 0)
rte_vmbus_chan_signal_read(rxq->chan, bytes_read);
rte_spinlock_unlock(&rxq->ring_lock);
}
static void hn_append_to_chim(struct hn_tx_queue *txq,
struct rndis_packet_msg *pkt,
const struct rte_mbuf *m)
{
struct hn_txdesc *txd = txq->agg_txd;
uint8_t *buf = (uint8_t *)pkt;
unsigned int data_offs;
hn_rndis_dump(pkt);
data_offs = RNDIS_PACKET_MSG_OFFSET_ABS(pkt->dataoffset);
txd->chim_size += pkt->len;
txd->data_size += m->pkt_len;
++txd->packets;
hn_update_packet_stats(&txq->stats, m);
for (; m; m = m->next) {
uint16_t len = rte_pktmbuf_data_len(m);
rte_memcpy(buf + data_offs,
rte_pktmbuf_mtod(m, const char *), len);
data_offs += len;
}
}
/*
* Send pending aggregated data in chimney buffer (if any).
* Returns error if send was unsuccessful because channel ring buffer
* was full.
*/
static int hn_flush_txagg(struct hn_tx_queue *txq, bool *need_sig)
{
struct hn_txdesc *txd = txq->agg_txd;
struct hn_nvs_rndis rndis;
int ret;
if (!txd)
return 0;
rndis = (struct hn_nvs_rndis) {
.type = NVS_TYPE_RNDIS,
.rndis_mtype = NVS_RNDIS_MTYPE_DATA,
.chim_idx = txd->chim_index,
.chim_sz = txd->chim_size,
};
PMD_TX_LOG(DEBUG, "port %u:%u tx %u size %u",
txq->port_id, txq->queue_id, txd->chim_index, txd->chim_size);
ret = hn_nvs_send(txq->chan, VMBUS_CHANPKT_FLAG_RC,
&rndis, sizeof(rndis), (uintptr_t)txd, need_sig);
if (likely(ret == 0))
hn_reset_txagg(txq);
else
PMD_TX_LOG(NOTICE, "port %u:%u send failed: %d",
txq->port_id, txq->queue_id, ret);
return ret;
}
static struct hn_txdesc *hn_new_txd(struct hn_data *hv,
struct hn_tx_queue *txq)
{
struct hn_txdesc *txd;
if (rte_mempool_get(hv->tx_pool, (void **)&txd)) {
++txq->stats.nomemory;
PMD_TX_LOG(DEBUG, "tx pool exhausted!");
return NULL;
}
txd->m = NULL;
txd->queue_id = txq->queue_id;
txd->packets = 0;
txd->data_size = 0;
txd->chim_size = 0;
return txd;
}
static void *
hn_try_txagg(struct hn_data *hv, struct hn_tx_queue *txq, uint32_t pktsize)
{
struct hn_txdesc *agg_txd = txq->agg_txd;
struct rndis_packet_msg *pkt;
void *chim;
if (agg_txd) {
unsigned int padding, olen;
/*
* Update the previous RNDIS packet's total length,
* it can be increased due to the mandatory alignment
* padding for this RNDIS packet. And update the
* aggregating txdesc's chimney sending buffer size
* accordingly.
*
* Zero-out the padding, as required by the RNDIS spec.
*/
pkt = txq->agg_prevpkt;
olen = pkt->len;
padding = RTE_ALIGN(olen, txq->agg_align) - olen;
if (padding > 0) {
agg_txd->chim_size += padding;
pkt->len += padding;
memset((uint8_t *)pkt + olen, 0, padding);
}
chim = (uint8_t *)pkt + pkt->len;
txq->agg_pktleft--;
txq->agg_szleft -= pktsize;
if (txq->agg_szleft < HN_PKTSIZE_MIN(txq->agg_align)) {
/*
* Probably can't aggregate more packets,
* flush this aggregating txdesc proactively.
*/
txq->agg_pktleft = 0;
}
} else {
agg_txd = hn_new_txd(hv, txq);
if (!agg_txd)
return NULL;
chim = (uint8_t *)hv->chim_res->addr
+ agg_txd->chim_index * hv->chim_szmax;
txq->agg_txd = agg_txd;
txq->agg_pktleft = txq->agg_pktmax - 1;
txq->agg_szleft = txq->agg_szmax - pktsize;
}
txq->agg_prevpkt = chim;
return chim;
}
static inline void *
hn_rndis_pktinfo_append(struct rndis_packet_msg *pkt,
uint32_t pi_dlen, uint32_t pi_type)
{
const uint32_t pi_size = RNDIS_PKTINFO_SIZE(pi_dlen);
struct rndis_pktinfo *pi;
/*
* Per-packet-info does not move; it only grows.
*
* NOTE:
* pktinfooffset in this phase counts from the beginning
* of rndis_packet_msg.
*/
pi = (struct rndis_pktinfo *)((uint8_t *)pkt + hn_rndis_pktlen(pkt));
pkt->pktinfolen += pi_size;
pi->size = pi_size;
pi->type = pi_type;
pi->offset = RNDIS_PKTINFO_OFFSET;
return pi->data;
}
/* Put RNDIS header and packet info on packet */
static void hn_encap(struct rndis_packet_msg *pkt,
uint16_t queue_id,
const struct rte_mbuf *m)
{
unsigned int hlen = m->l2_len + m->l3_len;
uint32_t *pi_data;
uint32_t pkt_hlen;
pkt->type = RNDIS_PACKET_MSG;
pkt->len = m->pkt_len;
pkt->dataoffset = 0;
pkt->datalen = m->pkt_len;
pkt->oobdataoffset = 0;
pkt->oobdatalen = 0;
pkt->oobdataelements = 0;
pkt->pktinfooffset = sizeof(*pkt);
pkt->pktinfolen = 0;
pkt->vchandle = 0;
pkt->reserved = 0;
/*
* Set the hash value for this packet, to the queue_id to cause
* TX done event for this packet on the right channel.
*/
pi_data = hn_rndis_pktinfo_append(pkt, NDIS_HASH_VALUE_SIZE,
NDIS_PKTINFO_TYPE_HASHVAL);
*pi_data = queue_id;
if (m->ol_flags & PKT_TX_VLAN_PKT) {
pi_data = hn_rndis_pktinfo_append(pkt, NDIS_VLAN_INFO_SIZE,
NDIS_PKTINFO_TYPE_VLAN);
*pi_data = m->vlan_tci;
}
if (m->ol_flags & PKT_TX_TCP_SEG) {
pi_data = hn_rndis_pktinfo_append(pkt, NDIS_LSO2_INFO_SIZE,
NDIS_PKTINFO_TYPE_LSO);
if (m->ol_flags & PKT_TX_IPV6) {
*pi_data = NDIS_LSO2_INFO_MAKEIPV6(hlen,
m->tso_segsz);
} else {
*pi_data = NDIS_LSO2_INFO_MAKEIPV4(hlen,
m->tso_segsz);
}
} else if (m->ol_flags &
(PKT_TX_TCP_CKSUM | PKT_TX_UDP_CKSUM | PKT_TX_IP_CKSUM)) {
pi_data = hn_rndis_pktinfo_append(pkt, NDIS_TXCSUM_INFO_SIZE,
NDIS_PKTINFO_TYPE_CSUM);
*pi_data = 0;
if (m->ol_flags & PKT_TX_IPV6)
*pi_data |= NDIS_TXCSUM_INFO_IPV6;
if (m->ol_flags & PKT_TX_IPV4) {
*pi_data |= NDIS_TXCSUM_INFO_IPV4;
if (m->ol_flags & PKT_TX_IP_CKSUM)
*pi_data |= NDIS_TXCSUM_INFO_IPCS;
}
if (m->ol_flags & PKT_TX_TCP_CKSUM)
*pi_data |= NDIS_TXCSUM_INFO_MKTCPCS(hlen);
else if (m->ol_flags & PKT_TX_UDP_CKSUM)
*pi_data |= NDIS_TXCSUM_INFO_MKUDPCS(hlen);
}
pkt_hlen = pkt->pktinfooffset + pkt->pktinfolen;
/* Fixup RNDIS packet message total length */
pkt->len += pkt_hlen;
/* Convert RNDIS packet message offsets */
pkt->dataoffset = hn_rndis_pktmsg_offset(pkt_hlen);
pkt->pktinfooffset = hn_rndis_pktmsg_offset(pkt->pktinfooffset);
}
/* How many scatter gather list elements ar needed */
static unsigned int hn_get_slots(const struct rte_mbuf *m)
{
unsigned int slots = 1; /* for RNDIS header */
while (m) {
unsigned int size = rte_pktmbuf_data_len(m);
unsigned int offs = rte_mbuf_data_iova(m) & PAGE_MASK;
slots += (offs + size + PAGE_SIZE - 1) / PAGE_SIZE;
m = m->next;
}
return slots;
}
/* Build scatter gather list from chained mbuf */
static unsigned int hn_fill_sg(struct vmbus_gpa *sg,
const struct rte_mbuf *m)
{
unsigned int segs = 0;
while (m) {
rte_iova_t addr = rte_mbuf_data_iova(m);
unsigned int page = addr / PAGE_SIZE;
unsigned int offset = addr & PAGE_MASK;
unsigned int len = rte_pktmbuf_data_len(m);
while (len > 0) {
unsigned int bytes = RTE_MIN(len, PAGE_SIZE - offset);
sg[segs].page = page;
sg[segs].ofs = offset;
sg[segs].len = bytes;
segs++;
++page;
offset = 0;
len -= bytes;
}
m = m->next;
}
return segs;
}
/* Transmit directly from mbuf */
static int hn_xmit_sg(struct hn_tx_queue *txq,
const struct hn_txdesc *txd, const struct rte_mbuf *m,
bool *need_sig)
{
struct vmbus_gpa sg[hn_get_slots(m)];
struct hn_nvs_rndis nvs_rndis = {
.type = NVS_TYPE_RNDIS,
.rndis_mtype = NVS_RNDIS_MTYPE_DATA,
.chim_sz = txd->chim_size,
};
rte_iova_t addr;
unsigned int segs;
/* attach aggregation data if present */
if (txd->chim_size > 0)
nvs_rndis.chim_idx = txd->chim_index;
else
nvs_rndis.chim_idx = NVS_CHIM_IDX_INVALID;
hn_rndis_dump(txd->rndis_pkt);
/* pass IOVA of rndis header in first segment */
addr = rte_malloc_virt2iova(txd->rndis_pkt);
if (unlikely(addr == RTE_BAD_IOVA)) {
PMD_DRV_LOG(ERR, "RNDIS transmit can not get iova");
return -EINVAL;
}
sg[0].page = addr / PAGE_SIZE;
sg[0].ofs = addr & PAGE_MASK;
sg[0].len = RNDIS_PACKET_MSG_OFFSET_ABS(hn_rndis_pktlen(txd->rndis_pkt));
segs = 1;
hn_update_packet_stats(&txq->stats, m);
segs += hn_fill_sg(sg + 1, m);
PMD_TX_LOG(DEBUG, "port %u:%u tx %u segs %u size %u",
txq->port_id, txq->queue_id, txd->chim_index,
segs, nvs_rndis.chim_sz);
return hn_nvs_send_sglist(txq->chan, sg, segs,
&nvs_rndis, sizeof(nvs_rndis),
(uintptr_t)txd, need_sig);
}
uint16_t
hn_xmit_pkts(void *ptxq, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)
{
struct hn_tx_queue *txq = ptxq;
struct hn_data *hv = txq->hv;
bool need_sig = false;
uint16_t nb_tx;
int ret;
if (unlikely(hv->closed))
return 0;
if (rte_mempool_avail_count(hv->tx_pool) <= txq->free_thresh)
hn_process_events(hv, txq->queue_id);
for (nb_tx = 0; nb_tx < nb_pkts; nb_tx++) {
struct rte_mbuf *m = tx_pkts[nb_tx];
uint32_t pkt_size = m->pkt_len + HN_RNDIS_PKT_LEN;
struct rndis_packet_msg *pkt;
/* For small packets aggregate them in chimney buffer */
if (m->pkt_len < HN_TXCOPY_THRESHOLD && pkt_size <= txq->agg_szmax) {
/* If this packet will not fit, then flush */
if (txq->agg_pktleft == 0 ||
RTE_ALIGN(pkt_size, txq->agg_align) > txq->agg_szleft) {
if (hn_flush_txagg(txq, &need_sig))
goto fail;
}
pkt = hn_try_txagg(hv, txq, pkt_size);
if (unlikely(!pkt))
break;
hn_encap(pkt, txq->queue_id, m);
hn_append_to_chim(txq, pkt, m);
rte_pktmbuf_free(m);
/* if buffer is full, flush */
if (txq->agg_pktleft == 0 &&
hn_flush_txagg(txq, &need_sig))
goto fail;
} else {
struct hn_txdesc *txd;
/* can send chimney data and large packet at once */
txd = txq->agg_txd;
if (txd) {
hn_reset_txagg(txq);
} else {
txd = hn_new_txd(hv, txq);
if (unlikely(!txd))
break;
}
pkt = txd->rndis_pkt;
txd->m = m;
txd->data_size += m->pkt_len;
++txd->packets;
hn_encap(pkt, txq->queue_id, m);
ret = hn_xmit_sg(txq, txd, m, &need_sig);
if (unlikely(ret != 0)) {
PMD_TX_LOG(NOTICE, "sg send failed: %d", ret);
++txq->stats.errors;
rte_mempool_put(hv->tx_pool, txd);
goto fail;
}
}
}
/* If partial buffer left, then try and send it.
* if that fails, then reuse it on next send.
*/
hn_flush_txagg(txq, &need_sig);
fail:
if (need_sig)
rte_vmbus_chan_signal_tx(txq->chan);
return nb_tx;
}
uint16_t
hn_recv_pkts(void *prxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts)
{
struct hn_rx_queue *rxq = prxq;
struct hn_data *hv = rxq->hv;
if (unlikely(hv->closed))
return 0;
/* If ring is empty then process more */
if (rte_ring_count(rxq->rx_ring) < nb_pkts)
hn_process_events(hv, rxq->queue_id);
/* Get mbufs off staging ring */
return rte_ring_sc_dequeue_burst(rxq->rx_ring, (void **)rx_pkts,
nb_pkts, NULL);
}