295968d174
Add 'RTE_ETH' namespace to all enums & macros in a backward compatible way. The macros for backward compatibility can be removed in next LTS. Also updated some struct names to have 'rte_eth' prefix. All internal components switched to using new names. Syntax fixed on lines that this patch touches. Signed-off-by: Ferruh Yigit <ferruh.yigit@intel.com> Acked-by: Tyler Retzlaff <roretzla@linux.microsoft.com> Acked-by: Andrew Rybchenko <andrew.rybchenko@oktetlabs.ru> Acked-by: Ajit Khaparde <ajit.khaparde@broadcom.com> Acked-by: Jerin Jacob <jerinj@marvell.com> Acked-by: Wisam Jaddo <wisamm@nvidia.com> Acked-by: Rosen Xu <rosen.xu@intel.com> Acked-by: Chenbo Xia <chenbo.xia@intel.com> Acked-by: Hemant Agrawal <hemant.agrawal@nxp.com> Acked-by: Somnath Kotur <somnath.kotur@broadcom.com>
1059 lines
26 KiB
C
1059 lines
26 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(C) 2021 Marvell.
|
|
*/
|
|
|
|
#include <unistd.h>
|
|
|
|
#include <rte_eal.h>
|
|
#include <rte_mempool.h>
|
|
#include <rte_mbuf.h>
|
|
#include <rte_io.h>
|
|
#include <rte_net.h>
|
|
#include <ethdev_pci.h>
|
|
|
|
#include "otx_ep_common.h"
|
|
#include "otx_ep_vf.h"
|
|
#include "otx2_ep_vf.h"
|
|
#include "otx_ep_rxtx.h"
|
|
|
|
/* SDP_LENGTH_S specifies packet length and is of 8-byte size */
|
|
#define INFO_SIZE 8
|
|
#define DROQ_REFILL_THRESHOLD 16
|
|
|
|
static void
|
|
otx_ep_dmazone_free(const struct rte_memzone *mz)
|
|
{
|
|
const struct rte_memzone *mz_tmp;
|
|
int ret = 0;
|
|
|
|
if (mz == NULL) {
|
|
otx_ep_err("Memzone: NULL\n");
|
|
return;
|
|
}
|
|
|
|
mz_tmp = rte_memzone_lookup(mz->name);
|
|
if (mz_tmp == NULL) {
|
|
otx_ep_err("Memzone %s Not Found\n", mz->name);
|
|
return;
|
|
}
|
|
|
|
ret = rte_memzone_free(mz);
|
|
if (ret)
|
|
otx_ep_err("Memzone free failed : ret = %d\n", ret);
|
|
}
|
|
|
|
/* Free IQ resources */
|
|
int
|
|
otx_ep_delete_iqs(struct otx_ep_device *otx_ep, uint32_t iq_no)
|
|
{
|
|
struct otx_ep_instr_queue *iq;
|
|
|
|
iq = otx_ep->instr_queue[iq_no];
|
|
if (iq == NULL) {
|
|
otx_ep_err("Invalid IQ[%d]\n", iq_no);
|
|
return -EINVAL;
|
|
}
|
|
|
|
rte_free(iq->req_list);
|
|
iq->req_list = NULL;
|
|
|
|
if (iq->iq_mz) {
|
|
otx_ep_dmazone_free(iq->iq_mz);
|
|
iq->iq_mz = NULL;
|
|
}
|
|
|
|
rte_free(otx_ep->instr_queue[iq_no]);
|
|
otx_ep->instr_queue[iq_no] = NULL;
|
|
|
|
otx_ep->nb_tx_queues--;
|
|
|
|
otx_ep_info("IQ[%d] is deleted\n", iq_no);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* IQ initialization */
|
|
static int
|
|
otx_ep_init_instr_queue(struct otx_ep_device *otx_ep, int iq_no, int num_descs,
|
|
unsigned int socket_id)
|
|
{
|
|
const struct otx_ep_config *conf;
|
|
struct otx_ep_instr_queue *iq;
|
|
uint32_t q_size;
|
|
|
|
conf = otx_ep->conf;
|
|
iq = otx_ep->instr_queue[iq_no];
|
|
q_size = conf->iq.instr_type * num_descs;
|
|
|
|
/* IQ memory creation for Instruction submission to OCTEON TX2 */
|
|
iq->iq_mz = rte_eth_dma_zone_reserve(otx_ep->eth_dev,
|
|
"instr_queue", iq_no, q_size,
|
|
OTX_EP_PCI_RING_ALIGN,
|
|
socket_id);
|
|
if (iq->iq_mz == NULL) {
|
|
otx_ep_err("IQ[%d] memzone alloc failed\n", iq_no);
|
|
goto iq_init_fail;
|
|
}
|
|
|
|
iq->base_addr_dma = iq->iq_mz->iova;
|
|
iq->base_addr = (uint8_t *)iq->iq_mz->addr;
|
|
|
|
if (num_descs & (num_descs - 1)) {
|
|
otx_ep_err("IQ[%d] descs not in power of 2\n", iq_no);
|
|
goto iq_init_fail;
|
|
}
|
|
|
|
iq->nb_desc = num_descs;
|
|
|
|
/* Create a IQ request list to hold requests that have been
|
|
* posted to OCTEON TX2. This list will be used for freeing the IQ
|
|
* data buffer(s) later once the OCTEON TX2 fetched the requests.
|
|
*/
|
|
iq->req_list = rte_zmalloc_socket("request_list",
|
|
(iq->nb_desc * OTX_EP_IQREQ_LIST_SIZE),
|
|
RTE_CACHE_LINE_SIZE,
|
|
rte_socket_id());
|
|
if (iq->req_list == NULL) {
|
|
otx_ep_err("IQ[%d] req_list alloc failed\n", iq_no);
|
|
goto iq_init_fail;
|
|
}
|
|
|
|
otx_ep_info("IQ[%d]: base: %p basedma: %lx count: %d\n",
|
|
iq_no, iq->base_addr, (unsigned long)iq->base_addr_dma,
|
|
iq->nb_desc);
|
|
|
|
iq->otx_ep_dev = otx_ep;
|
|
iq->q_no = iq_no;
|
|
iq->fill_cnt = 0;
|
|
iq->host_write_index = 0;
|
|
iq->otx_read_index = 0;
|
|
iq->flush_index = 0;
|
|
iq->instr_pending = 0;
|
|
|
|
otx_ep->io_qmask.iq |= (1ull << iq_no);
|
|
|
|
/* Set 32B/64B mode for each input queue */
|
|
if (conf->iq.instr_type == 64)
|
|
otx_ep->io_qmask.iq64B |= (1ull << iq_no);
|
|
|
|
iq->iqcmd_64B = (conf->iq.instr_type == 64);
|
|
|
|
/* Set up IQ registers */
|
|
otx_ep->fn_list.setup_iq_regs(otx_ep, iq_no);
|
|
|
|
return 0;
|
|
|
|
iq_init_fail:
|
|
return -ENOMEM;
|
|
}
|
|
|
|
int
|
|
otx_ep_setup_iqs(struct otx_ep_device *otx_ep, uint32_t iq_no, int num_descs,
|
|
unsigned int socket_id)
|
|
{
|
|
struct otx_ep_instr_queue *iq;
|
|
|
|
iq = (struct otx_ep_instr_queue *)rte_zmalloc("otx_ep_IQ", sizeof(*iq),
|
|
RTE_CACHE_LINE_SIZE);
|
|
if (iq == NULL)
|
|
return -ENOMEM;
|
|
|
|
otx_ep->instr_queue[iq_no] = iq;
|
|
|
|
if (otx_ep_init_instr_queue(otx_ep, iq_no, num_descs, socket_id)) {
|
|
otx_ep_err("IQ init is failed\n");
|
|
goto delete_IQ;
|
|
}
|
|
otx_ep->nb_tx_queues++;
|
|
|
|
otx_ep_info("IQ[%d] is created.\n", iq_no);
|
|
|
|
return 0;
|
|
|
|
delete_IQ:
|
|
otx_ep_delete_iqs(otx_ep, iq_no);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static void
|
|
otx_ep_droq_reset_indices(struct otx_ep_droq *droq)
|
|
{
|
|
droq->read_idx = 0;
|
|
droq->write_idx = 0;
|
|
droq->refill_idx = 0;
|
|
droq->refill_count = 0;
|
|
droq->last_pkt_count = 0;
|
|
droq->pkts_pending = 0;
|
|
}
|
|
|
|
static void
|
|
otx_ep_droq_destroy_ring_buffers(struct otx_ep_droq *droq)
|
|
{
|
|
uint32_t idx;
|
|
|
|
for (idx = 0; idx < droq->nb_desc; idx++) {
|
|
if (droq->recv_buf_list[idx]) {
|
|
rte_pktmbuf_free(droq->recv_buf_list[idx]);
|
|
droq->recv_buf_list[idx] = NULL;
|
|
}
|
|
}
|
|
|
|
otx_ep_droq_reset_indices(droq);
|
|
}
|
|
|
|
/* Free OQs resources */
|
|
int
|
|
otx_ep_delete_oqs(struct otx_ep_device *otx_ep, uint32_t oq_no)
|
|
{
|
|
struct otx_ep_droq *droq;
|
|
|
|
droq = otx_ep->droq[oq_no];
|
|
if (droq == NULL) {
|
|
otx_ep_err("Invalid droq[%d]\n", oq_no);
|
|
return -EINVAL;
|
|
}
|
|
|
|
otx_ep_droq_destroy_ring_buffers(droq);
|
|
rte_free(droq->recv_buf_list);
|
|
droq->recv_buf_list = NULL;
|
|
|
|
if (droq->desc_ring_mz) {
|
|
otx_ep_dmazone_free(droq->desc_ring_mz);
|
|
droq->desc_ring_mz = NULL;
|
|
}
|
|
|
|
memset(droq, 0, OTX_EP_DROQ_SIZE);
|
|
|
|
rte_free(otx_ep->droq[oq_no]);
|
|
otx_ep->droq[oq_no] = NULL;
|
|
|
|
otx_ep->nb_rx_queues--;
|
|
|
|
otx_ep_info("OQ[%d] is deleted\n", oq_no);
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
otx_ep_droq_setup_ring_buffers(struct otx_ep_droq *droq)
|
|
{
|
|
struct otx_ep_droq_desc *desc_ring = droq->desc_ring;
|
|
struct otx_ep_droq_info *info;
|
|
struct rte_mbuf *buf;
|
|
uint32_t idx;
|
|
|
|
for (idx = 0; idx < droq->nb_desc; idx++) {
|
|
buf = rte_pktmbuf_alloc(droq->mpool);
|
|
if (buf == NULL) {
|
|
otx_ep_err("OQ buffer alloc failed\n");
|
|
droq->stats.rx_alloc_failure++;
|
|
return -ENOMEM;
|
|
}
|
|
|
|
droq->recv_buf_list[idx] = buf;
|
|
info = rte_pktmbuf_mtod(buf, struct otx_ep_droq_info *);
|
|
memset(info, 0, sizeof(*info));
|
|
desc_ring[idx].buffer_ptr = rte_mbuf_data_iova_default(buf);
|
|
}
|
|
|
|
otx_ep_droq_reset_indices(droq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* OQ initialization */
|
|
static int
|
|
otx_ep_init_droq(struct otx_ep_device *otx_ep, uint32_t q_no,
|
|
uint32_t num_descs, uint32_t desc_size,
|
|
struct rte_mempool *mpool, unsigned int socket_id)
|
|
{
|
|
const struct otx_ep_config *conf = otx_ep->conf;
|
|
uint32_t c_refill_threshold;
|
|
struct otx_ep_droq *droq;
|
|
uint32_t desc_ring_size;
|
|
|
|
otx_ep_info("OQ[%d] Init start\n", q_no);
|
|
|
|
droq = otx_ep->droq[q_no];
|
|
droq->otx_ep_dev = otx_ep;
|
|
droq->q_no = q_no;
|
|
droq->mpool = mpool;
|
|
|
|
droq->nb_desc = num_descs;
|
|
droq->buffer_size = desc_size;
|
|
c_refill_threshold = RTE_MAX(conf->oq.refill_threshold,
|
|
droq->nb_desc / 2);
|
|
|
|
/* OQ desc_ring set up */
|
|
desc_ring_size = droq->nb_desc * OTX_EP_DROQ_DESC_SIZE;
|
|
droq->desc_ring_mz = rte_eth_dma_zone_reserve(otx_ep->eth_dev, "droq",
|
|
q_no, desc_ring_size,
|
|
OTX_EP_PCI_RING_ALIGN,
|
|
socket_id);
|
|
|
|
if (droq->desc_ring_mz == NULL) {
|
|
otx_ep_err("OQ:%d desc_ring allocation failed\n", q_no);
|
|
goto init_droq_fail;
|
|
}
|
|
|
|
droq->desc_ring_dma = droq->desc_ring_mz->iova;
|
|
droq->desc_ring = (struct otx_ep_droq_desc *)droq->desc_ring_mz->addr;
|
|
|
|
otx_ep_dbg("OQ[%d]: desc_ring: virt: 0x%p, dma: %lx\n",
|
|
q_no, droq->desc_ring, (unsigned long)droq->desc_ring_dma);
|
|
otx_ep_dbg("OQ[%d]: num_desc: %d\n", q_no, droq->nb_desc);
|
|
|
|
/* OQ buf_list set up */
|
|
droq->recv_buf_list = rte_zmalloc_socket("recv_buf_list",
|
|
(droq->nb_desc * sizeof(struct rte_mbuf *)),
|
|
RTE_CACHE_LINE_SIZE, socket_id);
|
|
if (droq->recv_buf_list == NULL) {
|
|
otx_ep_err("OQ recv_buf_list alloc failed\n");
|
|
goto init_droq_fail;
|
|
}
|
|
|
|
if (otx_ep_droq_setup_ring_buffers(droq))
|
|
goto init_droq_fail;
|
|
|
|
droq->refill_threshold = c_refill_threshold;
|
|
|
|
/* Set up OQ registers */
|
|
otx_ep->fn_list.setup_oq_regs(otx_ep, q_no);
|
|
|
|
otx_ep->io_qmask.oq |= (1ull << q_no);
|
|
|
|
return 0;
|
|
|
|
init_droq_fail:
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* OQ configuration and setup */
|
|
int
|
|
otx_ep_setup_oqs(struct otx_ep_device *otx_ep, int oq_no, int num_descs,
|
|
int desc_size, struct rte_mempool *mpool,
|
|
unsigned int socket_id)
|
|
{
|
|
struct otx_ep_droq *droq;
|
|
|
|
/* Allocate new droq. */
|
|
droq = (struct otx_ep_droq *)rte_zmalloc("otx_ep_OQ",
|
|
sizeof(*droq), RTE_CACHE_LINE_SIZE);
|
|
if (droq == NULL) {
|
|
otx_ep_err("Droq[%d] Creation Failed\n", oq_no);
|
|
return -ENOMEM;
|
|
}
|
|
otx_ep->droq[oq_no] = droq;
|
|
|
|
if (otx_ep_init_droq(otx_ep, oq_no, num_descs, desc_size, mpool,
|
|
socket_id)) {
|
|
otx_ep_err("Droq[%d] Initialization failed\n", oq_no);
|
|
goto delete_OQ;
|
|
}
|
|
otx_ep_info("OQ[%d] is created.\n", oq_no);
|
|
|
|
otx_ep->nb_rx_queues++;
|
|
|
|
return 0;
|
|
|
|
delete_OQ:
|
|
otx_ep_delete_oqs(otx_ep, oq_no);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static inline void
|
|
otx_ep_iqreq_delete(struct otx_ep_instr_queue *iq, uint32_t idx)
|
|
{
|
|
uint32_t reqtype;
|
|
void *buf;
|
|
struct otx_ep_buf_free_info *finfo;
|
|
|
|
buf = iq->req_list[idx].buf;
|
|
reqtype = iq->req_list[idx].reqtype;
|
|
|
|
switch (reqtype) {
|
|
case OTX_EP_REQTYPE_NORESP_NET:
|
|
rte_pktmbuf_free((struct rte_mbuf *)buf);
|
|
otx_ep_dbg("IQ buffer freed at idx[%d]\n", idx);
|
|
break;
|
|
|
|
case OTX_EP_REQTYPE_NORESP_GATHER:
|
|
finfo = (struct otx_ep_buf_free_info *)buf;
|
|
/* This will take care of multiple segments also */
|
|
rte_pktmbuf_free(finfo->mbuf);
|
|
rte_free(finfo->g.sg);
|
|
rte_free(finfo);
|
|
break;
|
|
|
|
case OTX_EP_REQTYPE_NONE:
|
|
default:
|
|
otx_ep_info("This iqreq mode is not supported:%d\n", reqtype);
|
|
}
|
|
|
|
/* Reset the request list at this index */
|
|
iq->req_list[idx].buf = NULL;
|
|
iq->req_list[idx].reqtype = 0;
|
|
}
|
|
|
|
static inline void
|
|
otx_ep_iqreq_add(struct otx_ep_instr_queue *iq, void *buf,
|
|
uint32_t reqtype, int index)
|
|
{
|
|
iq->req_list[index].buf = buf;
|
|
iq->req_list[index].reqtype = reqtype;
|
|
}
|
|
|
|
static uint32_t
|
|
otx_vf_update_read_index(struct otx_ep_instr_queue *iq)
|
|
{
|
|
uint32_t new_idx = rte_read32(iq->inst_cnt_reg);
|
|
if (unlikely(new_idx == 0xFFFFFFFFU))
|
|
rte_write32(new_idx, iq->inst_cnt_reg);
|
|
/* Modulo of the new index with the IQ size will give us
|
|
* the new index.
|
|
*/
|
|
new_idx &= (iq->nb_desc - 1);
|
|
|
|
return new_idx;
|
|
}
|
|
|
|
static void
|
|
otx_ep_flush_iq(struct otx_ep_instr_queue *iq)
|
|
{
|
|
uint32_t instr_processed = 0;
|
|
|
|
iq->otx_read_index = otx_vf_update_read_index(iq);
|
|
while (iq->flush_index != iq->otx_read_index) {
|
|
/* Free the IQ data buffer to the pool */
|
|
otx_ep_iqreq_delete(iq, iq->flush_index);
|
|
iq->flush_index =
|
|
otx_ep_incr_index(iq->flush_index, 1, iq->nb_desc);
|
|
|
|
instr_processed++;
|
|
}
|
|
|
|
iq->stats.instr_processed = instr_processed;
|
|
iq->instr_pending -= instr_processed;
|
|
}
|
|
|
|
static inline void
|
|
otx_ep_ring_doorbell(struct otx_ep_device *otx_ep __rte_unused,
|
|
struct otx_ep_instr_queue *iq)
|
|
{
|
|
rte_wmb();
|
|
rte_write64(iq->fill_cnt, iq->doorbell_reg);
|
|
iq->fill_cnt = 0;
|
|
}
|
|
|
|
static inline int
|
|
post_iqcmd(struct otx_ep_instr_queue *iq, uint8_t *iqcmd)
|
|
{
|
|
uint8_t *iqptr, cmdsize;
|
|
|
|
/* This ensures that the read index does not wrap around to
|
|
* the same position if queue gets full before OCTEON TX2 could
|
|
* fetch any instr.
|
|
*/
|
|
if (iq->instr_pending > (iq->nb_desc - 1))
|
|
return OTX_EP_IQ_SEND_FAILED;
|
|
|
|
/* Copy cmd into iq */
|
|
cmdsize = 64;
|
|
iqptr = iq->base_addr + (iq->host_write_index << 6);
|
|
|
|
rte_memcpy(iqptr, iqcmd, cmdsize);
|
|
|
|
/* Increment the host write index */
|
|
iq->host_write_index =
|
|
otx_ep_incr_index(iq->host_write_index, 1, iq->nb_desc);
|
|
|
|
iq->fill_cnt++;
|
|
|
|
/* Flush the command into memory. We need to be sure the data
|
|
* is in memory before indicating that the instruction is
|
|
* pending.
|
|
*/
|
|
iq->instr_pending++;
|
|
/* OTX_EP_IQ_SEND_SUCCESS */
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int
|
|
otx_ep_send_data(struct otx_ep_device *otx_ep, struct otx_ep_instr_queue *iq,
|
|
void *cmd, int dbell)
|
|
{
|
|
uint32_t ret;
|
|
|
|
/* Submit IQ command */
|
|
ret = post_iqcmd(iq, cmd);
|
|
|
|
if (ret == OTX_EP_IQ_SEND_SUCCESS) {
|
|
if (dbell)
|
|
otx_ep_ring_doorbell(otx_ep, iq);
|
|
iq->stats.instr_posted++;
|
|
|
|
} else {
|
|
iq->stats.instr_dropped++;
|
|
if (iq->fill_cnt)
|
|
otx_ep_ring_doorbell(otx_ep, iq);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static inline void
|
|
set_sg_size(struct otx_ep_sg_entry *sg_entry, uint16_t size, uint32_t pos)
|
|
{
|
|
#if RTE_BYTE_ORDER == RTE_BIG_ENDIAN
|
|
sg_entry->u.size[pos] = size;
|
|
#elif RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN
|
|
sg_entry->u.size[3 - pos] = size;
|
|
#endif
|
|
}
|
|
|
|
/* Enqueue requests/packets to OTX_EP IQ queue.
|
|
* returns number of requests enqueued successfully
|
|
*/
|
|
uint16_t
|
|
otx_ep_xmit_pkts(void *tx_queue, struct rte_mbuf **pkts, uint16_t nb_pkts)
|
|
{
|
|
struct otx_ep_instr_64B iqcmd;
|
|
struct otx_ep_instr_queue *iq;
|
|
struct otx_ep_device *otx_ep;
|
|
struct rte_mbuf *m;
|
|
|
|
uint32_t iqreq_type, sgbuf_sz;
|
|
int dbell, index, count = 0;
|
|
unsigned int pkt_len, i;
|
|
int gather, gsz;
|
|
void *iqreq_buf;
|
|
uint64_t dptr;
|
|
|
|
iq = (struct otx_ep_instr_queue *)tx_queue;
|
|
otx_ep = iq->otx_ep_dev;
|
|
|
|
iqcmd.ih.u64 = 0;
|
|
iqcmd.pki_ih3.u64 = 0;
|
|
iqcmd.irh.u64 = 0;
|
|
|
|
/* ih invars */
|
|
iqcmd.ih.s.fsz = OTX_EP_FSZ;
|
|
iqcmd.ih.s.pkind = otx_ep->pkind; /* The SDK decided PKIND value */
|
|
|
|
/* pki ih3 invars */
|
|
iqcmd.pki_ih3.s.w = 1;
|
|
iqcmd.pki_ih3.s.utt = 1;
|
|
iqcmd.pki_ih3.s.tagtype = ORDERED_TAG;
|
|
/* sl will be sizeof(pki_ih3) */
|
|
iqcmd.pki_ih3.s.sl = OTX_EP_FSZ + OTX_CUST_DATA_LEN;
|
|
|
|
/* irh invars */
|
|
iqcmd.irh.s.opcode = OTX_EP_NW_PKT_OP;
|
|
|
|
for (i = 0; i < nb_pkts; i++) {
|
|
m = pkts[i];
|
|
if (m->nb_segs == 1) {
|
|
/* dptr */
|
|
dptr = rte_mbuf_data_iova(m);
|
|
pkt_len = rte_pktmbuf_data_len(m);
|
|
iqreq_buf = m;
|
|
iqreq_type = OTX_EP_REQTYPE_NORESP_NET;
|
|
gather = 0;
|
|
gsz = 0;
|
|
} else {
|
|
struct otx_ep_buf_free_info *finfo;
|
|
int j, frags, num_sg;
|
|
|
|
if (!(otx_ep->tx_offloads & RTE_ETH_TX_OFFLOAD_MULTI_SEGS))
|
|
goto xmit_fail;
|
|
|
|
finfo = (struct otx_ep_buf_free_info *)rte_malloc(NULL,
|
|
sizeof(*finfo), 0);
|
|
if (finfo == NULL) {
|
|
otx_ep_err("free buffer alloc failed\n");
|
|
goto xmit_fail;
|
|
}
|
|
num_sg = (m->nb_segs + 3) / 4;
|
|
sgbuf_sz = sizeof(struct otx_ep_sg_entry) * num_sg;
|
|
finfo->g.sg =
|
|
rte_zmalloc(NULL, sgbuf_sz, OTX_EP_SG_ALIGN);
|
|
if (finfo->g.sg == NULL) {
|
|
rte_free(finfo);
|
|
otx_ep_err("sg entry alloc failed\n");
|
|
goto xmit_fail;
|
|
}
|
|
gather = 1;
|
|
gsz = m->nb_segs;
|
|
finfo->g.num_sg = num_sg;
|
|
finfo->g.sg[0].ptr[0] = rte_mbuf_data_iova(m);
|
|
set_sg_size(&finfo->g.sg[0], m->data_len, 0);
|
|
pkt_len = m->data_len;
|
|
finfo->mbuf = m;
|
|
|
|
frags = m->nb_segs - 1;
|
|
j = 1;
|
|
m = m->next;
|
|
while (frags--) {
|
|
finfo->g.sg[(j >> 2)].ptr[(j & 3)] =
|
|
rte_mbuf_data_iova(m);
|
|
set_sg_size(&finfo->g.sg[(j >> 2)],
|
|
m->data_len, (j & 3));
|
|
pkt_len += m->data_len;
|
|
j++;
|
|
m = m->next;
|
|
}
|
|
dptr = rte_mem_virt2iova(finfo->g.sg);
|
|
iqreq_buf = finfo;
|
|
iqreq_type = OTX_EP_REQTYPE_NORESP_GATHER;
|
|
if (pkt_len > OTX_EP_MAX_PKT_SZ) {
|
|
rte_free(finfo->g.sg);
|
|
rte_free(finfo);
|
|
otx_ep_err("failed\n");
|
|
goto xmit_fail;
|
|
}
|
|
}
|
|
/* ih vars */
|
|
iqcmd.ih.s.tlen = pkt_len + iqcmd.ih.s.fsz;
|
|
iqcmd.ih.s.gather = gather;
|
|
iqcmd.ih.s.gsz = gsz;
|
|
|
|
iqcmd.dptr = dptr;
|
|
otx_ep_swap_8B_data(&iqcmd.irh.u64, 1);
|
|
|
|
#ifdef OTX_EP_IO_DEBUG
|
|
otx_ep_dbg("After swapping\n");
|
|
otx_ep_dbg("Word0 [dptr]: 0x%016lx\n",
|
|
(unsigned long)iqcmd.dptr);
|
|
otx_ep_dbg("Word1 [ihtx]: 0x%016lx\n", (unsigned long)iqcmd.ih);
|
|
otx_ep_dbg("Word2 [pki_ih3]: 0x%016lx\n",
|
|
(unsigned long)iqcmd.pki_ih3);
|
|
otx_ep_dbg("Word3 [rptr]: 0x%016lx\n",
|
|
(unsigned long)iqcmd.rptr);
|
|
otx_ep_dbg("Word4 [irh]: 0x%016lx\n", (unsigned long)iqcmd.irh);
|
|
otx_ep_dbg("Word5 [exhdr[0]]: 0x%016lx\n",
|
|
(unsigned long)iqcmd.exhdr[0]);
|
|
rte_pktmbuf_dump(stdout, m, rte_pktmbuf_pkt_len(m));
|
|
#endif
|
|
dbell = (i == (unsigned int)(nb_pkts - 1)) ? 1 : 0;
|
|
index = iq->host_write_index;
|
|
if (otx_ep_send_data(otx_ep, iq, &iqcmd, dbell))
|
|
goto xmit_fail;
|
|
otx_ep_iqreq_add(iq, iqreq_buf, iqreq_type, index);
|
|
iq->stats.tx_pkts++;
|
|
iq->stats.tx_bytes += pkt_len;
|
|
count++;
|
|
}
|
|
|
|
xmit_fail:
|
|
if (iq->instr_pending >= OTX_EP_MAX_INSTR)
|
|
otx_ep_flush_iq(iq);
|
|
|
|
/* Return no# of instructions posted successfully. */
|
|
return count;
|
|
}
|
|
|
|
/* Enqueue requests/packets to OTX_EP IQ queue.
|
|
* returns number of requests enqueued successfully
|
|
*/
|
|
uint16_t
|
|
otx2_ep_xmit_pkts(void *tx_queue, struct rte_mbuf **pkts, uint16_t nb_pkts)
|
|
{
|
|
struct otx2_ep_instr_64B iqcmd2;
|
|
struct otx_ep_instr_queue *iq;
|
|
struct otx_ep_device *otx_ep;
|
|
uint64_t dptr;
|
|
int count = 0;
|
|
unsigned int i;
|
|
struct rte_mbuf *m;
|
|
unsigned int pkt_len;
|
|
void *iqreq_buf;
|
|
uint32_t iqreq_type, sgbuf_sz;
|
|
int gather, gsz;
|
|
int dbell;
|
|
int index;
|
|
|
|
iq = (struct otx_ep_instr_queue *)tx_queue;
|
|
otx_ep = iq->otx_ep_dev;
|
|
|
|
iqcmd2.ih.u64 = 0;
|
|
iqcmd2.irh.u64 = 0;
|
|
|
|
/* ih invars */
|
|
iqcmd2.ih.s.fsz = OTX2_EP_FSZ;
|
|
iqcmd2.ih.s.pkind = otx_ep->pkind; /* The SDK decided PKIND value */
|
|
/* irh invars */
|
|
iqcmd2.irh.s.opcode = OTX_EP_NW_PKT_OP;
|
|
|
|
for (i = 0; i < nb_pkts; i++) {
|
|
m = pkts[i];
|
|
if (m->nb_segs == 1) {
|
|
/* dptr */
|
|
dptr = rte_mbuf_data_iova(m);
|
|
pkt_len = rte_pktmbuf_data_len(m);
|
|
iqreq_buf = m;
|
|
iqreq_type = OTX_EP_REQTYPE_NORESP_NET;
|
|
gather = 0;
|
|
gsz = 0;
|
|
} else {
|
|
struct otx_ep_buf_free_info *finfo;
|
|
int j, frags, num_sg;
|
|
|
|
if (!(otx_ep->tx_offloads & RTE_ETH_TX_OFFLOAD_MULTI_SEGS))
|
|
goto xmit_fail;
|
|
|
|
finfo = (struct otx_ep_buf_free_info *)
|
|
rte_malloc(NULL, sizeof(*finfo), 0);
|
|
if (finfo == NULL) {
|
|
otx_ep_err("free buffer alloc failed\n");
|
|
goto xmit_fail;
|
|
}
|
|
num_sg = (m->nb_segs + 3) / 4;
|
|
sgbuf_sz = sizeof(struct otx_ep_sg_entry) * num_sg;
|
|
finfo->g.sg =
|
|
rte_zmalloc(NULL, sgbuf_sz, OTX_EP_SG_ALIGN);
|
|
if (finfo->g.sg == NULL) {
|
|
rte_free(finfo);
|
|
otx_ep_err("sg entry alloc failed\n");
|
|
goto xmit_fail;
|
|
}
|
|
gather = 1;
|
|
gsz = m->nb_segs;
|
|
finfo->g.num_sg = num_sg;
|
|
finfo->g.sg[0].ptr[0] = rte_mbuf_data_iova(m);
|
|
set_sg_size(&finfo->g.sg[0], m->data_len, 0);
|
|
pkt_len = m->data_len;
|
|
finfo->mbuf = m;
|
|
|
|
frags = m->nb_segs - 1;
|
|
j = 1;
|
|
m = m->next;
|
|
while (frags--) {
|
|
finfo->g.sg[(j >> 2)].ptr[(j & 3)] =
|
|
rte_mbuf_data_iova(m);
|
|
set_sg_size(&finfo->g.sg[(j >> 2)],
|
|
m->data_len, (j & 3));
|
|
pkt_len += m->data_len;
|
|
j++;
|
|
m = m->next;
|
|
}
|
|
dptr = rte_mem_virt2iova(finfo->g.sg);
|
|
iqreq_buf = finfo;
|
|
iqreq_type = OTX_EP_REQTYPE_NORESP_GATHER;
|
|
if (pkt_len > OTX_EP_MAX_PKT_SZ) {
|
|
rte_free(finfo->g.sg);
|
|
rte_free(finfo);
|
|
otx_ep_err("failed\n");
|
|
goto xmit_fail;
|
|
}
|
|
}
|
|
/* ih vars */
|
|
iqcmd2.ih.s.tlen = pkt_len + iqcmd2.ih.s.fsz;
|
|
iqcmd2.ih.s.gather = gather;
|
|
iqcmd2.ih.s.gsz = gsz;
|
|
iqcmd2.dptr = dptr;
|
|
otx_ep_swap_8B_data(&iqcmd2.irh.u64, 1);
|
|
|
|
#ifdef OTX_EP_IO_DEBUG
|
|
otx_ep_dbg("After swapping\n");
|
|
otx_ep_dbg("Word0 [dptr]: 0x%016lx\n",
|
|
(unsigned long)iqcmd.dptr);
|
|
otx_ep_dbg("Word1 [ihtx]: 0x%016lx\n", (unsigned long)iqcmd.ih);
|
|
otx_ep_dbg("Word2 [pki_ih3]: 0x%016lx\n",
|
|
(unsigned long)iqcmd.pki_ih3);
|
|
otx_ep_dbg("Word3 [rptr]: 0x%016lx\n",
|
|
(unsigned long)iqcmd.rptr);
|
|
otx_ep_dbg("Word4 [irh]: 0x%016lx\n", (unsigned long)iqcmd.irh);
|
|
otx_ep_dbg("Word5 [exhdr[0]]: 0x%016lx\n",
|
|
(unsigned long)iqcmd.exhdr[0]);
|
|
#endif
|
|
index = iq->host_write_index;
|
|
dbell = (i == (unsigned int)(nb_pkts - 1)) ? 1 : 0;
|
|
if (otx_ep_send_data(otx_ep, iq, &iqcmd2, dbell))
|
|
goto xmit_fail;
|
|
otx_ep_iqreq_add(iq, iqreq_buf, iqreq_type, index);
|
|
iq->stats.tx_pkts++;
|
|
iq->stats.tx_bytes += pkt_len;
|
|
count++;
|
|
}
|
|
|
|
xmit_fail:
|
|
if (iq->instr_pending >= OTX_EP_MAX_INSTR)
|
|
otx_ep_flush_iq(iq);
|
|
|
|
/* Return no# of instructions posted successfully. */
|
|
return count;
|
|
}
|
|
|
|
static uint32_t
|
|
otx_ep_droq_refill(struct otx_ep_droq *droq)
|
|
{
|
|
struct otx_ep_droq_desc *desc_ring;
|
|
struct otx_ep_droq_info *info;
|
|
struct rte_mbuf *buf = NULL;
|
|
uint32_t desc_refilled = 0;
|
|
|
|
desc_ring = droq->desc_ring;
|
|
|
|
while (droq->refill_count && (desc_refilled < droq->nb_desc)) {
|
|
/* If a valid buffer exists (happens if there is no dispatch),
|
|
* reuse the buffer, else allocate.
|
|
*/
|
|
if (droq->recv_buf_list[droq->refill_idx] != NULL)
|
|
break;
|
|
|
|
buf = rte_pktmbuf_alloc(droq->mpool);
|
|
/* If a buffer could not be allocated, no point in
|
|
* continuing
|
|
*/
|
|
if (buf == NULL) {
|
|
droq->stats.rx_alloc_failure++;
|
|
break;
|
|
}
|
|
info = rte_pktmbuf_mtod(buf, struct otx_ep_droq_info *);
|
|
memset(info, 0, sizeof(*info));
|
|
|
|
droq->recv_buf_list[droq->refill_idx] = buf;
|
|
desc_ring[droq->refill_idx].buffer_ptr =
|
|
rte_mbuf_data_iova_default(buf);
|
|
|
|
|
|
droq->refill_idx = otx_ep_incr_index(droq->refill_idx, 1,
|
|
droq->nb_desc);
|
|
|
|
desc_refilled++;
|
|
droq->refill_count--;
|
|
}
|
|
|
|
return desc_refilled;
|
|
}
|
|
|
|
static struct rte_mbuf *
|
|
otx_ep_droq_read_packet(struct otx_ep_device *otx_ep,
|
|
struct otx_ep_droq *droq, int next_fetch)
|
|
{
|
|
volatile struct otx_ep_droq_info *info;
|
|
struct rte_mbuf *droq_pkt2 = NULL;
|
|
struct rte_mbuf *droq_pkt = NULL;
|
|
struct rte_net_hdr_lens hdr_lens;
|
|
struct otx_ep_droq_info *info2;
|
|
uint64_t total_pkt_len;
|
|
uint32_t pkt_len = 0;
|
|
int next_idx;
|
|
|
|
droq_pkt = droq->recv_buf_list[droq->read_idx];
|
|
droq_pkt2 = droq->recv_buf_list[droq->read_idx];
|
|
info = rte_pktmbuf_mtod(droq_pkt, struct otx_ep_droq_info *);
|
|
/* make sure info is available */
|
|
rte_rmb();
|
|
if (unlikely(!info->length)) {
|
|
int retry = OTX_EP_MAX_DELAYED_PKT_RETRIES;
|
|
/* otx_ep_dbg("OCTEON DROQ[%d]: read_idx: %d; Data not ready "
|
|
* "yet, Retry; pending=%lu\n", droq->q_no, droq->read_idx,
|
|
* droq->pkts_pending);
|
|
*/
|
|
droq->stats.pkts_delayed_data++;
|
|
while (retry && !info->length)
|
|
retry--;
|
|
if (!retry && !info->length) {
|
|
otx_ep_err("OCTEON DROQ[%d]: read_idx: %d; Retry failed !!\n",
|
|
droq->q_no, droq->read_idx);
|
|
/* May be zero length packet; drop it */
|
|
rte_pktmbuf_free(droq_pkt);
|
|
droq->recv_buf_list[droq->read_idx] = NULL;
|
|
droq->read_idx = otx_ep_incr_index(droq->read_idx, 1,
|
|
droq->nb_desc);
|
|
droq->stats.dropped_zlp++;
|
|
droq->refill_count++;
|
|
goto oq_read_fail;
|
|
}
|
|
}
|
|
if (next_fetch) {
|
|
next_idx = otx_ep_incr_index(droq->read_idx, 1, droq->nb_desc);
|
|
droq_pkt2 = droq->recv_buf_list[next_idx];
|
|
info2 = rte_pktmbuf_mtod(droq_pkt2, struct otx_ep_droq_info *);
|
|
rte_prefetch_non_temporal((const void *)info2);
|
|
}
|
|
|
|
info->length = rte_bswap64(info->length);
|
|
/* Deduce the actual data size */
|
|
total_pkt_len = info->length + INFO_SIZE;
|
|
if (total_pkt_len <= droq->buffer_size) {
|
|
info->length -= OTX_EP_RH_SIZE;
|
|
droq_pkt = droq->recv_buf_list[droq->read_idx];
|
|
if (likely(droq_pkt != NULL)) {
|
|
droq_pkt->data_off += OTX_EP_DROQ_INFO_SIZE;
|
|
/* otx_ep_dbg("OQ: pkt_len[%ld], buffer_size %d\n",
|
|
* (long)info->length, droq->buffer_size);
|
|
*/
|
|
pkt_len = (uint32_t)info->length;
|
|
droq_pkt->pkt_len = pkt_len;
|
|
droq_pkt->data_len = pkt_len;
|
|
droq_pkt->port = otx_ep->port_id;
|
|
droq->recv_buf_list[droq->read_idx] = NULL;
|
|
droq->read_idx = otx_ep_incr_index(droq->read_idx, 1,
|
|
droq->nb_desc);
|
|
droq->refill_count++;
|
|
}
|
|
} else {
|
|
struct rte_mbuf *first_buf = NULL;
|
|
struct rte_mbuf *last_buf = NULL;
|
|
|
|
while (pkt_len < total_pkt_len) {
|
|
int cpy_len = 0;
|
|
|
|
cpy_len = ((pkt_len + droq->buffer_size) >
|
|
total_pkt_len)
|
|
? ((uint32_t)total_pkt_len -
|
|
pkt_len)
|
|
: droq->buffer_size;
|
|
|
|
droq_pkt = droq->recv_buf_list[droq->read_idx];
|
|
droq->recv_buf_list[droq->read_idx] = NULL;
|
|
|
|
if (likely(droq_pkt != NULL)) {
|
|
/* Note the first seg */
|
|
if (!pkt_len)
|
|
first_buf = droq_pkt;
|
|
|
|
droq_pkt->port = otx_ep->port_id;
|
|
if (!pkt_len) {
|
|
droq_pkt->data_off +=
|
|
OTX_EP_DROQ_INFO_SIZE;
|
|
droq_pkt->pkt_len =
|
|
cpy_len - OTX_EP_DROQ_INFO_SIZE;
|
|
droq_pkt->data_len =
|
|
cpy_len - OTX_EP_DROQ_INFO_SIZE;
|
|
} else {
|
|
droq_pkt->pkt_len = cpy_len;
|
|
droq_pkt->data_len = cpy_len;
|
|
}
|
|
|
|
if (pkt_len) {
|
|
first_buf->nb_segs++;
|
|
first_buf->pkt_len += droq_pkt->pkt_len;
|
|
}
|
|
|
|
if (last_buf)
|
|
last_buf->next = droq_pkt;
|
|
|
|
last_buf = droq_pkt;
|
|
} else {
|
|
otx_ep_err("no buf\n");
|
|
}
|
|
|
|
pkt_len += cpy_len;
|
|
droq->read_idx = otx_ep_incr_index(droq->read_idx, 1,
|
|
droq->nb_desc);
|
|
droq->refill_count++;
|
|
}
|
|
droq_pkt = first_buf;
|
|
}
|
|
droq_pkt->packet_type = rte_net_get_ptype(droq_pkt, &hdr_lens,
|
|
RTE_PTYPE_ALL_MASK);
|
|
droq_pkt->l2_len = hdr_lens.l2_len;
|
|
droq_pkt->l3_len = hdr_lens.l3_len;
|
|
droq_pkt->l4_len = hdr_lens.l4_len;
|
|
|
|
if (droq_pkt->nb_segs > 1 &&
|
|
!(otx_ep->rx_offloads & RTE_ETH_RX_OFFLOAD_SCATTER)) {
|
|
rte_pktmbuf_free(droq_pkt);
|
|
goto oq_read_fail;
|
|
}
|
|
|
|
return droq_pkt;
|
|
|
|
oq_read_fail:
|
|
return NULL;
|
|
}
|
|
|
|
static inline uint32_t
|
|
otx_ep_check_droq_pkts(struct otx_ep_droq *droq)
|
|
{
|
|
volatile uint64_t pkt_count;
|
|
uint32_t new_pkts;
|
|
|
|
/* Latest available OQ packets */
|
|
pkt_count = rte_read32(droq->pkts_sent_reg);
|
|
rte_write32(pkt_count, droq->pkts_sent_reg);
|
|
new_pkts = pkt_count;
|
|
droq->pkts_pending += new_pkts;
|
|
return new_pkts;
|
|
}
|
|
|
|
/* Check for response arrival from OCTEON TX2
|
|
* returns number of requests completed
|
|
*/
|
|
uint16_t
|
|
otx_ep_recv_pkts(void *rx_queue,
|
|
struct rte_mbuf **rx_pkts,
|
|
uint16_t budget)
|
|
{
|
|
struct otx_ep_droq *droq = rx_queue;
|
|
struct otx_ep_device *otx_ep;
|
|
struct rte_mbuf *oq_pkt;
|
|
|
|
uint32_t pkts = 0;
|
|
uint32_t new_pkts = 0;
|
|
int next_fetch;
|
|
|
|
otx_ep = droq->otx_ep_dev;
|
|
|
|
if (droq->pkts_pending > budget) {
|
|
new_pkts = budget;
|
|
} else {
|
|
new_pkts = droq->pkts_pending;
|
|
new_pkts += otx_ep_check_droq_pkts(droq);
|
|
if (new_pkts > budget)
|
|
new_pkts = budget;
|
|
}
|
|
|
|
if (!new_pkts)
|
|
goto update_credit; /* No pkts at this moment */
|
|
|
|
for (pkts = 0; pkts < new_pkts; pkts++) {
|
|
/* Push the received pkt to application */
|
|
next_fetch = (pkts == new_pkts - 1) ? 0 : 1;
|
|
oq_pkt = otx_ep_droq_read_packet(otx_ep, droq, next_fetch);
|
|
if (!oq_pkt) {
|
|
RTE_LOG_DP(ERR, PMD,
|
|
"DROQ read pkt failed pending %" PRIu64
|
|
"last_pkt_count %" PRIu64 "new_pkts %d.\n",
|
|
droq->pkts_pending, droq->last_pkt_count,
|
|
new_pkts);
|
|
droq->pkts_pending -= pkts;
|
|
droq->stats.rx_err++;
|
|
goto finish;
|
|
}
|
|
rx_pkts[pkts] = oq_pkt;
|
|
/* Stats */
|
|
droq->stats.pkts_received++;
|
|
droq->stats.bytes_received += oq_pkt->pkt_len;
|
|
}
|
|
droq->pkts_pending -= pkts;
|
|
|
|
/* Refill DROQ buffers */
|
|
update_credit:
|
|
if (droq->refill_count >= DROQ_REFILL_THRESHOLD) {
|
|
int desc_refilled = otx_ep_droq_refill(droq);
|
|
|
|
/* Flush the droq descriptor data to memory to be sure
|
|
* that when we update the credits the data in memory is
|
|
* accurate.
|
|
*/
|
|
rte_wmb();
|
|
rte_write32(desc_refilled, droq->pkts_credit_reg);
|
|
} else {
|
|
/*
|
|
* SDP output goes into DROP state when output doorbell count
|
|
* goes below drop count. When door bell count is written with
|
|
* a value greater than drop count SDP output should come out
|
|
* of DROP state. Due to a race condition this is not happening.
|
|
* Writing doorbell register with 0 again may make SDP output
|
|
* come out of this state.
|
|
*/
|
|
|
|
rte_write32(0, droq->pkts_credit_reg);
|
|
}
|
|
finish:
|
|
return pkts;
|
|
}
|