2014-05-27 18:18:41 +00:00
|
|
|
/*-
|
|
|
|
* Copyright (c) 2014 Chelsio Communications, Inc.
|
|
|
|
* All rights reserved.
|
|
|
|
* Written by: Navdeep Parhar <np@FreeBSD.org>
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
|
|
* SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <sys/cdefs.h>
|
|
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
|
|
|
|
#include "opt_inet.h"
|
|
|
|
#include "opt_inet6.h"
|
|
|
|
|
|
|
|
#ifdef DEV_NETMAP
|
|
|
|
#include <sys/param.h>
|
2015-12-03 00:02:01 +00:00
|
|
|
#include <sys/bus.h>
|
2014-05-27 18:18:41 +00:00
|
|
|
#include <sys/eventhandler.h>
|
|
|
|
#include <sys/lock.h>
|
|
|
|
#include <sys/mbuf.h>
|
2015-12-03 00:02:01 +00:00
|
|
|
#include <sys/module.h>
|
2014-05-27 18:18:41 +00:00
|
|
|
#include <sys/selinfo.h>
|
|
|
|
#include <sys/socket.h>
|
|
|
|
#include <sys/sockio.h>
|
|
|
|
#include <machine/bus.h>
|
|
|
|
#include <net/ethernet.h>
|
|
|
|
#include <net/if.h>
|
|
|
|
#include <net/if_media.h>
|
|
|
|
#include <net/if_var.h>
|
|
|
|
#include <net/if_clone.h>
|
|
|
|
#include <net/if_types.h>
|
|
|
|
#include <net/netmap.h>
|
|
|
|
#include <dev/netmap/netmap_kern.h>
|
|
|
|
|
|
|
|
#include "common/common.h"
|
|
|
|
#include "common/t4_regs.h"
|
|
|
|
#include "common/t4_regs_values.h"
|
|
|
|
|
|
|
|
extern int fl_pad; /* XXXNM */
|
|
|
|
|
2015-03-06 20:39:19 +00:00
|
|
|
SYSCTL_NODE(_hw, OID_AUTO, cxgbe, CTLFLAG_RD, 0, "cxgbe netmap parameters");
|
|
|
|
|
2015-03-06 20:41:28 +00:00
|
|
|
/*
|
|
|
|
* 0 = normal netmap rx
|
|
|
|
* 1 = black hole
|
|
|
|
* 2 = supermassive black hole (buffer packing enabled)
|
|
|
|
*/
|
|
|
|
int black_hole = 0;
|
|
|
|
SYSCTL_INT(_hw_cxgbe, OID_AUTO, nm_black_hole, CTLFLAG_RDTUN, &black_hole, 0,
|
|
|
|
"Sink incoming packets.");
|
|
|
|
|
2015-03-06 20:39:19 +00:00
|
|
|
int rx_ndesc = 256;
|
|
|
|
SYSCTL_INT(_hw_cxgbe, OID_AUTO, nm_rx_ndesc, CTLFLAG_RWTUN,
|
|
|
|
&rx_ndesc, 0, "# of rx descriptors after which the hw cidx is updated.");
|
|
|
|
|
|
|
|
int holdoff_tmr_idx = 2;
|
|
|
|
SYSCTL_INT(_hw_cxgbe, OID_AUTO, nm_holdoff_tmr_idx, CTLFLAG_RWTUN,
|
|
|
|
&holdoff_tmr_idx, 0, "Holdoff timer index for netmap rx queues.");
|
|
|
|
|
2015-07-06 20:56:59 +00:00
|
|
|
/*
|
|
|
|
* Congestion drops.
|
|
|
|
* -1: no congestion feedback (not recommended).
|
|
|
|
* 0: backpressure the channel instead of dropping packets right away.
|
|
|
|
* 1: no backpressure, drop packets for the congested queue immediately.
|
|
|
|
*/
|
|
|
|
static int nm_cong_drop = 1;
|
|
|
|
TUNABLE_INT("hw.cxgbe.nm_cong_drop", &nm_cong_drop);
|
|
|
|
|
2014-05-27 18:18:41 +00:00
|
|
|
static int
|
2015-12-03 00:02:01 +00:00
|
|
|
alloc_nm_rxq_hwq(struct vi_info *vi, struct sge_nm_rxq *nm_rxq, int cong)
|
2014-05-27 18:18:41 +00:00
|
|
|
{
|
2015-02-24 18:40:10 +00:00
|
|
|
int rc, cntxt_id, i;
|
2014-05-27 18:18:41 +00:00
|
|
|
__be32 v;
|
2015-12-03 00:02:01 +00:00
|
|
|
struct adapter *sc = vi->pi->adapter;
|
2016-03-08 00:23:56 +00:00
|
|
|
struct sge_params *sp = &sc->params.sge;
|
2015-12-03 00:02:01 +00:00
|
|
|
struct netmap_adapter *na = NA(vi->ifp);
|
2014-05-27 18:18:41 +00:00
|
|
|
struct fw_iq_cmd c;
|
|
|
|
|
|
|
|
MPASS(na != NULL);
|
|
|
|
MPASS(nm_rxq->iq_desc != NULL);
|
|
|
|
MPASS(nm_rxq->fl_desc != NULL);
|
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
bzero(nm_rxq->iq_desc, vi->qsize_rxq * IQ_ESIZE);
|
2016-03-08 00:23:56 +00:00
|
|
|
bzero(nm_rxq->fl_desc, na->num_rx_desc * EQ_ESIZE + sp->spg_len);
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
bzero(&c, sizeof(c));
|
|
|
|
c.op_to_vfn = htobe32(V_FW_CMD_OP(FW_IQ_CMD) | F_FW_CMD_REQUEST |
|
|
|
|
F_FW_CMD_WRITE | F_FW_CMD_EXEC | V_FW_IQ_CMD_PFN(sc->pf) |
|
|
|
|
V_FW_IQ_CMD_VFN(0));
|
|
|
|
c.alloc_to_len16 = htobe32(F_FW_IQ_CMD_ALLOC | F_FW_IQ_CMD_IQSTART |
|
|
|
|
FW_LEN16(c));
|
2015-12-03 00:02:01 +00:00
|
|
|
if (vi->flags & INTR_RXQ) {
|
2014-05-27 18:18:41 +00:00
|
|
|
KASSERT(nm_rxq->intr_idx < sc->intr_count,
|
|
|
|
("%s: invalid direct intr_idx %d", __func__,
|
|
|
|
nm_rxq->intr_idx));
|
|
|
|
v = V_FW_IQ_CMD_IQANDSTINDEX(nm_rxq->intr_idx);
|
|
|
|
} else {
|
|
|
|
CXGBE_UNIMPLEMENTED(__func__); /* XXXNM: needs review */
|
|
|
|
v = V_FW_IQ_CMD_IQANDSTINDEX(nm_rxq->intr_idx) |
|
|
|
|
F_FW_IQ_CMD_IQANDST;
|
|
|
|
}
|
|
|
|
c.type_to_iqandstindex = htobe32(v |
|
|
|
|
V_FW_IQ_CMD_TYPE(FW_IQ_TYPE_FL_INT_CAP) |
|
2015-12-03 00:02:01 +00:00
|
|
|
V_FW_IQ_CMD_VIID(vi->viid) |
|
2014-05-27 18:18:41 +00:00
|
|
|
V_FW_IQ_CMD_IQANUD(X_UPDATEDELIVERY_INTERRUPT));
|
2015-12-03 00:02:01 +00:00
|
|
|
c.iqdroprss_to_iqesize = htobe16(V_FW_IQ_CMD_IQPCIECH(vi->pi->tx_chan) |
|
2014-05-27 18:18:41 +00:00
|
|
|
F_FW_IQ_CMD_IQGTSMODE |
|
|
|
|
V_FW_IQ_CMD_IQINTCNTTHRESH(0) |
|
2014-08-02 00:56:34 +00:00
|
|
|
V_FW_IQ_CMD_IQESIZE(ilog2(IQ_ESIZE) - 4));
|
2015-12-03 00:02:01 +00:00
|
|
|
c.iqsize = htobe16(vi->qsize_rxq);
|
2014-05-27 18:18:41 +00:00
|
|
|
c.iqaddr = htobe64(nm_rxq->iq_ba);
|
2015-02-24 18:40:10 +00:00
|
|
|
if (cong >= 0) {
|
|
|
|
c.iqns_to_fl0congen = htobe32(F_FW_IQ_CMD_IQFLINTCONGEN |
|
|
|
|
V_FW_IQ_CMD_FL0CNGCHMAP(cong) | F_FW_IQ_CMD_FL0CONGCIF |
|
|
|
|
F_FW_IQ_CMD_FL0CONGEN);
|
|
|
|
}
|
2014-05-27 18:18:41 +00:00
|
|
|
c.iqns_to_fl0congen |=
|
|
|
|
htobe32(V_FW_IQ_CMD_FL0HOSTFCMODE(X_HOSTFCMODE_NONE) |
|
|
|
|
F_FW_IQ_CMD_FL0FETCHRO | F_FW_IQ_CMD_FL0DATARO |
|
2015-03-06 20:41:28 +00:00
|
|
|
(fl_pad ? F_FW_IQ_CMD_FL0PADEN : 0) |
|
|
|
|
(black_hole == 2 ? F_FW_IQ_CMD_FL0PACKEN : 0));
|
2014-05-27 18:18:41 +00:00
|
|
|
c.fl0dcaen_to_fl0cidxfthresh =
|
2015-06-05 00:37:46 +00:00
|
|
|
htobe16(V_FW_IQ_CMD_FL0FBMIN(X_FETCHBURSTMIN_128B) |
|
2014-05-27 18:18:41 +00:00
|
|
|
V_FW_IQ_CMD_FL0FBMAX(X_FETCHBURSTMAX_512B));
|
2016-03-08 00:23:56 +00:00
|
|
|
c.fl0size = htobe16(na->num_rx_desc / 8 + sp->spg_len / EQ_ESIZE);
|
2014-05-27 18:18:41 +00:00
|
|
|
c.fl0addr = htobe64(nm_rxq->fl_ba);
|
|
|
|
|
|
|
|
rc = -t4_wr_mbox(sc, sc->mbox, &c, sizeof(c), &c);
|
|
|
|
if (rc != 0) {
|
|
|
|
device_printf(sc->dev,
|
|
|
|
"failed to create netmap ingress queue: %d\n", rc);
|
|
|
|
return (rc);
|
|
|
|
}
|
|
|
|
|
|
|
|
nm_rxq->iq_cidx = 0;
|
2016-03-08 00:23:56 +00:00
|
|
|
MPASS(nm_rxq->iq_sidx == vi->qsize_rxq - sp->spg_len / IQ_ESIZE);
|
2014-05-27 18:18:41 +00:00
|
|
|
nm_rxq->iq_gen = F_RSPD_GEN;
|
|
|
|
nm_rxq->iq_cntxt_id = be16toh(c.iqid);
|
|
|
|
nm_rxq->iq_abs_id = be16toh(c.physiqid);
|
|
|
|
cntxt_id = nm_rxq->iq_cntxt_id - sc->sge.iq_start;
|
|
|
|
if (cntxt_id >= sc->sge.niq) {
|
|
|
|
panic ("%s: nm_rxq->iq_cntxt_id (%d) more than the max (%d)",
|
|
|
|
__func__, cntxt_id, sc->sge.niq - 1);
|
|
|
|
}
|
|
|
|
sc->sge.iqmap[cntxt_id] = (void *)nm_rxq;
|
|
|
|
|
|
|
|
nm_rxq->fl_cntxt_id = be16toh(c.fl0id);
|
|
|
|
nm_rxq->fl_pidx = nm_rxq->fl_cidx = 0;
|
|
|
|
MPASS(nm_rxq->fl_sidx == na->num_rx_desc);
|
|
|
|
cntxt_id = nm_rxq->fl_cntxt_id - sc->sge.eq_start;
|
|
|
|
if (cntxt_id >= sc->sge.neq) {
|
|
|
|
panic("%s: nm_rxq->fl_cntxt_id (%d) more than the max (%d)",
|
|
|
|
__func__, cntxt_id, sc->sge.neq - 1);
|
|
|
|
}
|
|
|
|
sc->sge.eqmap[cntxt_id] = (void *)nm_rxq;
|
|
|
|
|
2016-03-04 13:11:13 +00:00
|
|
|
nm_rxq->fl_db_val = V_QID(nm_rxq->fl_cntxt_id) |
|
|
|
|
sc->chip_params->sge_fl_db;
|
2014-05-27 18:18:41 +00:00
|
|
|
|
2015-02-24 18:40:10 +00:00
|
|
|
if (is_t5(sc) && cong >= 0) {
|
|
|
|
uint32_t param, val;
|
|
|
|
|
|
|
|
param = V_FW_PARAMS_MNEM(FW_PARAMS_MNEM_DMAQ) |
|
|
|
|
V_FW_PARAMS_PARAM_X(FW_PARAMS_PARAM_DMAQ_CONM_CTXT) |
|
|
|
|
V_FW_PARAMS_PARAM_YZ(nm_rxq->iq_cntxt_id);
|
|
|
|
param = V_FW_PARAMS_MNEM(FW_PARAMS_MNEM_DMAQ) |
|
|
|
|
V_FW_PARAMS_PARAM_X(FW_PARAMS_PARAM_DMAQ_CONM_CTXT) |
|
|
|
|
V_FW_PARAMS_PARAM_YZ(nm_rxq->iq_cntxt_id);
|
|
|
|
if (cong == 0)
|
|
|
|
val = 1 << 19;
|
|
|
|
else {
|
|
|
|
val = 2 << 19;
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
|
|
if (cong & (1 << i))
|
|
|
|
val |= 1 << (i << 2);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
rc = -t4_set_params(sc, sc->mbox, sc->pf, 0, 1, ¶m, &val);
|
|
|
|
if (rc != 0) {
|
|
|
|
/* report error but carry on */
|
|
|
|
device_printf(sc->dev,
|
|
|
|
"failed to set congestion manager context for "
|
|
|
|
"ingress queue %d: %d\n", nm_rxq->iq_cntxt_id, rc);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-02-24 18:32:17 +00:00
|
|
|
t4_write_reg(sc, MYPF_REG(A_SGE_PF_GTS),
|
2015-03-06 20:39:19 +00:00
|
|
|
V_INGRESSQID(nm_rxq->iq_cntxt_id) |
|
|
|
|
V_SEINTARM(V_QINTR_TIMER_IDX(holdoff_tmr_idx)));
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
return (rc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2015-12-03 00:02:01 +00:00
|
|
|
free_nm_rxq_hwq(struct vi_info *vi, struct sge_nm_rxq *nm_rxq)
|
2014-05-27 18:18:41 +00:00
|
|
|
{
|
2015-12-03 00:02:01 +00:00
|
|
|
struct adapter *sc = vi->pi->adapter;
|
2014-05-27 18:18:41 +00:00
|
|
|
int rc;
|
|
|
|
|
|
|
|
rc = -t4_iq_free(sc, sc->mbox, sc->pf, 0, FW_IQ_TYPE_FL_INT_CAP,
|
|
|
|
nm_rxq->iq_cntxt_id, nm_rxq->fl_cntxt_id, 0xffff);
|
|
|
|
if (rc != 0)
|
|
|
|
device_printf(sc->dev, "%s: failed for iq %d, fl %d: %d\n",
|
|
|
|
__func__, nm_rxq->iq_cntxt_id, nm_rxq->fl_cntxt_id, rc);
|
|
|
|
return (rc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2015-12-03 00:02:01 +00:00
|
|
|
alloc_nm_txq_hwq(struct vi_info *vi, struct sge_nm_txq *nm_txq)
|
2014-05-27 18:18:41 +00:00
|
|
|
{
|
|
|
|
int rc, cntxt_id;
|
|
|
|
size_t len;
|
2015-12-03 00:02:01 +00:00
|
|
|
struct adapter *sc = vi->pi->adapter;
|
|
|
|
struct netmap_adapter *na = NA(vi->ifp);
|
2014-05-27 18:18:41 +00:00
|
|
|
struct fw_eq_eth_cmd c;
|
|
|
|
|
|
|
|
MPASS(na != NULL);
|
|
|
|
MPASS(nm_txq->desc != NULL);
|
|
|
|
|
2016-03-08 00:23:56 +00:00
|
|
|
len = na->num_tx_desc * EQ_ESIZE + sc->params.sge.spg_len;
|
2014-05-27 18:18:41 +00:00
|
|
|
bzero(nm_txq->desc, len);
|
|
|
|
|
|
|
|
bzero(&c, sizeof(c));
|
|
|
|
c.op_to_vfn = htobe32(V_FW_CMD_OP(FW_EQ_ETH_CMD) | F_FW_CMD_REQUEST |
|
|
|
|
F_FW_CMD_WRITE | F_FW_CMD_EXEC | V_FW_EQ_ETH_CMD_PFN(sc->pf) |
|
|
|
|
V_FW_EQ_ETH_CMD_VFN(0));
|
|
|
|
c.alloc_to_len16 = htobe32(F_FW_EQ_ETH_CMD_ALLOC |
|
|
|
|
F_FW_EQ_ETH_CMD_EQSTART | FW_LEN16(c));
|
2015-07-01 00:34:14 +00:00
|
|
|
c.autoequiqe_to_viid = htobe32(F_FW_EQ_ETH_CMD_AUTOEQUIQE |
|
2015-12-03 00:02:01 +00:00
|
|
|
F_FW_EQ_ETH_CMD_AUTOEQUEQE | V_FW_EQ_ETH_CMD_VIID(vi->viid));
|
2014-05-27 18:18:41 +00:00
|
|
|
c.fetchszm_to_iqid =
|
|
|
|
htobe32(V_FW_EQ_ETH_CMD_HOSTFCMODE(X_HOSTFCMODE_NONE) |
|
2015-12-03 00:02:01 +00:00
|
|
|
V_FW_EQ_ETH_CMD_PCIECHN(vi->pi->tx_chan) | F_FW_EQ_ETH_CMD_FETCHRO |
|
2014-05-27 18:18:41 +00:00
|
|
|
V_FW_EQ_ETH_CMD_IQID(sc->sge.nm_rxq[nm_txq->iqidx].iq_cntxt_id));
|
|
|
|
c.dcaen_to_eqsize = htobe32(V_FW_EQ_ETH_CMD_FBMIN(X_FETCHBURSTMIN_64B) |
|
|
|
|
V_FW_EQ_ETH_CMD_FBMAX(X_FETCHBURSTMAX_512B) |
|
|
|
|
V_FW_EQ_ETH_CMD_EQSIZE(len / EQ_ESIZE));
|
|
|
|
c.eqaddr = htobe64(nm_txq->ba);
|
|
|
|
|
|
|
|
rc = -t4_wr_mbox(sc, sc->mbox, &c, sizeof(c), &c);
|
|
|
|
if (rc != 0) {
|
2015-12-03 00:02:01 +00:00
|
|
|
device_printf(vi->dev,
|
2014-05-27 18:18:41 +00:00
|
|
|
"failed to create netmap egress queue: %d\n", rc);
|
|
|
|
return (rc);
|
|
|
|
}
|
|
|
|
|
|
|
|
nm_txq->cntxt_id = G_FW_EQ_ETH_CMD_EQID(be32toh(c.eqid_pkd));
|
|
|
|
cntxt_id = nm_txq->cntxt_id - sc->sge.eq_start;
|
|
|
|
if (cntxt_id >= sc->sge.neq)
|
|
|
|
panic("%s: nm_txq->cntxt_id (%d) more than the max (%d)", __func__,
|
|
|
|
cntxt_id, sc->sge.neq - 1);
|
|
|
|
sc->sge.eqmap[cntxt_id] = (void *)nm_txq;
|
|
|
|
|
|
|
|
nm_txq->pidx = nm_txq->cidx = 0;
|
|
|
|
MPASS(nm_txq->sidx == na->num_tx_desc);
|
2014-09-09 18:36:00 +00:00
|
|
|
nm_txq->equiqidx = nm_txq->equeqidx = nm_txq->dbidx = 0;
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
nm_txq->doorbells = sc->doorbells;
|
|
|
|
if (isset(&nm_txq->doorbells, DOORBELL_UDB) ||
|
|
|
|
isset(&nm_txq->doorbells, DOORBELL_UDBWC) ||
|
|
|
|
isset(&nm_txq->doorbells, DOORBELL_WCWR)) {
|
2016-03-08 00:23:56 +00:00
|
|
|
uint32_t s_qpp = sc->params.sge.eq_s_qpp;
|
2014-05-27 18:18:41 +00:00
|
|
|
uint32_t mask = (1 << s_qpp) - 1;
|
|
|
|
volatile uint8_t *udb;
|
|
|
|
|
|
|
|
udb = sc->udbs_base + UDBS_DB_OFFSET;
|
|
|
|
udb += (nm_txq->cntxt_id >> s_qpp) << PAGE_SHIFT;
|
|
|
|
nm_txq->udb_qid = nm_txq->cntxt_id & mask;
|
2014-08-02 01:48:25 +00:00
|
|
|
if (nm_txq->udb_qid >= PAGE_SIZE / UDBS_SEG_SIZE)
|
2014-05-27 18:18:41 +00:00
|
|
|
clrbit(&nm_txq->doorbells, DOORBELL_WCWR);
|
|
|
|
else {
|
|
|
|
udb += nm_txq->udb_qid << UDBS_SEG_SHIFT;
|
|
|
|
nm_txq->udb_qid = 0;
|
|
|
|
}
|
|
|
|
nm_txq->udb = (volatile void *)udb;
|
|
|
|
}
|
|
|
|
|
|
|
|
return (rc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2015-12-03 00:02:01 +00:00
|
|
|
free_nm_txq_hwq(struct vi_info *vi, struct sge_nm_txq *nm_txq)
|
2014-05-27 18:18:41 +00:00
|
|
|
{
|
2015-12-03 00:02:01 +00:00
|
|
|
struct adapter *sc = vi->pi->adapter;
|
2014-05-27 18:18:41 +00:00
|
|
|
int rc;
|
|
|
|
|
|
|
|
rc = -t4_eth_eq_free(sc, sc->mbox, sc->pf, 0, nm_txq->cntxt_id);
|
|
|
|
if (rc != 0)
|
|
|
|
device_printf(sc->dev, "%s: failed for eq %d: %d\n", __func__,
|
|
|
|
nm_txq->cntxt_id, rc);
|
|
|
|
return (rc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2015-12-03 00:02:01 +00:00
|
|
|
cxgbe_netmap_on(struct adapter *sc, struct vi_info *vi, struct ifnet *ifp,
|
2014-05-27 18:18:41 +00:00
|
|
|
struct netmap_adapter *na)
|
|
|
|
{
|
|
|
|
struct netmap_slot *slot;
|
|
|
|
struct sge_nm_rxq *nm_rxq;
|
|
|
|
struct sge_nm_txq *nm_txq;
|
|
|
|
int rc, i, j, hwidx;
|
|
|
|
struct hw_buf_info *hwb;
|
|
|
|
|
|
|
|
ASSERT_SYNCHRONIZED_OP(sc);
|
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
if ((vi->flags & VI_INIT_DONE) == 0 ||
|
2014-05-27 18:18:41 +00:00
|
|
|
(ifp->if_drv_flags & IFF_DRV_RUNNING) == 0)
|
|
|
|
return (EAGAIN);
|
|
|
|
|
|
|
|
hwb = &sc->sge.hw_buf_info[0];
|
|
|
|
for (i = 0; i < SGE_FLBUF_SIZES; i++, hwb++) {
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
if (hwb->size == NETMAP_BUF_SIZE(na))
|
2014-05-27 18:18:41 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (i >= SGE_FLBUF_SIZES) {
|
|
|
|
if_printf(ifp, "no hwidx for netmap buffer size %d.\n",
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
NETMAP_BUF_SIZE(na));
|
2014-05-27 18:18:41 +00:00
|
|
|
return (ENXIO);
|
|
|
|
}
|
|
|
|
hwidx = i;
|
|
|
|
|
|
|
|
/* Must set caps before calling netmap_reset */
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
nm_set_native_flags(na);
|
2014-05-27 18:18:41 +00:00
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
for_each_nm_rxq(vi, i, nm_rxq) {
|
2016-06-23 02:53:00 +00:00
|
|
|
struct irq *irq = &sc->irq[vi->first_intr + i];
|
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
alloc_nm_rxq_hwq(vi, nm_rxq, tnl_cong(vi->pi, nm_cong_drop));
|
2014-05-27 18:18:41 +00:00
|
|
|
nm_rxq->fl_hwidx = hwidx;
|
|
|
|
slot = netmap_reset(na, NR_RX, i, 0);
|
|
|
|
MPASS(slot != NULL); /* XXXNM: error check, not assert */
|
|
|
|
|
|
|
|
/* We deal with 8 bufs at a time */
|
|
|
|
MPASS((na->num_rx_desc & 7) == 0);
|
|
|
|
MPASS(na->num_rx_desc == nm_rxq->fl_sidx);
|
2015-03-06 20:39:19 +00:00
|
|
|
for (j = 0; j < nm_rxq->fl_sidx; j++) {
|
2014-05-27 18:18:41 +00:00
|
|
|
uint64_t ba;
|
|
|
|
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
PNMB(na, &slot[j], &ba);
|
2015-03-06 20:39:19 +00:00
|
|
|
MPASS(ba != 0);
|
2014-05-27 18:18:41 +00:00
|
|
|
nm_rxq->fl_desc[j] = htobe64(ba | hwidx);
|
|
|
|
}
|
2015-03-06 20:39:19 +00:00
|
|
|
j = nm_rxq->fl_pidx = nm_rxq->fl_sidx - 8;
|
2014-05-27 18:18:41 +00:00
|
|
|
MPASS((j & 7) == 0);
|
|
|
|
j /= 8; /* driver pidx to hardware pidx */
|
|
|
|
wmb();
|
|
|
|
t4_write_reg(sc, MYPF_REG(A_SGE_PF_KDOORBELL),
|
|
|
|
nm_rxq->fl_db_val | V_PIDX(j));
|
2016-06-23 02:53:00 +00:00
|
|
|
|
|
|
|
atomic_cmpset_int(&irq->nm_state, NM_OFF, NM_ON);
|
2014-05-27 18:18:41 +00:00
|
|
|
}
|
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
for_each_nm_txq(vi, i, nm_txq) {
|
|
|
|
alloc_nm_txq_hwq(vi, nm_txq);
|
2014-05-27 18:18:41 +00:00
|
|
|
slot = netmap_reset(na, NR_TX, i, 0);
|
|
|
|
MPASS(slot != NULL); /* XXXNM: error check, not assert */
|
|
|
|
}
|
|
|
|
|
2016-06-23 02:53:00 +00:00
|
|
|
if (vi->nm_rss == NULL) {
|
|
|
|
vi->nm_rss = malloc(vi->rss_size * sizeof(uint16_t), M_CXGBE,
|
|
|
|
M_ZERO | M_WAITOK);
|
|
|
|
}
|
2015-12-03 00:02:01 +00:00
|
|
|
for (i = 0; i < vi->rss_size;) {
|
|
|
|
for_each_nm_rxq(vi, j, nm_rxq) {
|
2016-06-23 02:53:00 +00:00
|
|
|
vi->nm_rss[i++] = nm_rxq->iq_abs_id;
|
2015-12-03 00:02:01 +00:00
|
|
|
if (i == vi->rss_size)
|
2014-05-27 18:18:41 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2015-12-03 00:02:01 +00:00
|
|
|
rc = -t4_config_rss_range(sc, sc->mbox, vi->viid, 0, vi->rss_size,
|
2016-06-23 02:53:00 +00:00
|
|
|
vi->nm_rss, vi->rss_size);
|
2014-05-27 18:18:41 +00:00
|
|
|
if (rc != 0)
|
|
|
|
if_printf(ifp, "netmap rss_config failed: %d\n", rc);
|
|
|
|
|
|
|
|
return (rc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2015-12-03 00:02:01 +00:00
|
|
|
cxgbe_netmap_off(struct adapter *sc, struct vi_info *vi, struct ifnet *ifp,
|
2014-05-27 18:18:41 +00:00
|
|
|
struct netmap_adapter *na)
|
|
|
|
{
|
|
|
|
int rc, i;
|
|
|
|
struct sge_nm_txq *nm_txq;
|
|
|
|
struct sge_nm_rxq *nm_rxq;
|
|
|
|
|
|
|
|
ASSERT_SYNCHRONIZED_OP(sc);
|
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
if ((vi->flags & VI_INIT_DONE) == 0)
|
|
|
|
return (0);
|
|
|
|
|
2016-06-23 02:53:00 +00:00
|
|
|
rc = -t4_config_rss_range(sc, sc->mbox, vi->viid, 0, vi->rss_size,
|
|
|
|
vi->rss, vi->rss_size);
|
2014-05-27 18:18:41 +00:00
|
|
|
if (rc != 0)
|
2016-06-23 02:53:00 +00:00
|
|
|
if_printf(ifp, "failed to restore RSS config: %d\n", rc);
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
nm_clear_native_flags(na);
|
2014-05-27 18:18:41 +00:00
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
for_each_nm_txq(vi, i, nm_txq) {
|
2015-02-24 18:22:24 +00:00
|
|
|
struct sge_qstat *spg = (void *)&nm_txq->desc[nm_txq->sidx];
|
|
|
|
|
|
|
|
/* Wait for hw pidx to catch up ... */
|
|
|
|
while (be16toh(nm_txq->pidx) != spg->pidx)
|
|
|
|
pause("nmpidx", 1);
|
|
|
|
|
|
|
|
/* ... and then for the cidx. */
|
|
|
|
while (spg->pidx != spg->cidx)
|
|
|
|
pause("nmcidx", 1);
|
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
free_nm_txq_hwq(vi, nm_txq);
|
2014-05-27 18:18:41 +00:00
|
|
|
}
|
2015-12-03 00:02:01 +00:00
|
|
|
for_each_nm_rxq(vi, i, nm_rxq) {
|
2016-06-23 02:53:00 +00:00
|
|
|
struct irq *irq = &sc->irq[vi->first_intr + i];
|
|
|
|
|
|
|
|
while (!atomic_cmpset_int(&irq->nm_state, NM_ON, NM_OFF))
|
|
|
|
pause("nmst", 1);
|
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
free_nm_rxq_hwq(vi, nm_rxq);
|
2014-05-27 18:18:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return (rc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
cxgbe_netmap_reg(struct netmap_adapter *na, int on)
|
|
|
|
{
|
|
|
|
struct ifnet *ifp = na->ifp;
|
2015-12-03 00:02:01 +00:00
|
|
|
struct vi_info *vi = ifp->if_softc;
|
|
|
|
struct adapter *sc = vi->pi->adapter;
|
2014-05-27 18:18:41 +00:00
|
|
|
int rc;
|
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
rc = begin_synchronized_op(sc, vi, SLEEP_OK | INTR_OK, "t4nmreg");
|
2014-05-27 18:18:41 +00:00
|
|
|
if (rc != 0)
|
|
|
|
return (rc);
|
|
|
|
if (on)
|
2015-12-03 00:02:01 +00:00
|
|
|
rc = cxgbe_netmap_on(sc, vi, ifp, na);
|
2014-05-27 18:18:41 +00:00
|
|
|
else
|
2015-12-03 00:02:01 +00:00
|
|
|
rc = cxgbe_netmap_off(sc, vi, ifp, na);
|
2014-05-27 18:18:41 +00:00
|
|
|
end_synchronized_op(sc, 0);
|
|
|
|
|
|
|
|
return (rc);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* How many packets can a single type1 WR carry in n descriptors */
|
|
|
|
static inline int
|
|
|
|
ndesc_to_npkt(const int n)
|
|
|
|
{
|
|
|
|
|
|
|
|
MPASS(n > 0 && n <= SGE_MAX_WR_NDESC);
|
|
|
|
|
|
|
|
return (n * 2 - 1);
|
|
|
|
}
|
|
|
|
#define MAX_NPKT_IN_TYPE1_WR (ndesc_to_npkt(SGE_MAX_WR_NDESC))
|
|
|
|
|
|
|
|
/* Space (in descriptors) needed for a type1 WR that carries n packets */
|
|
|
|
static inline int
|
|
|
|
npkt_to_ndesc(const int n)
|
|
|
|
{
|
|
|
|
|
|
|
|
MPASS(n > 0 && n <= MAX_NPKT_IN_TYPE1_WR);
|
|
|
|
|
|
|
|
return ((n + 2) / 2);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Space (in 16B units) needed for a type1 WR that carries n packets */
|
|
|
|
static inline int
|
|
|
|
npkt_to_len16(const int n)
|
|
|
|
{
|
|
|
|
|
|
|
|
MPASS(n > 0 && n <= MAX_NPKT_IN_TYPE1_WR);
|
|
|
|
|
|
|
|
return (n * 2 + 1);
|
|
|
|
}
|
|
|
|
|
2014-08-02 00:56:34 +00:00
|
|
|
#define NMIDXDIFF(q, idx) IDXDIFF((q)->pidx, (q)->idx, (q)->sidx)
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
static void
|
|
|
|
ring_nm_txq_db(struct adapter *sc, struct sge_nm_txq *nm_txq)
|
|
|
|
{
|
|
|
|
int n;
|
|
|
|
u_int db = nm_txq->doorbells;
|
|
|
|
|
|
|
|
MPASS(nm_txq->pidx != nm_txq->dbidx);
|
|
|
|
|
2014-08-02 00:56:34 +00:00
|
|
|
n = NMIDXDIFF(nm_txq, dbidx);
|
2014-05-27 18:18:41 +00:00
|
|
|
if (n > 1)
|
|
|
|
clrbit(&db, DOORBELL_WCWR);
|
|
|
|
wmb();
|
|
|
|
|
|
|
|
switch (ffs(db) - 1) {
|
|
|
|
case DOORBELL_UDB:
|
|
|
|
*nm_txq->udb = htole32(V_QID(nm_txq->udb_qid) | V_PIDX(n));
|
|
|
|
break;
|
|
|
|
|
|
|
|
case DOORBELL_WCWR: {
|
|
|
|
volatile uint64_t *dst, *src;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Queues whose 128B doorbell segment fits in the page do not
|
|
|
|
* use relative qid (udb_qid is always 0). Only queues with
|
|
|
|
* doorbell segments can do WCWR.
|
|
|
|
*/
|
|
|
|
KASSERT(nm_txq->udb_qid == 0 && n == 1,
|
|
|
|
("%s: inappropriate doorbell (0x%x, %d, %d) for nm_txq %p",
|
|
|
|
__func__, nm_txq->doorbells, n, nm_txq->pidx, nm_txq));
|
|
|
|
|
|
|
|
dst = (volatile void *)((uintptr_t)nm_txq->udb +
|
|
|
|
UDBS_WR_OFFSET - UDBS_DB_OFFSET);
|
|
|
|
src = (void *)&nm_txq->desc[nm_txq->dbidx];
|
|
|
|
while (src != (void *)&nm_txq->desc[nm_txq->dbidx + 1])
|
|
|
|
*dst++ = *src++;
|
|
|
|
wmb();
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case DOORBELL_UDBWC:
|
|
|
|
*nm_txq->udb = htole32(V_QID(nm_txq->udb_qid) | V_PIDX(n));
|
|
|
|
wmb();
|
|
|
|
break;
|
|
|
|
|
|
|
|
case DOORBELL_KDB:
|
|
|
|
t4_write_reg(sc, MYPF_REG(A_SGE_PF_KDOORBELL),
|
|
|
|
V_QID(nm_txq->cntxt_id) | V_PIDX(n));
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
nm_txq->dbidx = nm_txq->pidx;
|
|
|
|
}
|
|
|
|
|
|
|
|
int lazy_tx_credit_flush = 1;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Write work requests to send 'npkt' frames and ring the doorbell to send them
|
|
|
|
* on their way. No need to check for wraparound.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
cxgbe_nm_tx(struct adapter *sc, struct sge_nm_txq *nm_txq,
|
2015-02-24 21:31:13 +00:00
|
|
|
struct netmap_kring *kring, int npkt, int npkt_remaining, int txcsum)
|
2014-05-27 18:18:41 +00:00
|
|
|
{
|
|
|
|
struct netmap_ring *ring = kring->ring;
|
|
|
|
struct netmap_slot *slot;
|
|
|
|
const u_int lim = kring->nkr_num_slots - 1;
|
|
|
|
struct fw_eth_tx_pkts_wr *wr = (void *)&nm_txq->desc[nm_txq->pidx];
|
|
|
|
uint16_t len;
|
|
|
|
uint64_t ba;
|
|
|
|
struct cpl_tx_pkt_core *cpl;
|
|
|
|
struct ulptx_sgl *usgl;
|
|
|
|
int i, n;
|
|
|
|
|
|
|
|
while (npkt) {
|
|
|
|
n = min(npkt, MAX_NPKT_IN_TYPE1_WR);
|
|
|
|
len = 0;
|
|
|
|
|
|
|
|
wr = (void *)&nm_txq->desc[nm_txq->pidx];
|
|
|
|
wr->op_pkd = htobe32(V_FW_WR_OP(FW_ETH_TX_PKTS_WR));
|
|
|
|
wr->equiq_to_len16 = htobe32(V_FW_WR_LEN16(npkt_to_len16(n)));
|
|
|
|
wr->npkt = n;
|
|
|
|
wr->r3 = 0;
|
|
|
|
wr->type = 1;
|
|
|
|
cpl = (void *)(wr + 1);
|
|
|
|
|
|
|
|
for (i = 0; i < n; i++) {
|
|
|
|
slot = &ring->slot[kring->nr_hwcur];
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
PNMB(kring->na, slot, &ba);
|
2015-03-06 20:39:19 +00:00
|
|
|
MPASS(ba != 0);
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
cpl->ctrl0 = nm_txq->cpl_ctrl0;
|
|
|
|
cpl->pack = 0;
|
|
|
|
cpl->len = htobe16(slot->len);
|
|
|
|
/*
|
|
|
|
* netmap(4) says "netmap does not use features such as
|
|
|
|
* checksum offloading, TCP segmentation offloading,
|
|
|
|
* encryption, VLAN encapsulation/decapsulation, etc."
|
|
|
|
*
|
2015-02-24 21:31:13 +00:00
|
|
|
* So the ncxl interfaces have tx hardware checksumming
|
|
|
|
* disabled by default. But you can override netmap by
|
|
|
|
* enabling IFCAP_TXCSUM on the interface manully.
|
2014-05-27 18:18:41 +00:00
|
|
|
*/
|
2015-02-24 21:31:13 +00:00
|
|
|
cpl->ctrl1 = txcsum ? 0 :
|
|
|
|
htobe64(F_TXPKT_IPCSUM_DIS | F_TXPKT_L4CSUM_DIS);
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
usgl = (void *)(cpl + 1);
|
|
|
|
usgl->cmd_nsge = htobe32(V_ULPTX_CMD(ULP_TX_SC_DSGL) |
|
|
|
|
V_ULPTX_NSGE(1));
|
|
|
|
usgl->len0 = htobe32(slot->len);
|
|
|
|
usgl->addr0 = htobe64(ba);
|
|
|
|
|
|
|
|
slot->flags &= ~(NS_REPORT | NS_BUF_CHANGED);
|
|
|
|
cpl = (void *)(usgl + 1);
|
|
|
|
MPASS(slot->len + len <= UINT16_MAX);
|
|
|
|
len += slot->len;
|
|
|
|
kring->nr_hwcur = nm_next(kring->nr_hwcur, lim);
|
|
|
|
}
|
|
|
|
wr->plen = htobe16(len);
|
|
|
|
|
|
|
|
npkt -= n;
|
|
|
|
nm_txq->pidx += npkt_to_ndesc(n);
|
|
|
|
MPASS(nm_txq->pidx <= nm_txq->sidx);
|
|
|
|
if (__predict_false(nm_txq->pidx == nm_txq->sidx)) {
|
|
|
|
/*
|
|
|
|
* This routine doesn't know how to write WRs that wrap
|
|
|
|
* around. Make sure it wasn't asked to.
|
|
|
|
*/
|
|
|
|
MPASS(npkt == 0);
|
|
|
|
nm_txq->pidx = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (npkt == 0 && npkt_remaining == 0) {
|
|
|
|
/* All done. */
|
|
|
|
if (lazy_tx_credit_flush == 0) {
|
|
|
|
wr->equiq_to_len16 |= htobe32(F_FW_WR_EQUEQ |
|
|
|
|
F_FW_WR_EQUIQ);
|
|
|
|
nm_txq->equeqidx = nm_txq->pidx;
|
|
|
|
nm_txq->equiqidx = nm_txq->pidx;
|
|
|
|
}
|
|
|
|
ring_nm_txq_db(sc, nm_txq);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2014-08-02 00:56:34 +00:00
|
|
|
if (NMIDXDIFF(nm_txq, equiqidx) >= nm_txq->sidx / 2) {
|
2014-05-27 18:18:41 +00:00
|
|
|
wr->equiq_to_len16 |= htobe32(F_FW_WR_EQUEQ |
|
|
|
|
F_FW_WR_EQUIQ);
|
|
|
|
nm_txq->equeqidx = nm_txq->pidx;
|
|
|
|
nm_txq->equiqidx = nm_txq->pidx;
|
2014-08-02 00:56:34 +00:00
|
|
|
} else if (NMIDXDIFF(nm_txq, equeqidx) >= 64) {
|
2014-05-27 18:18:41 +00:00
|
|
|
wr->equiq_to_len16 |= htobe32(F_FW_WR_EQUEQ);
|
|
|
|
nm_txq->equeqidx = nm_txq->pidx;
|
|
|
|
}
|
2014-08-02 00:56:34 +00:00
|
|
|
if (NMIDXDIFF(nm_txq, dbidx) >= 2 * SGE_MAX_WR_NDESC)
|
2014-05-27 18:18:41 +00:00
|
|
|
ring_nm_txq_db(sc, nm_txq);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Will get called again. */
|
|
|
|
MPASS(npkt_remaining);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* How many contiguous free descriptors starting at pidx */
|
|
|
|
static inline int
|
|
|
|
contiguous_ndesc_available(struct sge_nm_txq *nm_txq)
|
|
|
|
{
|
|
|
|
|
|
|
|
if (nm_txq->cidx > nm_txq->pidx)
|
|
|
|
return (nm_txq->cidx - nm_txq->pidx - 1);
|
|
|
|
else if (nm_txq->cidx > 0)
|
|
|
|
return (nm_txq->sidx - nm_txq->pidx);
|
|
|
|
else
|
|
|
|
return (nm_txq->sidx - nm_txq->pidx - 1);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
reclaim_nm_tx_desc(struct sge_nm_txq *nm_txq)
|
|
|
|
{
|
|
|
|
struct sge_qstat *spg = (void *)&nm_txq->desc[nm_txq->sidx];
|
|
|
|
uint16_t hw_cidx = spg->cidx; /* snapshot */
|
|
|
|
struct fw_eth_tx_pkts_wr *wr;
|
|
|
|
int n = 0;
|
|
|
|
|
|
|
|
hw_cidx = be16toh(hw_cidx);
|
|
|
|
|
|
|
|
while (nm_txq->cidx != hw_cidx) {
|
|
|
|
wr = (void *)&nm_txq->desc[nm_txq->cidx];
|
|
|
|
|
|
|
|
MPASS(wr->op_pkd == htobe32(V_FW_WR_OP(FW_ETH_TX_PKTS_WR)));
|
|
|
|
MPASS(wr->type == 1);
|
|
|
|
MPASS(wr->npkt > 0 && wr->npkt <= MAX_NPKT_IN_TYPE1_WR);
|
|
|
|
|
|
|
|
n += wr->npkt;
|
|
|
|
nm_txq->cidx += npkt_to_ndesc(wr->npkt);
|
2014-08-02 00:56:34 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* We never sent a WR that wrapped around so the credits coming
|
|
|
|
* back, WR by WR, should never cause the cidx to wrap around
|
|
|
|
* either.
|
|
|
|
*/
|
|
|
|
MPASS(nm_txq->cidx <= nm_txq->sidx);
|
|
|
|
if (__predict_false(nm_txq->cidx == nm_txq->sidx))
|
|
|
|
nm_txq->cidx = 0;
|
2014-05-27 18:18:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return (n);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
cxgbe_netmap_txsync(struct netmap_kring *kring, int flags)
|
2014-05-27 18:18:41 +00:00
|
|
|
{
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
struct netmap_adapter *na = kring->na;
|
2014-05-27 18:18:41 +00:00
|
|
|
struct ifnet *ifp = na->ifp;
|
2015-12-03 00:02:01 +00:00
|
|
|
struct vi_info *vi = ifp->if_softc;
|
|
|
|
struct adapter *sc = vi->pi->adapter;
|
2016-06-23 02:53:00 +00:00
|
|
|
struct sge_nm_txq *nm_txq = &sc->sge.nm_txq[vi->first_nm_txq + kring->ring_id];
|
2014-05-27 18:18:41 +00:00
|
|
|
const u_int head = kring->rhead;
|
|
|
|
u_int reclaimed = 0;
|
2015-02-24 21:31:13 +00:00
|
|
|
int n, d, npkt_remaining, ndesc_remaining, txcsum;
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Tx was at kring->nr_hwcur last time around and now we need to advance
|
|
|
|
* to kring->rhead. Note that the driver's pidx moves independent of
|
|
|
|
* netmap's kring->nr_hwcur (pidx counts descriptors and the relation
|
|
|
|
* between descriptors and frames isn't 1:1).
|
|
|
|
*/
|
|
|
|
|
|
|
|
npkt_remaining = head >= kring->nr_hwcur ? head - kring->nr_hwcur :
|
|
|
|
kring->nkr_num_slots - kring->nr_hwcur + head;
|
2015-02-24 21:31:13 +00:00
|
|
|
txcsum = ifp->if_capenable & (IFCAP_TXCSUM | IFCAP_TXCSUM_IPV6);
|
2014-05-27 18:18:41 +00:00
|
|
|
while (npkt_remaining) {
|
|
|
|
reclaimed += reclaim_nm_tx_desc(nm_txq);
|
|
|
|
ndesc_remaining = contiguous_ndesc_available(nm_txq);
|
|
|
|
/* Can't run out of descriptors with packets still remaining */
|
|
|
|
MPASS(ndesc_remaining > 0);
|
|
|
|
|
|
|
|
/* # of desc needed to tx all remaining packets */
|
|
|
|
d = (npkt_remaining / MAX_NPKT_IN_TYPE1_WR) * SGE_MAX_WR_NDESC;
|
|
|
|
if (npkt_remaining % MAX_NPKT_IN_TYPE1_WR)
|
|
|
|
d += npkt_to_ndesc(npkt_remaining % MAX_NPKT_IN_TYPE1_WR);
|
|
|
|
|
|
|
|
if (d <= ndesc_remaining)
|
|
|
|
n = npkt_remaining;
|
|
|
|
else {
|
|
|
|
/* Can't send all, calculate how many can be sent */
|
|
|
|
n = (ndesc_remaining / SGE_MAX_WR_NDESC) *
|
|
|
|
MAX_NPKT_IN_TYPE1_WR;
|
|
|
|
if (ndesc_remaining % SGE_MAX_WR_NDESC)
|
|
|
|
n += ndesc_to_npkt(ndesc_remaining % SGE_MAX_WR_NDESC);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Send n packets and update nm_txq->pidx and kring->nr_hwcur */
|
|
|
|
npkt_remaining -= n;
|
2015-02-24 21:31:13 +00:00
|
|
|
cxgbe_nm_tx(sc, nm_txq, kring, n, npkt_remaining, txcsum);
|
2014-05-27 18:18:41 +00:00
|
|
|
}
|
|
|
|
MPASS(npkt_remaining == 0);
|
|
|
|
MPASS(kring->nr_hwcur == head);
|
|
|
|
MPASS(nm_txq->dbidx == nm_txq->pidx);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Second part: reclaim buffers for completed transmissions.
|
|
|
|
*/
|
|
|
|
if (reclaimed || flags & NAF_FORCE_RECLAIM || nm_kr_txempty(kring)) {
|
|
|
|
reclaimed += reclaim_nm_tx_desc(nm_txq);
|
|
|
|
kring->nr_hwtail += reclaimed;
|
|
|
|
if (kring->nr_hwtail >= kring->nkr_num_slots)
|
|
|
|
kring->nr_hwtail -= kring->nkr_num_slots;
|
|
|
|
}
|
|
|
|
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
cxgbe_netmap_rxsync(struct netmap_kring *kring, int flags)
|
2014-05-27 18:18:41 +00:00
|
|
|
{
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
struct netmap_adapter *na = kring->na;
|
2014-05-27 18:18:41 +00:00
|
|
|
struct netmap_ring *ring = kring->ring;
|
|
|
|
struct ifnet *ifp = na->ifp;
|
2015-12-03 00:02:01 +00:00
|
|
|
struct vi_info *vi = ifp->if_softc;
|
|
|
|
struct adapter *sc = vi->pi->adapter;
|
2016-06-23 02:53:00 +00:00
|
|
|
struct sge_nm_rxq *nm_rxq = &sc->sge.nm_rxq[vi->first_nm_rxq + kring->ring_id];
|
Sync netmap sources with the version in our private tree.
This commit contains large contributions from Giuseppe Lettieri and
Stefano Garzarella, is partly supported by grants from Verisign and Cisco,
and brings in the following:
- fix zerocopy monitor ports and introduce copying monitor ports
(the latter are lower performance but give access to all traffic
in parallel with the application)
- exclusive open mode, useful to implement solutions that recover
from crashes of the main netmap client (suggested by Patrick Kelsey)
- revised memory allocator in preparation for the 'passthrough mode'
(ptnetmap) recently presented at bsdcan. ptnetmap is described in
S. Garzarella, G. Lettieri, L. Rizzo;
Virtual device passthrough for high speed VM networking,
ACM/IEEE ANCS 2015, Oakland (CA) May 2015
http://info.iet.unipi.it/~luigi/research.html
- fix rx CRC handing on ixl
- add module dependencies for netmap when building drivers as modules
- minor simplifications to device-specific routines (*txsync, *rxsync)
- general code cleanup (remove unused variables, introduce macros
to access rings and remove duplicate code,
Applications do not need to be recompiled, unless of course
they want to use the new features (monitors and exclusive open).
Those willing to try this code on stable/10 can just update the
sys/dev/netmap/*, sys/net/netmap* with the version in HEAD
and apply the small patches to individual device drivers.
MFC after: 1 month
Sponsored by: (partly) Verisign, Cisco
2015-07-10 05:51:36 +00:00
|
|
|
u_int const head = kring->rhead;
|
2014-05-27 18:18:41 +00:00
|
|
|
u_int n;
|
|
|
|
int force_update = (flags & NAF_FORCE_READ) || kring->nr_kflags & NKR_PENDINTR;
|
|
|
|
|
2015-03-06 20:41:28 +00:00
|
|
|
if (black_hole)
|
|
|
|
return (0); /* No updates ever. */
|
|
|
|
|
2014-05-27 18:18:41 +00:00
|
|
|
if (netmap_no_pendintr || force_update) {
|
|
|
|
kring->nr_hwtail = atomic_load_acq_32(&nm_rxq->fl_cidx);
|
|
|
|
kring->nr_kflags &= ~NKR_PENDINTR;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Userspace done with buffers from kring->nr_hwcur to head */
|
|
|
|
n = head >= kring->nr_hwcur ? head - kring->nr_hwcur :
|
|
|
|
kring->nkr_num_slots - kring->nr_hwcur + head;
|
|
|
|
n &= ~7U;
|
|
|
|
if (n > 0) {
|
|
|
|
u_int fl_pidx = nm_rxq->fl_pidx;
|
|
|
|
struct netmap_slot *slot = &ring->slot[fl_pidx];
|
|
|
|
uint64_t ba;
|
|
|
|
int i, dbinc = 0, hwidx = nm_rxq->fl_hwidx;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We always deal with 8 buffers at a time. We must have
|
|
|
|
* stopped at an 8B boundary (fl_pidx) last time around and we
|
|
|
|
* must have a multiple of 8B buffers to give to the freelist.
|
|
|
|
*/
|
|
|
|
MPASS((fl_pidx & 7) == 0);
|
|
|
|
MPASS((n & 7) == 0);
|
|
|
|
|
2014-08-02 00:56:34 +00:00
|
|
|
IDXINCR(kring->nr_hwcur, n, kring->nkr_num_slots);
|
|
|
|
IDXINCR(nm_rxq->fl_pidx, n, nm_rxq->fl_sidx);
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
while (n > 0) {
|
|
|
|
for (i = 0; i < 8; i++, fl_pidx++, slot++) {
|
Update to the current version of netmap.
Mostly bugfixes or features developed in the past 6 months,
so this is a 10.1 candidate.
Basically no user API changes (some bugfixes in sys/net/netmap_user.h).
In detail:
1. netmap support for virtio-net, including in netmap mode.
Under bhyve and with a netmap backend [2] we reach over 1Mpps
with standard APIs (e.g. libpcap), and 5-8 Mpps in netmap mode.
2. (kernel) add support for multiple memory allocators, so we can
better partition physical and virtual interfaces giving access
to separate users. The most visible effect is one additional
argument to the various kernel functions to compute buffer
addresses. All netmap-supported drivers are affected, but changes
are mechanical and trivial
3. (kernel) simplify the prototype for *txsync() and *rxsync()
driver methods. All netmap drivers affected, changes mostly mechanical.
4. add support for netmap-monitor ports. Think of it as a mirroring
port on a physical switch: a netmap monitor port replicates traffic
present on the main port. Restrictions apply. Drive carefully.
5. if_lem.c: support for various paravirtualization features,
experimental and disabled by default.
Most of these are described in our ANCS'13 paper [1].
Paravirtualized support in netmap mode is new, and beats the
numbers in the paper by a large factor (under qemu-kvm,
we measured gues-host throughput up to 10-12 Mpps).
A lot of refactoring and additional documentation in the files
in sys/dev/netmap, but apart from #2 and #3 above, almost nothing
of this stuff is visible to other kernel parts.
Example programs in tools/tools/netmap have been updated with bugfixes
and to support more of the existing features.
This is meant to go into 10.1 so we plan an MFC before the Aug.22 deadline.
A lot of this code has been contributed by my colleagues at UNIPI,
including Giuseppe Lettieri, Vincenzo Maffione, Stefano Garzarella.
MFC after: 3 days.
2014-08-16 15:00:01 +00:00
|
|
|
PNMB(na, slot, &ba);
|
2015-03-06 20:39:19 +00:00
|
|
|
MPASS(ba != 0);
|
2014-05-27 18:18:41 +00:00
|
|
|
nm_rxq->fl_desc[fl_pidx] = htobe64(ba | hwidx);
|
|
|
|
slot->flags &= ~NS_BUF_CHANGED;
|
|
|
|
MPASS(fl_pidx <= nm_rxq->fl_sidx);
|
|
|
|
}
|
|
|
|
n -= 8;
|
|
|
|
if (fl_pidx == nm_rxq->fl_sidx) {
|
|
|
|
fl_pidx = 0;
|
|
|
|
slot = &ring->slot[0];
|
|
|
|
}
|
|
|
|
if (++dbinc == 8 && n >= 32) {
|
|
|
|
wmb();
|
|
|
|
t4_write_reg(sc, MYPF_REG(A_SGE_PF_KDOORBELL),
|
|
|
|
nm_rxq->fl_db_val | V_PIDX(dbinc));
|
|
|
|
dbinc = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
MPASS(nm_rxq->fl_pidx == fl_pidx);
|
|
|
|
|
|
|
|
if (dbinc > 0) {
|
|
|
|
wmb();
|
|
|
|
t4_write_reg(sc, MYPF_REG(A_SGE_PF_KDOORBELL),
|
|
|
|
nm_rxq->fl_db_val | V_PIDX(dbinc));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
2016-06-23 02:53:00 +00:00
|
|
|
void
|
|
|
|
cxgbe_nm_attach(struct vi_info *vi)
|
2014-05-27 18:18:41 +00:00
|
|
|
{
|
2015-12-03 00:02:01 +00:00
|
|
|
struct port_info *pi;
|
|
|
|
struct adapter *sc;
|
2014-05-27 18:18:41 +00:00
|
|
|
struct netmap_adapter na;
|
|
|
|
|
2016-06-23 02:53:00 +00:00
|
|
|
MPASS(vi->nnmrxq > 0);
|
|
|
|
MPASS(vi->ifp != NULL);
|
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
pi = vi->pi;
|
|
|
|
sc = pi->adapter;
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
bzero(&na, sizeof(na));
|
|
|
|
|
2016-06-23 02:53:00 +00:00
|
|
|
na.ifp = vi->ifp;
|
2014-05-27 18:18:41 +00:00
|
|
|
na.na_flags = NAF_BDG_MAYSLEEP;
|
|
|
|
|
|
|
|
/* Netmap doesn't know about the space reserved for the status page. */
|
2016-03-08 00:23:56 +00:00
|
|
|
na.num_tx_desc = vi->qsize_txq - sc->params.sge.spg_len / EQ_ESIZE;
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* The freelist's cidx/pidx drives netmap's rx cidx/pidx. So
|
|
|
|
* num_rx_desc is based on the number of buffers that can be held in the
|
|
|
|
* freelist, and not the number of entries in the iq. (These two are
|
|
|
|
* not exactly the same due to the space taken up by the status page).
|
|
|
|
*/
|
2016-04-30 14:41:18 +00:00
|
|
|
na.num_rx_desc = rounddown(vi->qsize_rxq, 8);
|
2014-05-27 18:18:41 +00:00
|
|
|
na.nm_txsync = cxgbe_netmap_txsync;
|
|
|
|
na.nm_rxsync = cxgbe_netmap_rxsync;
|
|
|
|
na.nm_register = cxgbe_netmap_reg;
|
2016-06-23 02:53:00 +00:00
|
|
|
na.num_tx_rings = vi->nnmtxq;
|
|
|
|
na.num_rx_rings = vi->nnmrxq;
|
2014-05-27 18:18:41 +00:00
|
|
|
netmap_attach(&na); /* This adds IFCAP_NETMAP to if_capabilities */
|
|
|
|
}
|
|
|
|
|
2016-06-23 02:53:00 +00:00
|
|
|
void
|
|
|
|
cxgbe_nm_detach(struct vi_info *vi)
|
2014-05-27 18:18:41 +00:00
|
|
|
{
|
2015-12-03 00:02:01 +00:00
|
|
|
|
2016-06-23 02:53:00 +00:00
|
|
|
MPASS(vi->nnmrxq > 0);
|
|
|
|
MPASS(vi->ifp != NULL);
|
2014-05-27 18:18:41 +00:00
|
|
|
|
2015-12-03 00:02:01 +00:00
|
|
|
netmap_detach(vi->ifp);
|
2014-05-27 18:18:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
handle_nm_fw6_msg(struct adapter *sc, struct ifnet *ifp,
|
|
|
|
const struct cpl_fw6_msg *cpl)
|
|
|
|
{
|
|
|
|
const struct cpl_sge_egr_update *egr;
|
|
|
|
uint32_t oq;
|
|
|
|
struct sge_nm_txq *nm_txq;
|
|
|
|
|
|
|
|
if (cpl->type != FW_TYPE_RSSCPL && cpl->type != FW6_TYPE_RSSCPL)
|
|
|
|
panic("%s: FW_TYPE 0x%x on nm_rxq.", __func__, cpl->type);
|
|
|
|
|
|
|
|
/* data[0] is RSS header */
|
|
|
|
egr = (const void *)&cpl->data[1];
|
|
|
|
oq = be32toh(egr->opcode_qid);
|
|
|
|
MPASS(G_CPL_OPCODE(oq) == CPL_SGE_EGR_UPDATE);
|
|
|
|
nm_txq = (void *)sc->sge.eqmap[G_EGR_QID(oq) - sc->sge.eq_start];
|
|
|
|
|
|
|
|
netmap_tx_irq(ifp, nm_txq->nid);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
t4_nm_intr(void *arg)
|
|
|
|
{
|
|
|
|
struct sge_nm_rxq *nm_rxq = arg;
|
2015-12-03 00:02:01 +00:00
|
|
|
struct vi_info *vi = nm_rxq->vi;
|
|
|
|
struct adapter *sc = vi->pi->adapter;
|
|
|
|
struct ifnet *ifp = vi->ifp;
|
2014-05-27 18:18:41 +00:00
|
|
|
struct netmap_adapter *na = NA(ifp);
|
|
|
|
struct netmap_kring *kring = &na->rx_rings[nm_rxq->nid];
|
|
|
|
struct netmap_ring *ring = kring->ring;
|
2014-08-02 00:56:34 +00:00
|
|
|
struct iq_desc *d = &nm_rxq->iq_desc[nm_rxq->iq_cidx];
|
2014-05-27 18:18:41 +00:00
|
|
|
uint32_t lq;
|
2015-03-06 20:39:19 +00:00
|
|
|
u_int n = 0, work = 0;
|
2014-05-27 18:18:41 +00:00
|
|
|
uint8_t opcode;
|
|
|
|
uint32_t fl_cidx = atomic_load_acq_32(&nm_rxq->fl_cidx);
|
2015-03-06 20:41:28 +00:00
|
|
|
u_int fl_credits = fl_cidx & 7;
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
while ((d->rsp.u.type_gen & F_RSPD_GEN) == nm_rxq->iq_gen) {
|
|
|
|
|
|
|
|
rmb();
|
|
|
|
|
|
|
|
lq = be32toh(d->rsp.pldbuflen_qid);
|
|
|
|
opcode = d->rss.opcode;
|
|
|
|
|
|
|
|
switch (G_RSPD_TYPE(d->rsp.u.type_gen)) {
|
|
|
|
case X_RSPD_TYPE_FLBUF:
|
2015-03-06 20:41:28 +00:00
|
|
|
if (black_hole != 2) {
|
|
|
|
/* No buffer packing so new buf every time */
|
|
|
|
MPASS(lq & F_RSPD_NEWBUF);
|
|
|
|
}
|
2014-05-27 18:18:41 +00:00
|
|
|
|
|
|
|
/* fall through */
|
|
|
|
|
|
|
|
case X_RSPD_TYPE_CPL:
|
|
|
|
MPASS(opcode < NUM_CPL_CMDS);
|
|
|
|
|
|
|
|
switch (opcode) {
|
|
|
|
case CPL_FW4_MSG:
|
|
|
|
case CPL_FW6_MSG:
|
2014-08-02 00:56:34 +00:00
|
|
|
handle_nm_fw6_msg(sc, ifp,
|
|
|
|
(const void *)&d->cpl[0]);
|
2014-05-27 18:18:41 +00:00
|
|
|
break;
|
|
|
|
case CPL_RX_PKT:
|
2016-03-08 00:23:56 +00:00
|
|
|
ring->slot[fl_cidx].len = G_RSPD_LEN(lq) -
|
|
|
|
sc->params.sge.fl_pktshift;
|
2014-05-27 18:18:41 +00:00
|
|
|
ring->slot[fl_cidx].flags = kring->nkr_slot_flags;
|
2015-03-06 20:41:28 +00:00
|
|
|
fl_cidx += (lq & F_RSPD_NEWBUF) ? 1 : 0;
|
|
|
|
fl_credits += (lq & F_RSPD_NEWBUF) ? 1 : 0;
|
|
|
|
if (__predict_false(fl_cidx == nm_rxq->fl_sidx))
|
2014-05-27 18:18:41 +00:00
|
|
|
fl_cidx = 0;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
panic("%s: unexpected opcode 0x%x on nm_rxq %p",
|
|
|
|
__func__, opcode, nm_rxq);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case X_RSPD_TYPE_INTR:
|
|
|
|
/* Not equipped to handle forwarded interrupts. */
|
|
|
|
panic("%s: netmap queue received interrupt for iq %u\n",
|
|
|
|
__func__, lq);
|
|
|
|
|
|
|
|
default:
|
|
|
|
panic("%s: illegal response type %d on nm_rxq %p",
|
|
|
|
__func__, G_RSPD_TYPE(d->rsp.u.type_gen), nm_rxq);
|
|
|
|
}
|
|
|
|
|
|
|
|
d++;
|
|
|
|
if (__predict_false(++nm_rxq->iq_cidx == nm_rxq->iq_sidx)) {
|
|
|
|
nm_rxq->iq_cidx = 0;
|
|
|
|
d = &nm_rxq->iq_desc[0];
|
|
|
|
nm_rxq->iq_gen ^= F_RSPD_GEN;
|
|
|
|
}
|
|
|
|
|
2015-03-06 20:39:19 +00:00
|
|
|
if (__predict_false(++n == rx_ndesc)) {
|
|
|
|
atomic_store_rel_32(&nm_rxq->fl_cidx, fl_cidx);
|
2015-03-06 20:41:28 +00:00
|
|
|
if (black_hole && fl_credits >= 8) {
|
|
|
|
fl_credits /= 8;
|
|
|
|
IDXINCR(nm_rxq->fl_pidx, fl_credits * 8,
|
|
|
|
nm_rxq->fl_sidx);
|
|
|
|
t4_write_reg(sc, MYPF_REG(A_SGE_PF_KDOORBELL),
|
|
|
|
nm_rxq->fl_db_val | V_PIDX(fl_credits));
|
|
|
|
fl_credits = fl_cidx & 7;
|
|
|
|
} else if (!black_hole) {
|
|
|
|
netmap_rx_irq(ifp, nm_rxq->nid, &work);
|
|
|
|
MPASS(work != 0);
|
|
|
|
}
|
2014-05-27 18:18:41 +00:00
|
|
|
t4_write_reg(sc, MYPF_REG(A_SGE_PF_GTS),
|
|
|
|
V_CIDXINC(n) | V_INGRESSQID(nm_rxq->iq_cntxt_id) |
|
|
|
|
V_SEINTARM(V_QINTR_TIMER_IDX(X_TIMERREG_UPDATE_CIDX)));
|
|
|
|
n = 0;
|
|
|
|
}
|
|
|
|
}
|
2015-03-06 20:41:28 +00:00
|
|
|
|
|
|
|
atomic_store_rel_32(&nm_rxq->fl_cidx, fl_cidx);
|
|
|
|
if (black_hole) {
|
|
|
|
fl_credits /= 8;
|
|
|
|
IDXINCR(nm_rxq->fl_pidx, fl_credits * 8, nm_rxq->fl_sidx);
|
|
|
|
t4_write_reg(sc, MYPF_REG(A_SGE_PF_KDOORBELL),
|
|
|
|
nm_rxq->fl_db_val | V_PIDX(fl_credits));
|
|
|
|
} else
|
2015-03-06 20:39:19 +00:00
|
|
|
netmap_rx_irq(ifp, nm_rxq->nid, &work);
|
2015-03-06 20:41:28 +00:00
|
|
|
|
2014-05-27 18:18:41 +00:00
|
|
|
t4_write_reg(sc, MYPF_REG(A_SGE_PF_GTS), V_CIDXINC(n) |
|
2015-02-24 18:32:17 +00:00
|
|
|
V_INGRESSQID((u32)nm_rxq->iq_cntxt_id) |
|
2015-03-06 20:39:19 +00:00
|
|
|
V_SEINTARM(V_QINTR_TIMER_IDX(holdoff_tmr_idx)));
|
2014-05-27 18:18:41 +00:00
|
|
|
}
|
|
|
|
#endif
|