Add optimizing LRO wrapper:

- Add optimizing LRO wrapper which pre-sorts all incoming packets
  according to the hash type and flowid. This prevents exhaustion of
  the LRO entries due to too many connections at the same time.
  Testing using a larger number of higher bandwidth TCP connections
  showed that the incoming ACK packet aggregation rate increased from
  ~1.3:1 to almost 3:1. Another test showed that for a number of TCP
  connections greater than 16 per hardware receive ring, where 8 TCP
  connections was the LRO active entry limit, there was a significant
  improvement in throughput due to being able to fully aggregate more
  than 8 TCP stream. For very few very high bandwidth TCP streams, the
  optimizing LRO wrapper will add CPU usage instead of reducing CPU
  usage. This is expected. Network drivers which want to use the
  optimizing LRO wrapper needs to call "tcp_lro_queue_mbuf()" instead
  of "tcp_lro_rx()" and "tcp_lro_flush_all()" instead of
  "tcp_lro_flush()". Further the LRO control structure must be
  initialized using "tcp_lro_init_args()" passing a non-zero number
  into the "lro_mbufs" argument.

- Make LRO statistics 64-bit. Previously 32-bit integers were used for
  statistics which can be prone to wrap-around. Fix this while at it
  and update all SYSCTL's which expose LRO statistics.

- Ensure all data is freed when destroying a LRO control structures,
  especially leftover LRO entries.

- Reduce number of memory allocations needed when setting up a LRO
  control structure by precomputing the total amount of memory needed.

- Add own memory allocation counter for LRO.

- Bump the FreeBSD version to force recompilation of all KLDs due to
  change of the LRO control structure size.

Sponsored by:	Mellanox Technologies
Reviewed by:	gallatin, sbruno, rrs, gnn, transport
Tested by:	Netflix
Differential Revision:	https://reviews.freebsd.org/D4914
This commit is contained in:
Hans Petter Selasky 2016-01-19 15:33:28 +00:00
parent f16550ad09
commit e936121d31
Notes: svn2git 2020-12-20 02:59:44 +00:00
svn path=/head/; revision=294327
10 changed files with 199 additions and 44 deletions

View File

@ -3579,11 +3579,11 @@ t3_add_configured_sysctls(adapter_t *sc)
CTLTYPE_STRING | CTLFLAG_RD, &qs->txq[TXQ_CTRL], CTLTYPE_STRING | CTLFLAG_RD, &qs->txq[TXQ_CTRL],
0, t3_dump_txq_ctrl, "A", "dump of the transmit queue"); 0, t3_dump_txq_ctrl, "A", "dump of the transmit queue");
SYSCTL_ADD_INT(ctx, lropoidlist, OID_AUTO, "lro_queued", SYSCTL_ADD_U64(ctx, lropoidlist, OID_AUTO, "lro_queued",
CTLFLAG_RD, &qs->lro.ctrl.lro_queued, 0, NULL); CTLFLAG_RD, &qs->lro.ctrl.lro_queued, 0, NULL);
SYSCTL_ADD_INT(ctx, lropoidlist, OID_AUTO, "lro_flushed", SYSCTL_ADD_U64(ctx, lropoidlist, OID_AUTO, "lro_flushed",
CTLFLAG_RD, &qs->lro.ctrl.lro_flushed, 0, NULL); CTLFLAG_RD, &qs->lro.ctrl.lro_flushed, 0, NULL);
SYSCTL_ADD_INT(ctx, lropoidlist, OID_AUTO, "lro_bad_csum", SYSCTL_ADD_U64(ctx, lropoidlist, OID_AUTO, "lro_bad_csum",
CTLFLAG_RD, &qs->lro.ctrl.lro_bad_csum, 0, NULL); CTLFLAG_RD, &qs->lro.ctrl.lro_bad_csum, 0, NULL);
SYSCTL_ADD_INT(ctx, lropoidlist, OID_AUTO, "lro_cnt", SYSCTL_ADD_INT(ctx, lropoidlist, OID_AUTO, "lro_cnt",
CTLFLAG_RD, &qs->lro.ctrl.lro_cnt, 0, NULL); CTLFLAG_RD, &qs->lro.ctrl.lro_cnt, 0, NULL);

View File

@ -2939,9 +2939,9 @@ alloc_rxq(struct vi_info *vi, struct sge_rxq *rxq, int intr_idx, int idx,
CTLTYPE_INT | CTLFLAG_RD, &rxq->iq.cidx, 0, sysctl_uint16, "I", CTLTYPE_INT | CTLFLAG_RD, &rxq->iq.cidx, 0, sysctl_uint16, "I",
"consumer index"); "consumer index");
#if defined(INET) || defined(INET6) #if defined(INET) || defined(INET6)
SYSCTL_ADD_INT(&vi->ctx, children, OID_AUTO, "lro_queued", CTLFLAG_RD, SYSCTL_ADD_U64(&vi->ctx, children, OID_AUTO, "lro_queued", CTLFLAG_RD,
&rxq->lro.lro_queued, 0, NULL); &rxq->lro.lro_queued, 0, NULL);
SYSCTL_ADD_INT(&vi->ctx, children, OID_AUTO, "lro_flushed", CTLFLAG_RD, SYSCTL_ADD_U64(&vi->ctx, children, OID_AUTO, "lro_flushed", CTLFLAG_RD,
&rxq->lro.lro_flushed, 0, NULL); &rxq->lro.lro_flushed, 0, NULL);
#endif #endif
SYSCTL_ADD_UQUAD(&vi->ctx, children, OID_AUTO, "rxcsum", CTLFLAG_RD, SYSCTL_ADD_UQUAD(&vi->ctx, children, OID_AUTO, "rxcsum", CTLFLAG_RD,

View File

@ -5914,10 +5914,10 @@ igb_add_hw_stats(struct adapter *adapter)
SYSCTL_ADD_QUAD(ctx, queue_list, OID_AUTO, "rx_bytes", SYSCTL_ADD_QUAD(ctx, queue_list, OID_AUTO, "rx_bytes",
CTLFLAG_RD, &rxr->rx_bytes, CTLFLAG_RD, &rxr->rx_bytes,
"Queue Bytes Received"); "Queue Bytes Received");
SYSCTL_ADD_UINT(ctx, queue_list, OID_AUTO, "lro_queued", SYSCTL_ADD_U64(ctx, queue_list, OID_AUTO, "lro_queued",
CTLFLAG_RD, &lro->lro_queued, 0, CTLFLAG_RD, &lro->lro_queued, 0,
"LRO Queued"); "LRO Queued");
SYSCTL_ADD_UINT(ctx, queue_list, OID_AUTO, "lro_flushed", SYSCTL_ADD_U64(ctx, queue_list, OID_AUTO, "lro_flushed",
CTLFLAG_RD, &lro->lro_flushed, 0, CTLFLAG_RD, &lro->lro_flushed, 0,
"LRO Flushed"); "LRO Flushed");
} }

View File

@ -405,9 +405,9 @@ netvsc_attach(device_t dev)
ctx = device_get_sysctl_ctx(dev); ctx = device_get_sysctl_ctx(dev);
child = SYSCTL_CHILDREN(device_get_sysctl_tree(dev)); child = SYSCTL_CHILDREN(device_get_sysctl_tree(dev));
SYSCTL_ADD_INT(ctx, child, OID_AUTO, "lro_queued", SYSCTL_ADD_U64(ctx, child, OID_AUTO, "lro_queued",
CTLFLAG_RW, &sc->hn_lro.lro_queued, 0, "LRO queued"); CTLFLAG_RW, &sc->hn_lro.lro_queued, 0, "LRO queued");
SYSCTL_ADD_INT(ctx, child, OID_AUTO, "lro_flushed", SYSCTL_ADD_U64(ctx, child, OID_AUTO, "lro_flushed",
CTLFLAG_RW, &sc->hn_lro.lro_flushed, 0, "LRO flushed"); CTLFLAG_RW, &sc->hn_lro.lro_flushed, 0, "LRO flushed");
SYSCTL_ADD_ULONG(ctx, child, OID_AUTO, "lro_tried", SYSCTL_ADD_ULONG(ctx, child, OID_AUTO, "lro_tried",
CTLFLAG_RW, &sc->hn_lro_tried, "# of LRO tries"); CTLFLAG_RW, &sc->hn_lro_tried, "# of LRO tries");

View File

@ -4476,10 +4476,10 @@ ixgbe_add_hw_stats(struct adapter *adapter)
SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "rx_copies", SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "rx_copies",
CTLFLAG_RD, &rxr->rx_copies, CTLFLAG_RD, &rxr->rx_copies,
"Copied RX Frames"); "Copied RX Frames");
SYSCTL_ADD_INT(ctx, queue_list, OID_AUTO, "lro_queued", SYSCTL_ADD_U64(ctx, queue_list, OID_AUTO, "lro_queued",
CTLFLAG_RD, &lro->lro_queued, 0, CTLFLAG_RD, &lro->lro_queued, 0,
"LRO Queued"); "LRO Queued");
SYSCTL_ADD_INT(ctx, queue_list, OID_AUTO, "lro_flushed", SYSCTL_ADD_U64(ctx, queue_list, OID_AUTO, "lro_flushed",
CTLFLAG_RD, &lro->lro_flushed, 0, CTLFLAG_RD, &lro->lro_flushed, 0,
"LRO Flushed"); "LRO Flushed");
} }

View File

@ -2167,10 +2167,10 @@ ixv_print_debug_info(struct adapter *adapter)
rxr->me, (long long)rxr->rx_packets); rxr->me, (long long)rxr->rx_packets);
device_printf(dev,"RX(%d) Bytes Received: %lu\n", device_printf(dev,"RX(%d) Bytes Received: %lu\n",
rxr->me, (long)rxr->rx_bytes); rxr->me, (long)rxr->rx_bytes);
device_printf(dev,"RX(%d) LRO Queued= %d\n", device_printf(dev,"RX(%d) LRO Queued= %lld\n",
rxr->me, lro->lro_queued); rxr->me, (long long)lro->lro_queued);
device_printf(dev,"RX(%d) LRO Flushed= %d\n", device_printf(dev,"RX(%d) LRO Flushed= %lld\n",
rxr->me, lro->lro_flushed); rxr->me, (long long)lro->lro_flushed);
device_printf(dev,"TX(%d) Packets Sent: %lu\n", device_printf(dev,"TX(%d) Packets Sent: %lu\n",
txr->me, (long)txr->total_packets); txr->me, (long)txr->total_packets);
device_printf(dev,"TX(%d) NO Desc Avail: %lu\n", device_printf(dev,"TX(%d) NO Desc Avail: %lu\n",

View File

@ -1637,15 +1637,15 @@ mxge_add_sysctls(mxge_softc_t *sc)
"rx_big_cnt", "rx_big_cnt",
CTLFLAG_RD, &ss->rx_big.cnt, CTLFLAG_RD, &ss->rx_big.cnt,
0, "rx_small_cnt"); 0, "rx_small_cnt");
SYSCTL_ADD_INT(ctx, children, OID_AUTO, SYSCTL_ADD_U64(ctx, children, OID_AUTO,
"lro_flushed", CTLFLAG_RD, &ss->lc.lro_flushed, "lro_flushed", CTLFLAG_RD, &ss->lc.lro_flushed,
0, "number of lro merge queues flushed"); 0, "number of lro merge queues flushed");
SYSCTL_ADD_INT(ctx, children, OID_AUTO, SYSCTL_ADD_U64(ctx, children, OID_AUTO,
"lro_bad_csum", CTLFLAG_RD, &ss->lc.lro_bad_csum, "lro_bad_csum", CTLFLAG_RD, &ss->lc.lro_bad_csum,
0, "number of bad csums preventing LRO"); 0, "number of bad csums preventing LRO");
SYSCTL_ADD_INT(ctx, children, OID_AUTO, SYSCTL_ADD_U64(ctx, children, OID_AUTO,
"lro_queued", CTLFLAG_RD, &ss->lc.lro_queued, "lro_queued", CTLFLAG_RD, &ss->lc.lro_queued,
0, "number of frames appended to lro merge" 0, "number of frames appended to lro merge"
"queues"); "queues");

View File

@ -2,6 +2,7 @@
* Copyright (c) 2007, Myricom Inc. * Copyright (c) 2007, Myricom Inc.
* Copyright (c) 2008, Intel Corporation. * Copyright (c) 2008, Intel Corporation.
* Copyright (c) 2012 The FreeBSD Foundation * Copyright (c) 2012 The FreeBSD Foundation
* Copyright (c) 2016 Mellanox Technologies.
* All rights reserved. * All rights reserved.
* *
* Portions of this software were developed by Bjoern Zeeb * Portions of this software were developed by Bjoern Zeeb
@ -58,9 +59,7 @@ __FBSDID("$FreeBSD$");
#include <machine/in_cksum.h> #include <machine/in_cksum.h>
#ifndef LRO_ENTRIES static MALLOC_DEFINE(M_LRO, "LRO", "LRO control structures");
#define LRO_ENTRIES 8 /* # of LRO entries per RX queue. */
#endif
#define TCP_LRO_UPDATE_CSUM 1 #define TCP_LRO_UPDATE_CSUM 1
#ifndef TCP_LRO_UPDATE_CSUM #ifndef TCP_LRO_UPDATE_CSUM
@ -69,43 +68,74 @@ __FBSDID("$FreeBSD$");
int int
tcp_lro_init(struct lro_ctrl *lc) tcp_lro_init(struct lro_ctrl *lc)
{
return (tcp_lro_init_args(lc, NULL, TCP_LRO_ENTRIES, 0));
}
int
tcp_lro_init_args(struct lro_ctrl *lc, struct ifnet *ifp,
unsigned lro_entries, unsigned lro_mbufs)
{ {
struct lro_entry *le; struct lro_entry *le;
int error, i; size_t size;
unsigned i;
lc->lro_bad_csum = 0; lc->lro_bad_csum = 0;
lc->lro_queued = 0; lc->lro_queued = 0;
lc->lro_flushed = 0; lc->lro_flushed = 0;
lc->lro_cnt = 0; lc->lro_cnt = 0;
lc->lro_mbuf_count = 0;
lc->lro_mbuf_max = lro_mbufs;
lc->lro_cnt = lro_entries;
lc->ifp = ifp;
SLIST_INIT(&lc->lro_free); SLIST_INIT(&lc->lro_free);
SLIST_INIT(&lc->lro_active); SLIST_INIT(&lc->lro_active);
error = 0; /* compute size to allocate */
for (i = 0; i < LRO_ENTRIES; i++) { size = (lro_mbufs * sizeof(struct mbuf *)) +
le = (struct lro_entry *)malloc(sizeof(*le), M_DEVBUF, (lro_entries * sizeof(*le));
M_NOWAIT | M_ZERO); lc->lro_mbuf_data = (struct mbuf **)
if (le == NULL) { malloc(size, M_LRO, M_NOWAIT | M_ZERO);
if (i == 0)
error = ENOMEM;
break;
}
lc->lro_cnt = i + 1;
SLIST_INSERT_HEAD(&lc->lro_free, le, next);
}
return (error); /* check for out of memory */
if (lc->lro_mbuf_data == NULL) {
memset(lc, 0, sizeof(*lc));
return (ENOMEM);
}
/* compute offset for LRO entries */
le = (struct lro_entry *)
(lc->lro_mbuf_data + lro_mbufs);
/* setup linked list */
for (i = 0; i != lro_entries; i++)
SLIST_INSERT_HEAD(&lc->lro_free, le + i, next);
return (0);
} }
void void
tcp_lro_free(struct lro_ctrl *lc) tcp_lro_free(struct lro_ctrl *lc)
{ {
struct lro_entry *le; struct lro_entry *le;
unsigned x;
while (!SLIST_EMPTY(&lc->lro_free)) { /* reset LRO free list */
le = SLIST_FIRST(&lc->lro_free); SLIST_INIT(&lc->lro_free);
SLIST_REMOVE_HEAD(&lc->lro_free, next);
free(le, M_DEVBUF); /* free active mbufs, if any */
while ((le = SLIST_FIRST(&lc->lro_active)) != NULL) {
SLIST_REMOVE_HEAD(&lc->lro_active, next);
m_freem(le->m_head);
} }
/* free mbuf array, if any */
for (x = 0; x != lc->lro_mbuf_count; x++)
m_freem(lc->lro_mbuf_data[x]);
lc->lro_mbuf_count = 0;
/* free allocated memory, if any */
free(lc->lro_mbuf_data, M_LRO);
lc->lro_mbuf_data = NULL;
} }
#ifdef TCP_LRO_UPDATE_CSUM #ifdef TCP_LRO_UPDATE_CSUM
@ -305,6 +335,83 @@ tcp_lro_flush(struct lro_ctrl *lc, struct lro_entry *le)
SLIST_INSERT_HEAD(&lc->lro_free, le, next); SLIST_INSERT_HEAD(&lc->lro_free, le, next);
} }
static int
tcp_lro_mbuf_compare_header(const void *ppa, const void *ppb)
{
const struct mbuf *ma = *((const struct mbuf * const *)ppa);
const struct mbuf *mb = *((const struct mbuf * const *)ppb);
int ret;
ret = M_HASHTYPE_GET(ma) - M_HASHTYPE_GET(mb);
if (ret != 0)
goto done;
ret = ma->m_pkthdr.flowid - mb->m_pkthdr.flowid;
if (ret != 0)
goto done;
ret = TCP_LRO_SEQUENCE(ma) - TCP_LRO_SEQUENCE(mb);
done:
return (ret);
}
void
tcp_lro_flush_all(struct lro_ctrl *lc)
{
struct lro_entry *le;
uint32_t hashtype;
uint32_t flowid;
unsigned x;
/* check if no mbufs to flush */
if (__predict_false(lc->lro_mbuf_count == 0))
goto done;
/* sort all mbufs according to stream */
qsort(lc->lro_mbuf_data, lc->lro_mbuf_count, sizeof(struct mbuf *),
&tcp_lro_mbuf_compare_header);
/* input data into LRO engine, stream by stream */
flowid = 0;
hashtype = M_HASHTYPE_NONE;
for (x = 0; x != lc->lro_mbuf_count; x++) {
struct mbuf *mb;
mb = lc->lro_mbuf_data[x];
/* check for new stream */
if (mb->m_pkthdr.flowid != flowid ||
M_HASHTYPE_GET(mb) != hashtype) {
flowid = mb->m_pkthdr.flowid;
hashtype = M_HASHTYPE_GET(mb);
/* flush active streams */
while ((le = SLIST_FIRST(&lc->lro_active)) != NULL) {
SLIST_REMOVE_HEAD(&lc->lro_active, next);
tcp_lro_flush(lc, le);
}
}
#ifdef TCP_LRO_RESET_SEQUENCE
/* reset sequence number */
TCP_LRO_SEQUENCE(mb) = 0;
#endif
/* add packet to LRO engine */
if (tcp_lro_rx(lc, mb, 0) != 0) {
/* input packet to network layer */
(*lc->ifp->if_input)(lc->ifp, mb);
lc->lro_queued++;
lc->lro_flushed++;
}
}
done:
/* flush active streams */
while ((le = SLIST_FIRST(&lc->lro_active)) != NULL) {
SLIST_REMOVE_HEAD(&lc->lro_active, next);
tcp_lro_flush(lc, le);
}
lc->lro_mbuf_count = 0;
}
#ifdef INET6 #ifdef INET6
static int static int
tcp_lro_rx_ipv6(struct lro_ctrl *lc, struct mbuf *m, struct ip6_hdr *ip6, tcp_lro_rx_ipv6(struct lro_ctrl *lc, struct mbuf *m, struct ip6_hdr *ip6,
@ -633,4 +740,37 @@ tcp_lro_rx(struct lro_ctrl *lc, struct mbuf *m, uint32_t csum)
return (0); return (0);
} }
void
tcp_lro_queue_mbuf(struct lro_ctrl *lc, struct mbuf *mb)
{
/* sanity checks */
if (__predict_false(lc->ifp == NULL || lc->lro_mbuf_data == NULL ||
lc->lro_mbuf_max == 0)) {
/* packet drop */
m_freem(mb);
return;
}
/* check if packet is not LRO capable */
if (__predict_false(mb->m_pkthdr.csum_flags == 0 ||
(lc->ifp->if_capenable & IFCAP_LRO) == 0)) {
lc->lro_flushed++;
lc->lro_queued++;
/* input packet to network layer */
(*lc->ifp->if_input) (lc->ifp, mb);
return;
}
/* check if array is full */
if (__predict_false(lc->lro_mbuf_count == lc->lro_mbuf_max))
tcp_lro_flush_all(lc);
/* store sequence number */
TCP_LRO_SEQUENCE(mb) = lc->lro_mbuf_count;
/* enter mbuf */
lc->lro_mbuf_data[lc->lro_mbuf_count++] = mb;
}
/* end */ /* end */

View File

@ -1,6 +1,7 @@
/*- /*-
* Copyright (c) 2006, Myricom Inc. * Copyright (c) 2006, Myricom Inc.
* Copyright (c) 2008, Intel Corporation. * Copyright (c) 2008, Intel Corporation.
* Copyright (c) 2016 Mellanox Technologies.
* All rights reserved. * All rights reserved.
* *
* Redistribution and use in source and binary forms, with or without * Redistribution and use in source and binary forms, with or without
@ -32,6 +33,14 @@
#include <sys/time.h> #include <sys/time.h>
#ifndef TCP_LRO_ENTRIES
/* Define default number of LRO entries per RX queue */
#define TCP_LRO_ENTRIES 8
#endif
#define TCP_LRO_SEQUENCE(mb) \
(mb)->m_pkthdr.PH_loc.thirtytwo[0]
struct lro_entry struct lro_entry
{ {
SLIST_ENTRY(lro_entry) next; SLIST_ENTRY(lro_entry) next;
@ -75,20 +84,26 @@ SLIST_HEAD(lro_head, lro_entry);
/* NB: This is part of driver structs. */ /* NB: This is part of driver structs. */
struct lro_ctrl { struct lro_ctrl {
struct ifnet *ifp; struct ifnet *ifp;
int lro_queued; struct mbuf **lro_mbuf_data;
int lro_flushed; uint64_t lro_queued;
int lro_bad_csum; uint64_t lro_flushed;
int lro_cnt; uint64_t lro_bad_csum;
unsigned lro_cnt;
unsigned lro_mbuf_count;
unsigned lro_mbuf_max;
struct lro_head lro_active; struct lro_head lro_active;
struct lro_head lro_free; struct lro_head lro_free;
}; };
int tcp_lro_init(struct lro_ctrl *); int tcp_lro_init(struct lro_ctrl *);
int tcp_lro_init_args(struct lro_ctrl *, struct ifnet *, unsigned, unsigned);
void tcp_lro_free(struct lro_ctrl *); void tcp_lro_free(struct lro_ctrl *);
void tcp_lro_flush_inactive(struct lro_ctrl *, const struct timeval *); void tcp_lro_flush_inactive(struct lro_ctrl *, const struct timeval *);
void tcp_lro_flush(struct lro_ctrl *, struct lro_entry *); void tcp_lro_flush(struct lro_ctrl *, struct lro_entry *);
void tcp_lro_flush_all(struct lro_ctrl *);
int tcp_lro_rx(struct lro_ctrl *, struct mbuf *, uint32_t); int tcp_lro_rx(struct lro_ctrl *, struct mbuf *, uint32_t);
void tcp_lro_queue_mbuf(struct lro_ctrl *, struct mbuf *);
#define TCP_LRO_CANNOT -1 #define TCP_LRO_CANNOT -1
#define TCP_LRO_NOT_SUPPORTED 1 #define TCP_LRO_NOT_SUPPORTED 1

View File

@ -58,7 +58,7 @@
* in the range 5 to 9. * in the range 5 to 9.
*/ */
#undef __FreeBSD_version #undef __FreeBSD_version
#define __FreeBSD_version 1100094 /* Master, propagated to newvers */ #define __FreeBSD_version 1100095 /* Master, propagated to newvers */
/* /*
* __FreeBSD_kernel__ indicates that this system uses the kernel of FreeBSD, * __FreeBSD_kernel__ indicates that this system uses the kernel of FreeBSD,