d08b8680e1
This updates the driver to align with the version included in the "Intel Ethernet Adapter Complete Driver Pack", version 25.6. There are no major functional changes; this mostly contains bug fixes and changes to prepare for new features. This version of the driver uses the previously committed ice_ddp package 1.3.19.0. Signed-off-by: Eric Joyner <erj@FreeBSD.org> Tested by: jeffrey.e.pieper@intel.com Sponsored by: Intel Corporation Differential Revision: https://reviews.freebsd.org/D28640
425 lines
12 KiB
C
425 lines
12 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause */
|
|
/* Copyright (c) 2021, Intel Corporation
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* 1. Redistributions of source code must retain the above copyright notice,
|
|
* this list of conditions and the following disclaimer.
|
|
*
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* 3. Neither the name of the Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived from
|
|
* this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
/*$FreeBSD$*/
|
|
|
|
/**
|
|
* @file ice_common_txrx.h
|
|
* @brief common Tx/Rx utility functions
|
|
*
|
|
* Contains common utility functions for the Tx/Rx hot path.
|
|
*
|
|
* The functions do depend on the if_pkt_info_t structure. A suitable
|
|
* implementation of this structure must be provided if these functions are to
|
|
* be used without the iflib networking stack.
|
|
*/
|
|
|
|
#ifndef _ICE_COMMON_TXRX_H_
|
|
#define _ICE_COMMON_TXRX_H_
|
|
|
|
#include <netinet/udp.h>
|
|
#include <netinet/sctp.h>
|
|
|
|
/**
|
|
* ice_tso_detect_sparse - detect TSO packets with too many segments
|
|
* @pi: packet information
|
|
*
|
|
* Hardware only transmits packets with a maximum of 8 descriptors. For TSO
|
|
* packets, hardware needs to be able to build the split packets using 8 or
|
|
* fewer descriptors. Additionally, the header must be contained within at
|
|
* most 3 descriptors.
|
|
*
|
|
* To verify this, we walk the headers to find out how many descriptors the
|
|
* headers require (usually 1). Then we ensure that, for each TSO segment, its
|
|
* data plus the headers are contained within 8 or fewer descriptors.
|
|
*/
|
|
static inline int
|
|
ice_tso_detect_sparse(if_pkt_info_t pi)
|
|
{
|
|
int count, curseg, i, hlen, segsz, seglen, tsolen, hdrs, maxsegs;
|
|
bus_dma_segment_t *segs = pi->ipi_segs;
|
|
int nsegs = pi->ipi_nsegs;
|
|
|
|
curseg = hdrs = 0;
|
|
|
|
hlen = pi->ipi_ehdrlen + pi->ipi_ip_hlen + pi->ipi_tcp_hlen;
|
|
tsolen = pi->ipi_len - hlen;
|
|
|
|
/* First, count the number of descriptors for the header.
|
|
* Additionally, make sure it does not span more than 3 segments.
|
|
*/
|
|
i = 0;
|
|
curseg = segs[0].ds_len;
|
|
while (hlen > 0) {
|
|
hdrs++;
|
|
if (hdrs > ICE_MAX_TSO_HDR_SEGS)
|
|
return (1);
|
|
if (curseg == 0) {
|
|
i++;
|
|
if (__predict_false(i == nsegs))
|
|
return (1);
|
|
|
|
curseg = segs[i].ds_len;
|
|
}
|
|
seglen = min(curseg, hlen);
|
|
curseg -= seglen;
|
|
hlen -= seglen;
|
|
}
|
|
|
|
maxsegs = ICE_MAX_TX_SEGS - hdrs;
|
|
|
|
/* We must count the headers, in order to verify that they take up
|
|
* 3 or fewer descriptors. However, we don't need to check the data
|
|
* if the total segments is small.
|
|
*/
|
|
if (nsegs <= maxsegs)
|
|
return (0);
|
|
|
|
count = 0;
|
|
|
|
/* Now check the data to make sure that each TSO segment is made up of
|
|
* no more than maxsegs descriptors. This ensures that hardware will
|
|
* be capable of performing TSO offload.
|
|
*/
|
|
while (tsolen > 0) {
|
|
segsz = pi->ipi_tso_segsz;
|
|
while (segsz > 0 && tsolen != 0) {
|
|
count++;
|
|
if (count > maxsegs) {
|
|
return (1);
|
|
}
|
|
if (curseg == 0) {
|
|
i++;
|
|
if (__predict_false(i == nsegs)) {
|
|
return (1);
|
|
}
|
|
curseg = segs[i].ds_len;
|
|
}
|
|
seglen = min(curseg, segsz);
|
|
segsz -= seglen;
|
|
curseg -= seglen;
|
|
tsolen -= seglen;
|
|
}
|
|
count = 0;
|
|
}
|
|
|
|
return (0);
|
|
}
|
|
|
|
/**
|
|
* ice_tso_setup - Setup a context descriptor to prepare for a TSO packet
|
|
* @txq: the Tx queue to use
|
|
* @pi: the packet info to prepare for
|
|
*
|
|
* Setup a context descriptor in preparation for sending a Tx packet that
|
|
* requires the TSO offload. Returns the index of the descriptor to use when
|
|
* encapsulating the Tx packet data into descriptors.
|
|
*/
|
|
static inline int
|
|
ice_tso_setup(struct ice_tx_queue *txq, if_pkt_info_t pi)
|
|
{
|
|
struct ice_tx_ctx_desc *txd;
|
|
u32 cmd, mss, type, tsolen;
|
|
int idx;
|
|
u64 type_cmd_tso_mss;
|
|
|
|
idx = pi->ipi_pidx;
|
|
txd = (struct ice_tx_ctx_desc *)&txq->tx_base[idx];
|
|
tsolen = pi->ipi_len - (pi->ipi_ehdrlen + pi->ipi_ip_hlen + pi->ipi_tcp_hlen);
|
|
|
|
type = ICE_TX_DESC_DTYPE_CTX;
|
|
cmd = ICE_TX_CTX_DESC_TSO;
|
|
/* TSO MSS must not be less than 64 */
|
|
if (pi->ipi_tso_segsz < ICE_MIN_TSO_MSS) {
|
|
txq->stats.mss_too_small++;
|
|
pi->ipi_tso_segsz = ICE_MIN_TSO_MSS;
|
|
}
|
|
mss = pi->ipi_tso_segsz;
|
|
|
|
type_cmd_tso_mss = ((u64)type << ICE_TXD_CTX_QW1_DTYPE_S) |
|
|
((u64)cmd << ICE_TXD_CTX_QW1_CMD_S) |
|
|
((u64)tsolen << ICE_TXD_CTX_QW1_TSO_LEN_S) |
|
|
((u64)mss << ICE_TXD_CTX_QW1_MSS_S);
|
|
txd->qw1 = htole64(type_cmd_tso_mss);
|
|
|
|
txd->tunneling_params = htole32(0);
|
|
txq->tso++;
|
|
|
|
return ((idx + 1) & (txq->desc_count-1));
|
|
}
|
|
|
|
/**
|
|
* ice_tx_setup_offload - Setup register values for performing a Tx offload
|
|
* @txq: The Tx queue, used to track checksum offload stats
|
|
* @pi: the packet info to program for
|
|
* @cmd: the cmd register value to update
|
|
* @off: the off register value to update
|
|
*
|
|
* Based on the packet info provided, update the cmd and off values for
|
|
* enabling Tx offloads. This depends on the packet type and which offloads
|
|
* have been requested.
|
|
*
|
|
* We also track the total number of times that we've requested hardware
|
|
* offload a particular type of checksum for debugging purposes.
|
|
*/
|
|
static inline void
|
|
ice_tx_setup_offload(struct ice_tx_queue *txq, if_pkt_info_t pi, u32 *cmd, u32 *off)
|
|
{
|
|
u32 remaining_csum_flags = pi->ipi_csum_flags;
|
|
|
|
switch (pi->ipi_etype) {
|
|
#ifdef INET
|
|
case ETHERTYPE_IP:
|
|
if (pi->ipi_csum_flags & ICE_CSUM_IP) {
|
|
*cmd |= ICE_TX_DESC_CMD_IIPT_IPV4_CSUM;
|
|
txq->stats.cso[ICE_CSO_STAT_TX_IP4]++;
|
|
remaining_csum_flags &= ~CSUM_IP;
|
|
} else
|
|
*cmd |= ICE_TX_DESC_CMD_IIPT_IPV4;
|
|
break;
|
|
#endif
|
|
#ifdef INET6
|
|
case ETHERTYPE_IPV6:
|
|
*cmd |= ICE_TX_DESC_CMD_IIPT_IPV6;
|
|
/*
|
|
* This indicates that the IIPT flag was set to the IPV6 value;
|
|
* there's no checksum for IPv6 packets.
|
|
*/
|
|
txq->stats.cso[ICE_CSO_STAT_TX_IP6]++;
|
|
break;
|
|
#endif
|
|
default:
|
|
txq->stats.cso[ICE_CSO_STAT_TX_L3_ERR]++;
|
|
break;
|
|
}
|
|
|
|
*off |= (pi->ipi_ehdrlen >> 1) << ICE_TX_DESC_LEN_MACLEN_S;
|
|
*off |= (pi->ipi_ip_hlen >> 2) << ICE_TX_DESC_LEN_IPLEN_S;
|
|
|
|
if (!(remaining_csum_flags & ~ICE_RX_CSUM_FLAGS))
|
|
return;
|
|
|
|
switch (pi->ipi_ipproto) {
|
|
case IPPROTO_TCP:
|
|
if (pi->ipi_csum_flags & ICE_CSUM_TCP) {
|
|
*cmd |= ICE_TX_DESC_CMD_L4T_EOFT_TCP;
|
|
*off |= (pi->ipi_tcp_hlen >> 2) <<
|
|
ICE_TX_DESC_LEN_L4_LEN_S;
|
|
txq->stats.cso[ICE_CSO_STAT_TX_TCP]++;
|
|
}
|
|
break;
|
|
case IPPROTO_UDP:
|
|
if (pi->ipi_csum_flags & ICE_CSUM_UDP) {
|
|
*cmd |= ICE_TX_DESC_CMD_L4T_EOFT_UDP;
|
|
*off |= (sizeof(struct udphdr) >> 2) <<
|
|
ICE_TX_DESC_LEN_L4_LEN_S;
|
|
txq->stats.cso[ICE_CSO_STAT_TX_UDP]++;
|
|
}
|
|
break;
|
|
case IPPROTO_SCTP:
|
|
if (pi->ipi_csum_flags & ICE_CSUM_SCTP) {
|
|
*cmd |= ICE_TX_DESC_CMD_L4T_EOFT_SCTP;
|
|
*off |= (sizeof(struct sctphdr) >> 2) <<
|
|
ICE_TX_DESC_LEN_L4_LEN_S;
|
|
txq->stats.cso[ICE_CSO_STAT_TX_SCTP]++;
|
|
}
|
|
break;
|
|
default:
|
|
txq->stats.cso[ICE_CSO_STAT_TX_L4_ERR]++;
|
|
break;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ice_rx_checksum - verify hardware checksum is valid or not
|
|
* @rxq: the Rx queue structure
|
|
* @flags: checksum flags to update
|
|
* @data: checksum data to update
|
|
* @status0: descriptor status data
|
|
* @ptype: packet type
|
|
*
|
|
* Determine whether the hardware indicated that the Rx checksum is valid. If
|
|
* so, update the checksum flags and data, informing the stack of the status
|
|
* of the checksum so that it does not spend time verifying it manually.
|
|
*/
|
|
static void
|
|
ice_rx_checksum(struct ice_rx_queue *rxq, uint32_t *flags, uint32_t *data,
|
|
u16 status0, u16 ptype)
|
|
{
|
|
const u16 l3_error = (BIT(ICE_RX_FLEX_DESC_STATUS0_XSUM_IPE_S) |
|
|
BIT(ICE_RX_FLEX_DESC_STATUS0_XSUM_EIPE_S));
|
|
const u16 l4_error = (BIT(ICE_RX_FLEX_DESC_STATUS0_XSUM_L4E_S) |
|
|
BIT(ICE_RX_FLEX_DESC_STATUS0_XSUM_EUDPE_S));
|
|
const u16 xsum_errors = (l3_error | l4_error |
|
|
BIT(ICE_RX_FLEX_DESC_STATUS0_IPV6EXADD_S));
|
|
struct ice_rx_ptype_decoded decoded;
|
|
bool is_ipv4, is_ipv6;
|
|
|
|
/* No L3 or L4 checksum was calculated */
|
|
if (!(status0 & BIT(ICE_RX_FLEX_DESC_STATUS0_L3L4P_S))) {
|
|
return;
|
|
}
|
|
|
|
decoded = ice_decode_rx_desc_ptype(ptype);
|
|
*flags = 0;
|
|
|
|
if (!(decoded.known && decoded.outer_ip))
|
|
return;
|
|
|
|
is_ipv4 = (decoded.outer_ip == ICE_RX_PTYPE_OUTER_IP) &&
|
|
(decoded.outer_ip_ver == ICE_RX_PTYPE_OUTER_IPV4);
|
|
is_ipv6 = (decoded.outer_ip == ICE_RX_PTYPE_OUTER_IP) &&
|
|
(decoded.outer_ip_ver == ICE_RX_PTYPE_OUTER_IPV6);
|
|
|
|
/* No checksum errors were reported */
|
|
if (!(status0 & xsum_errors)) {
|
|
if (is_ipv4)
|
|
*flags |= CSUM_L3_CALC | CSUM_L3_VALID;
|
|
|
|
switch (decoded.inner_prot) {
|
|
case ICE_RX_PTYPE_INNER_PROT_TCP:
|
|
case ICE_RX_PTYPE_INNER_PROT_UDP:
|
|
case ICE_RX_PTYPE_INNER_PROT_SCTP:
|
|
*flags |= CSUM_L4_CALC | CSUM_L4_VALID;
|
|
*data |= htons(0xffff);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Certain IPv6 extension headers impact the validity of L4 checksums.
|
|
* If one of these headers exist, hardware will set the IPV6EXADD bit
|
|
* in the descriptor. If the bit is set then pretend like hardware
|
|
* didn't checksum this packet.
|
|
*/
|
|
if (is_ipv6 && (status0 & BIT(ICE_RX_FLEX_DESC_STATUS0_IPV6EXADD_S))) {
|
|
rxq->stats.cso[ICE_CSO_STAT_RX_IP6_ERR]++;
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* At this point, status0 must have at least one of the l3_error or
|
|
* l4_error bits set.
|
|
*/
|
|
|
|
if (status0 & l3_error) {
|
|
if (is_ipv4) {
|
|
rxq->stats.cso[ICE_CSO_STAT_RX_IP4_ERR]++;
|
|
*flags |= CSUM_L3_CALC;
|
|
} else {
|
|
/* Hardware indicated L3 error but this isn't IPv4? */
|
|
rxq->stats.cso[ICE_CSO_STAT_RX_L3_ERR]++;
|
|
}
|
|
/* don't bother reporting L4 errors if we got an L3 error */
|
|
return;
|
|
} else if (is_ipv4) {
|
|
*flags |= CSUM_L3_CALC | CSUM_L3_VALID;
|
|
}
|
|
|
|
if (status0 & l4_error) {
|
|
switch (decoded.inner_prot) {
|
|
case ICE_RX_PTYPE_INNER_PROT_TCP:
|
|
rxq->stats.cso[ICE_CSO_STAT_RX_TCP_ERR]++;
|
|
*flags |= CSUM_L4_CALC;
|
|
break;
|
|
case ICE_RX_PTYPE_INNER_PROT_UDP:
|
|
rxq->stats.cso[ICE_CSO_STAT_RX_UDP_ERR]++;
|
|
*flags |= CSUM_L4_CALC;
|
|
break;
|
|
case ICE_RX_PTYPE_INNER_PROT_SCTP:
|
|
rxq->stats.cso[ICE_CSO_STAT_RX_SCTP_ERR]++;
|
|
*flags |= CSUM_L4_CALC;
|
|
break;
|
|
default:
|
|
/*
|
|
* Hardware indicated L4 error, but this isn't one of
|
|
* the expected protocols.
|
|
*/
|
|
rxq->stats.cso[ICE_CSO_STAT_RX_L4_ERR]++;
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ice_ptype_to_hash - Convert packet type to a hash value
|
|
* @ptype: the packet type to convert
|
|
*
|
|
* Given the packet type, convert to a suitable hashtype to report to the
|
|
* upper stack via the iri_rsstype value of the if_rxd_info_t structure.
|
|
*
|
|
* If the hash type is unknown we'll report M_HASHTYPE_OPAQUE.
|
|
*/
|
|
static inline int
|
|
ice_ptype_to_hash(u16 ptype)
|
|
{
|
|
struct ice_rx_ptype_decoded decoded;
|
|
|
|
if (ptype >= ARRAY_SIZE(ice_ptype_lkup))
|
|
return M_HASHTYPE_OPAQUE;
|
|
|
|
decoded = ice_decode_rx_desc_ptype(ptype);
|
|
|
|
if (!decoded.known)
|
|
return M_HASHTYPE_OPAQUE;
|
|
|
|
if (decoded.outer_ip == ICE_RX_PTYPE_OUTER_L2)
|
|
return M_HASHTYPE_OPAQUE;
|
|
|
|
/* Note: anything that gets to this point is IP */
|
|
if (decoded.outer_ip_ver == ICE_RX_PTYPE_OUTER_IPV6) {
|
|
switch (decoded.inner_prot) {
|
|
case ICE_RX_PTYPE_INNER_PROT_TCP:
|
|
return M_HASHTYPE_RSS_TCP_IPV6;
|
|
case ICE_RX_PTYPE_INNER_PROT_UDP:
|
|
return M_HASHTYPE_RSS_UDP_IPV6;
|
|
default:
|
|
return M_HASHTYPE_RSS_IPV6;
|
|
}
|
|
}
|
|
if (decoded.outer_ip_ver == ICE_RX_PTYPE_OUTER_IPV4) {
|
|
switch (decoded.inner_prot) {
|
|
case ICE_RX_PTYPE_INNER_PROT_TCP:
|
|
return M_HASHTYPE_RSS_TCP_IPV4;
|
|
case ICE_RX_PTYPE_INNER_PROT_UDP:
|
|
return M_HASHTYPE_RSS_UDP_IPV4;
|
|
default:
|
|
return M_HASHTYPE_RSS_IPV4;
|
|
}
|
|
}
|
|
|
|
/* We should never get here!! */
|
|
return M_HASHTYPE_OPAQUE;
|
|
}
|
|
#endif
|