c0d002aed9
rte_gso_segment decreased refcnt of pkt by one, but it is wrong if pkt is external mbuf, pkt won't be freed because of incorrect refcnt, the result is application can't allocate mbuf from mempool because mbufs in mempool are run out of. One correct way is application should call rte_pktmbuf_free after calling rte_gso_segment to free pkt explicitly. rte_gso_segment must not handle it, this should be responsibility of application. This commit changed rte_gso_segment in functional behavior and return value, so the application must take appropriate actions according to return values, "ret < 0" means it should free and drop 'pkt', "ret == 0" means 'pkt' isn't GSOed but 'pkt' can be transmitted as a normal packet, "ret > 0" means 'pkt' has been GSOed into two or multiple segments, it should use "pkts_out" to transmit these segments. The application must free 'pkt' after call rte_gso_segment when return value isn't equal to 0. Fixes: 119583797b6a ("gso: support TCP/IPv4 GSO") Cc: stable@dpdk.org Signed-off-by: Yi Yang <yangyi01@inspur.com> Acked-by: Jiayu Hu <jiayu.hu@intel.com> Acked-by: Konstantin Ananyev <konstantin.ananyev@intel.com>
72 lines
1.9 KiB
C
72 lines
1.9 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright(c) 2017 Intel Corporation
|
|
*/
|
|
|
|
#include "gso_common.h"
|
|
#include "gso_tcp4.h"
|
|
|
|
static void
|
|
update_ipv4_tcp_headers(struct rte_mbuf *pkt, uint8_t ipid_delta,
|
|
struct rte_mbuf **segs, uint16_t nb_segs)
|
|
{
|
|
struct rte_ipv4_hdr *ipv4_hdr;
|
|
struct rte_tcp_hdr *tcp_hdr;
|
|
uint32_t sent_seq;
|
|
uint16_t id, tail_idx, i;
|
|
uint16_t l3_offset = pkt->l2_len;
|
|
uint16_t l4_offset = l3_offset + pkt->l3_len;
|
|
|
|
ipv4_hdr = (struct rte_ipv4_hdr *)(rte_pktmbuf_mtod(pkt, char*) +
|
|
l3_offset);
|
|
tcp_hdr = (struct rte_tcp_hdr *)((char *)ipv4_hdr + pkt->l3_len);
|
|
id = rte_be_to_cpu_16(ipv4_hdr->packet_id);
|
|
sent_seq = rte_be_to_cpu_32(tcp_hdr->sent_seq);
|
|
tail_idx = nb_segs - 1;
|
|
|
|
for (i = 0; i < nb_segs; i++) {
|
|
update_ipv4_header(segs[i], l3_offset, id);
|
|
update_tcp_header(segs[i], l4_offset, sent_seq, i < tail_idx);
|
|
id += ipid_delta;
|
|
sent_seq += (segs[i]->pkt_len - segs[i]->data_len);
|
|
}
|
|
}
|
|
|
|
int
|
|
gso_tcp4_segment(struct rte_mbuf *pkt,
|
|
uint16_t gso_size,
|
|
uint8_t ipid_delta,
|
|
struct rte_mempool *direct_pool,
|
|
struct rte_mempool *indirect_pool,
|
|
struct rte_mbuf **pkts_out,
|
|
uint16_t nb_pkts_out)
|
|
{
|
|
struct rte_ipv4_hdr *ipv4_hdr;
|
|
uint16_t pyld_unit_size, hdr_offset;
|
|
uint16_t frag_off;
|
|
int ret;
|
|
|
|
/* Don't process the fragmented packet */
|
|
ipv4_hdr = (struct rte_ipv4_hdr *)(rte_pktmbuf_mtod(pkt, char *) +
|
|
pkt->l2_len);
|
|
frag_off = rte_be_to_cpu_16(ipv4_hdr->fragment_offset);
|
|
if (unlikely(IS_FRAGMENTED(frag_off))) {
|
|
return 0;
|
|
}
|
|
|
|
/* Don't process the packet without data */
|
|
hdr_offset = pkt->l2_len + pkt->l3_len + pkt->l4_len;
|
|
if (unlikely(hdr_offset >= pkt->pkt_len)) {
|
|
return 0;
|
|
}
|
|
|
|
pyld_unit_size = gso_size - hdr_offset;
|
|
|
|
/* Segment the payload */
|
|
ret = gso_do_segment(pkt, hdr_offset, pyld_unit_size, direct_pool,
|
|
indirect_pool, pkts_out, nb_pkts_out);
|
|
if (ret > 1)
|
|
update_ipv4_tcp_headers(pkt, ipid_delta, pkts_out, ret);
|
|
|
|
return ret;
|
|
}
|