3a52e64742
Include rte_memory.h for lib files that use __rte_cache_aligned attribute. Consider the following code: struct per_core_foo { ... } __rte_cache_aligned; struct global_foo { struct per_core_foo foo[RTE_MAX_CORE]; }; If __rte_cache_aligned is not defined (rte_memory.h is not included), the code compiles but the structure is not aligned... it defines the structure and creates a global variable called __rte_cache_aligned. And this can lead to really bad things if this code is in a .h that is included by files that may or may not include rte_memory.h Signed-off-by: Jia Yu <jyu@vmware.com> Acked-by: Thomas Monjalon <thomas.monjalon@6wind.com>
242 lines
6.5 KiB
C
242 lines
6.5 KiB
C
/*-
|
|
* BSD LICENSE
|
|
*
|
|
* Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
#include <string.h>
|
|
|
|
#include <rte_ether.h>
|
|
#include <rte_ip_frag.h>
|
|
#include <rte_memory.h>
|
|
|
|
#include "rte_port_frag.h"
|
|
|
|
/* Default byte size for the IPv4 Maximum Transfer Unit (MTU).
|
|
* This value includes the size of IPv4 header. */
|
|
#define IPV4_MTU_DEFAULT ETHER_MTU
|
|
|
|
/* Max number of fragments per packet allowed */
|
|
#define IPV4_MAX_FRAGS_PER_PACKET 0x80
|
|
|
|
struct rte_port_ring_reader_ipv4_frag {
|
|
/* Input parameters */
|
|
struct rte_ring *ring;
|
|
uint32_t mtu;
|
|
uint32_t metadata_size;
|
|
struct rte_mempool *pool_direct;
|
|
struct rte_mempool *pool_indirect;
|
|
|
|
/* Internal buffers */
|
|
struct rte_mbuf *pkts[RTE_PORT_IN_BURST_SIZE_MAX];
|
|
struct rte_mbuf *frags[IPV4_MAX_FRAGS_PER_PACKET];
|
|
uint32_t n_pkts;
|
|
uint32_t pos_pkts;
|
|
uint32_t n_frags;
|
|
uint32_t pos_frags;
|
|
} __rte_cache_aligned;
|
|
|
|
static void *
|
|
rte_port_ring_reader_ipv4_frag_create(void *params, int socket_id)
|
|
{
|
|
struct rte_port_ring_reader_ipv4_frag_params *conf =
|
|
(struct rte_port_ring_reader_ipv4_frag_params *) params;
|
|
struct rte_port_ring_reader_ipv4_frag *port;
|
|
|
|
/* Check input parameters */
|
|
if (conf == NULL) {
|
|
RTE_LOG(ERR, PORT, "%s: Parameter conf is NULL\n", __func__);
|
|
return NULL;
|
|
}
|
|
if (conf->ring == NULL) {
|
|
RTE_LOG(ERR, PORT, "%s: Parameter ring is NULL\n", __func__);
|
|
return NULL;
|
|
}
|
|
if (conf->mtu == 0) {
|
|
RTE_LOG(ERR, PORT, "%s: Parameter mtu is invalid\n", __func__);
|
|
return NULL;
|
|
}
|
|
if (conf->pool_direct == NULL) {
|
|
RTE_LOG(ERR, PORT, "%s: Parameter pool_direct is NULL\n",
|
|
__func__);
|
|
return NULL;
|
|
}
|
|
if (conf->pool_indirect == NULL) {
|
|
RTE_LOG(ERR, PORT, "%s: Parameter pool_indirect is NULL\n",
|
|
__func__);
|
|
return NULL;
|
|
}
|
|
|
|
/* Memory allocation */
|
|
port = rte_zmalloc_socket("PORT", sizeof(*port), RTE_CACHE_LINE_SIZE,
|
|
socket_id);
|
|
if (port == NULL) {
|
|
RTE_LOG(ERR, PORT, "%s: port is NULL\n", __func__);
|
|
return NULL;
|
|
}
|
|
|
|
/* Initialization */
|
|
port->ring = conf->ring;
|
|
port->mtu = conf->mtu;
|
|
port->metadata_size = conf->metadata_size;
|
|
port->pool_direct = conf->pool_direct;
|
|
port->pool_indirect = conf->pool_indirect;
|
|
|
|
port->n_pkts = 0;
|
|
port->pos_pkts = 0;
|
|
port->n_frags = 0;
|
|
port->pos_frags = 0;
|
|
|
|
return port;
|
|
}
|
|
|
|
static int
|
|
rte_port_ring_reader_ipv4_frag_rx(void *port,
|
|
struct rte_mbuf **pkts,
|
|
uint32_t n_pkts)
|
|
{
|
|
struct rte_port_ring_reader_ipv4_frag *p =
|
|
(struct rte_port_ring_reader_ipv4_frag *) port;
|
|
uint32_t n_pkts_out;
|
|
|
|
n_pkts_out = 0;
|
|
|
|
/* Get packets from the "frag" buffer */
|
|
if (p->n_frags >= n_pkts) {
|
|
memcpy(pkts, &p->frags[p->pos_frags], n_pkts * sizeof(void *));
|
|
p->pos_frags += n_pkts;
|
|
p->n_frags -= n_pkts;
|
|
|
|
return n_pkts;
|
|
}
|
|
|
|
memcpy(pkts, &p->frags[p->pos_frags], p->n_frags * sizeof(void *));
|
|
n_pkts_out = p->n_frags;
|
|
p->n_frags = 0;
|
|
|
|
/* Look to "pkts" buffer to get more packets */
|
|
for ( ; ; ) {
|
|
struct rte_mbuf *pkt;
|
|
uint32_t n_pkts_to_provide, i;
|
|
int status;
|
|
|
|
/* If "pkts" buffer is empty, read packet burst from ring */
|
|
if (p->n_pkts == 0) {
|
|
p->n_pkts = rte_ring_sc_dequeue_burst(p->ring,
|
|
(void **) p->pkts, RTE_PORT_IN_BURST_SIZE_MAX);
|
|
if (p->n_pkts == 0)
|
|
return n_pkts_out;
|
|
p->pos_pkts = 0;
|
|
}
|
|
|
|
/* Read next packet from "pkts" buffer */
|
|
pkt = p->pkts[p->pos_pkts++];
|
|
p->n_pkts--;
|
|
|
|
/* If not jumbo, pass current packet to output */
|
|
if (pkt->pkt_len <= IPV4_MTU_DEFAULT) {
|
|
pkts[n_pkts_out++] = pkt;
|
|
|
|
n_pkts_to_provide = n_pkts - n_pkts_out;
|
|
if (n_pkts_to_provide == 0)
|
|
return n_pkts;
|
|
|
|
continue;
|
|
}
|
|
|
|
/* Fragment current packet into the "frags" buffer */
|
|
status = rte_ipv4_fragment_packet(
|
|
pkt,
|
|
p->frags,
|
|
IPV4_MAX_FRAGS_PER_PACKET,
|
|
p->mtu,
|
|
p->pool_direct,
|
|
p->pool_indirect
|
|
);
|
|
|
|
if (status < 0) {
|
|
rte_pktmbuf_free(pkt);
|
|
continue;
|
|
}
|
|
|
|
p->n_frags = (uint32_t) status;
|
|
p->pos_frags = 0;
|
|
|
|
/* Copy meta-data from input jumbo packet to its fragments */
|
|
for (i = 0; i < p->n_frags; i++) {
|
|
uint8_t *src = RTE_MBUF_METADATA_UINT8_PTR(pkt, 0);
|
|
uint8_t *dst =
|
|
RTE_MBUF_METADATA_UINT8_PTR(p->frags[i], 0);
|
|
|
|
memcpy(dst, src, p->metadata_size);
|
|
}
|
|
|
|
/* Free input jumbo packet */
|
|
rte_pktmbuf_free(pkt);
|
|
|
|
/* Get packets from "frag" buffer */
|
|
n_pkts_to_provide = n_pkts - n_pkts_out;
|
|
if (p->n_frags >= n_pkts_to_provide) {
|
|
memcpy(&pkts[n_pkts_out], p->frags,
|
|
n_pkts_to_provide * sizeof(void *));
|
|
p->n_frags -= n_pkts_to_provide;
|
|
p->pos_frags += n_pkts_to_provide;
|
|
|
|
return n_pkts;
|
|
}
|
|
|
|
memcpy(&pkts[n_pkts_out], p->frags,
|
|
p->n_frags * sizeof(void *));
|
|
n_pkts_out += p->n_frags;
|
|
p->n_frags = 0;
|
|
}
|
|
}
|
|
|
|
static int
|
|
rte_port_ring_reader_ipv4_frag_free(void *port)
|
|
{
|
|
if (port == NULL) {
|
|
RTE_LOG(ERR, PORT, "%s: Parameter port is NULL\n", __func__);
|
|
return -1;
|
|
}
|
|
|
|
rte_free(port);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Summary of port operations
|
|
*/
|
|
struct rte_port_in_ops rte_port_ring_reader_ipv4_frag_ops = {
|
|
.f_create = rte_port_ring_reader_ipv4_frag_create,
|
|
.f_free = rte_port_ring_reader_ipv4_frag_free,
|
|
.f_rx = rte_port_ring_reader_ipv4_frag_rx,
|
|
};
|