numam-dpdk/lib/librte_distributor/distributor_private.h
Lukasz Wojciechowski 480d5a7c81 distributor: handle worker shutdown in burst mode
The burst version of distributor implementation was missing proper
handling of worker shutdown. A worker processing packets received
from distributor can call rte_distributor_return_pkt() function
informing distributor that it want no more packets. Further calls to
rte_distributor_request_pkt() or rte_distributor_get_pkt() however
should inform distributor that new packets are requested again.

Lack of the proper implementation has caused that even after worker
informed about returning last packets, new packets were still sent
from distributor causing deadlocks as no one could get them on worker
side.

This patch adds handling shutdown of the worker in following way:
1) It fixes usage of RTE_DISTRIB_VALID_BUF handshake flag. This flag
was formerly unused in burst implementation and now it is used
for marking valid packets in retptr64 replacing invalid use
of RTE_DISTRIB_RETURN_BUF flag.
2) Uses RTE_DISTRIB_RETURN_BUF as a worker to distributor handshake
in retptr64 to indicate that worker has shutdown.
3) Worker that shuts down blocks also bufptr for itself with
RTE_DISTRIB_RETURN_BUF flag allowing distributor to retrieve any
in flight packets.
4) When distributor receives information about shutdown of a worker,
it: marks worker as not active; retrieves any in flight and backlog
packets and process them to different workers; unlocks bufptr64
by clearing RTE_DISTRIB_RETURN_BUF flag and allowing use in
the future if worker requests any new packets.
5) Do not allow to: send or add to backlog any packets for not
active workers. Such workers are also ignored if matched.
6) Adjust calls to handle_returns() and tags matching procedure
to react for possible activation deactivation of workers.

Fixes: 775003ad2f ("distributor: add new burst-capable library")
Cc: stable@dpdk.org

Signed-off-by: Lukasz Wojciechowski <l.wojciechow@partner.samsung.com>
Acked-by: David Hunt <david.hunt@intel.com>
2020-10-19 10:57:17 +02:00

178 lines
5.4 KiB
C

/* SPDX-License-Identifier: BSD-3-Clause
* Copyright(c) 2017 Intel Corporation
*/
#ifndef _DIST_PRIV_H_
#define _DIST_PRIV_H_
/**
* @file
* RTE distributor
*
* The distributor is a component which is designed to pass packets
* one-at-a-time to workers, with dynamic load balancing.
*/
#ifdef __cplusplus
extern "C" {
#endif
#define NO_FLAGS 0
#define RTE_DISTRIB_PREFIX "DT_"
/*
* We will use the bottom four bits of pointer for flags, shifting out
* the top four bits to make room (since a 64-bit pointer actually only uses
* 48 bits). An arithmetic-right-shift will then appropriately restore the
* original pointer value with proper sign extension into the top bits.
*/
#define RTE_DISTRIB_FLAG_BITS 4
#define RTE_DISTRIB_FLAGS_MASK (0x0F)
#define RTE_DISTRIB_NO_BUF 0 /**< empty flags: no buffer requested */
#define RTE_DISTRIB_GET_BUF (1) /**< worker requests a buffer, returns old */
#define RTE_DISTRIB_RETURN_BUF (2) /**< worker returns a buffer, no request */
#define RTE_DISTRIB_VALID_BUF (4) /**< set if bufptr contains ptr */
#define RTE_DISTRIB_BACKLOG_SIZE 8
#define RTE_DISTRIB_BACKLOG_MASK (RTE_DISTRIB_BACKLOG_SIZE - 1)
#define RTE_DISTRIB_MAX_RETURNS 128
#define RTE_DISTRIB_RETURNS_MASK (RTE_DISTRIB_MAX_RETURNS - 1)
/**
* Maximum number of workers allowed.
* Be aware of increasing the limit, because it is limited by how we track
* in-flight tags. See in_flight_bitmask and rte_distributor_process
*/
#define RTE_DISTRIB_MAX_WORKERS 64
#define RTE_DISTRIBUTOR_NAMESIZE 32 /**< Length of name for instance */
/**
* Buffer structure used to pass the pointer data between cores. This is cache
* line aligned, but to improve performance and prevent adjacent cache-line
* prefetches of buffers for other workers, e.g. when worker 1's buffer is on
* the next cache line to worker 0, we pad this out to three cache lines.
* Only 64-bits of the memory is actually used though.
*/
union rte_distributor_buffer_single {
volatile int64_t bufptr64;
char pad[RTE_CACHE_LINE_SIZE*3];
} __rte_cache_aligned;
/*
* Transfer up to 8 mbufs at a time to/from workers, and
* flow matching algorithm optimized for 8 flow IDs at a time
*/
#define RTE_DIST_BURST_SIZE 8
struct rte_distributor_backlog {
unsigned int start;
unsigned int count;
int64_t pkts[RTE_DIST_BURST_SIZE] __rte_cache_aligned;
uint16_t *tags; /* will point to second cacheline of inflights */
} __rte_cache_aligned;
struct rte_distributor_returned_pkts {
unsigned int start;
unsigned int count;
struct rte_mbuf *mbufs[RTE_DISTRIB_MAX_RETURNS];
};
struct rte_distributor_single {
TAILQ_ENTRY(rte_distributor_single) next; /**< Next in list. */
char name[RTE_DISTRIBUTOR_NAMESIZE]; /**< Name of the ring. */
unsigned int num_workers; /**< Number of workers polling */
uint32_t in_flight_tags[RTE_DISTRIB_MAX_WORKERS];
/**< Tracks the tag being processed per core */
uint64_t in_flight_bitmask;
/**< on/off bits for in-flight tags.
* Note that if RTE_DISTRIB_MAX_WORKERS is larger than 64 then
* the bitmask has to expand.
*/
struct rte_distributor_backlog backlog[RTE_DISTRIB_MAX_WORKERS];
union rte_distributor_buffer_single bufs[RTE_DISTRIB_MAX_WORKERS];
struct rte_distributor_returned_pkts returns;
};
/* All different signature compare functions */
enum rte_distributor_match_function {
RTE_DIST_MATCH_SCALAR = 0,
RTE_DIST_MATCH_VECTOR,
RTE_DIST_NUM_MATCH_FNS
};
/**
* Buffer structure used to pass the pointer data between cores. This is cache
* line aligned, but to improve performance and prevent adjacent cache-line
* prefetches of buffers for other workers, e.g. when worker 1's buffer is on
* the next cache line to worker 0, we pad this out to two cache lines.
* We can pass up to 8 mbufs at a time in one cacheline.
* There is a separate cacheline for returns in the burst API.
*/
struct rte_distributor_buffer {
volatile int64_t bufptr64[RTE_DIST_BURST_SIZE]
__rte_cache_aligned; /* <= outgoing to worker */
int64_t pad1 __rte_cache_aligned; /* <= one cache line */
volatile int64_t retptr64[RTE_DIST_BURST_SIZE]
__rte_cache_aligned; /* <= incoming from worker */
int64_t pad2 __rte_cache_aligned; /* <= one cache line */
int count __rte_cache_aligned; /* <= number of current mbufs */
};
struct rte_distributor {
TAILQ_ENTRY(rte_distributor) next; /**< Next in list. */
char name[RTE_DISTRIBUTOR_NAMESIZE]; /**< Name of the ring. */
unsigned int num_workers; /**< Number of workers polling */
unsigned int alg_type; /**< Number of alg types */
/**>
* First cache line in the this array are the tags inflight
* on the worker core. Second cache line are the backlog
* that are going to go to the worker core.
*/
uint16_t in_flight_tags[RTE_DISTRIB_MAX_WORKERS][RTE_DIST_BURST_SIZE*2]
__rte_cache_aligned;
struct rte_distributor_backlog backlog[RTE_DISTRIB_MAX_WORKERS]
__rte_cache_aligned;
struct rte_distributor_buffer bufs[RTE_DISTRIB_MAX_WORKERS];
struct rte_distributor_returned_pkts returns;
enum rte_distributor_match_function dist_match_fn;
struct rte_distributor_single *d_single;
uint8_t active[RTE_DISTRIB_MAX_WORKERS];
uint8_t activesum;
};
void
find_match_scalar(struct rte_distributor *d,
uint16_t *data_ptr,
uint16_t *output_ptr);
void
find_match_vec(struct rte_distributor *d,
uint16_t *data_ptr,
uint16_t *output_ptr);
#ifdef __cplusplus
}
#endif
#endif /* _DIST_PRIV_H_ */