numam-dpdk/lib/librte_acl/acl_run_scalar.c
Konstantin Ananyev 074f54ad03 acl: fix build and runtime for default target
Make ACL library to build/work on 'default' architecture:
- make rte_acl_classify_scalar really scalar
 (make sure it wouldn't use sse4 instrincts through resolve_priority()).
- Provide two versions of rte_acl_classify code path:
  rte_acl_classify_sse() - could be build and used only on systems with sse4.2
  and upper, return -ENOTSUP on lower arch.
  rte_acl_classify_scalar() - a slower version, but could be build and used
  on all systems.
- Addition of a new function rte_acl_classify_alg.  This function lets you
  specify an enum value to override the acl contexts default algorithm when doing
  a classification.  This allows an application to specify a classification
  algorithm without needing to publicize each method. I know there was concern
  over keeping those methods public, but we don't have a static ABI at the moment,
  so this seems to me a reasonable thing to do, as it gives us less of an ABI
  surface to worry about.
- keep common code shared between these two codepaths.

Signed-off-by: Konstantin Ananyev <konstantin.ananyev@intel.com>
Acked-by: Neil Horman <nhorman@tuxdriver.com>
2014-09-03 03:26:50 +02:00

194 lines
6.3 KiB
C

/*-
* BSD LICENSE
*
* Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in
* the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Intel Corporation nor the names of its
* contributors may be used to endorse or promote products derived
* from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "acl_run.h"
/*
* Resolve priority for multiple results (scalar version).
* This consists comparing the priority of the current traversal with the
* running set of results for the packet.
* For each result, keep a running array of the result (rule number) and
* its priority for each category.
*/
static inline void
resolve_priority_scalar(uint64_t transition, int n,
const struct rte_acl_ctx *ctx, struct parms *parms,
const struct rte_acl_match_results *p, uint32_t categories)
{
uint32_t i;
int32_t *saved_priority;
uint32_t *saved_results;
const int32_t *priority;
const uint32_t *results;
saved_results = parms[n].cmplt->results;
saved_priority = parms[n].cmplt->priority;
/* results and priorities for completed trie */
results = p[transition].results;
priority = p[transition].priority;
/* if this is not the first completed trie */
if (parms[n].cmplt->count != ctx->num_tries) {
for (i = 0; i < categories; i += RTE_ACL_RESULTS_MULTIPLIER) {
if (saved_priority[i] <= priority[i]) {
saved_priority[i] = priority[i];
saved_results[i] = results[i];
}
if (saved_priority[i + 1] <= priority[i + 1]) {
saved_priority[i + 1] = priority[i + 1];
saved_results[i + 1] = results[i + 1];
}
if (saved_priority[i + 2] <= priority[i + 2]) {
saved_priority[i + 2] = priority[i + 2];
saved_results[i + 2] = results[i + 2];
}
if (saved_priority[i + 3] <= priority[i + 3]) {
saved_priority[i + 3] = priority[i + 3];
saved_results[i + 3] = results[i + 3];
}
}
} else {
for (i = 0; i < categories; i += RTE_ACL_RESULTS_MULTIPLIER) {
saved_priority[i] = priority[i];
saved_priority[i + 1] = priority[i + 1];
saved_priority[i + 2] = priority[i + 2];
saved_priority[i + 3] = priority[i + 3];
saved_results[i] = results[i];
saved_results[i + 1] = results[i + 1];
saved_results[i + 2] = results[i + 2];
saved_results[i + 3] = results[i + 3];
}
}
}
/*
* When processing the transition, rather than using if/else
* construct, the offset is calculated for DFA and QRANGE and
* then conditionally added to the address based on node type.
* This is done to avoid branch mis-predictions. Since the
* offset is rather simple calculation it is more efficient
* to do the calculation and do a condition move rather than
* a conditional branch to determine which calculation to do.
*/
static inline uint32_t
scan_forward(uint32_t input, uint32_t max)
{
return (input == 0) ? max : rte_bsf32(input);
}
static inline uint64_t
scalar_transition(const uint64_t *trans_table, uint64_t transition,
uint8_t input)
{
uint32_t addr, index, ranges, x, a, b, c;
/* break transition into component parts */
ranges = transition >> (sizeof(index) * CHAR_BIT);
/* calc address for a QRANGE node */
c = input * SCALAR_QRANGE_MULT;
a = ranges | SCALAR_QRANGE_MIN;
index = transition & ~RTE_ACL_NODE_INDEX;
a -= (c & SCALAR_QRANGE_MASK);
b = c & SCALAR_QRANGE_MIN;
addr = transition ^ index;
a &= SCALAR_QRANGE_MIN;
a ^= (ranges ^ b) & (a ^ b);
x = scan_forward(a, 32) >> 3;
addr += (index == RTE_ACL_NODE_DFA) ? input : x;
/* pickup next transition */
transition = *(trans_table + addr);
return transition;
}
int
rte_acl_classify_scalar(const struct rte_acl_ctx *ctx, const uint8_t **data,
uint32_t *results, uint32_t num, uint32_t categories)
{
int n;
uint64_t transition0, transition1;
uint32_t input0, input1;
struct acl_flow_data flows;
uint64_t index_array[MAX_SEARCHES_SCALAR];
struct completion cmplt[MAX_SEARCHES_SCALAR];
struct parms parms[MAX_SEARCHES_SCALAR];
if (categories != 1 &&
((RTE_ACL_RESULTS_MULTIPLIER - 1) & categories) != 0)
return -EINVAL;
acl_set_flow(&flows, cmplt, RTE_DIM(cmplt), data, results, num,
categories, ctx->trans_table);
for (n = 0; n < MAX_SEARCHES_SCALAR; n++) {
cmplt[n].count = 0;
index_array[n] = acl_start_next_trie(&flows, parms, n, ctx);
}
transition0 = index_array[0];
transition1 = index_array[1];
while (flows.started > 0) {
input0 = GET_NEXT_4BYTES(parms, 0);
input1 = GET_NEXT_4BYTES(parms, 1);
for (n = 0; n < 4; n++) {
if (likely((transition0 & RTE_ACL_NODE_MATCH) == 0))
transition0 = scalar_transition(flows.trans,
transition0, (uint8_t)input0);
input0 >>= CHAR_BIT;
if (likely((transition1 & RTE_ACL_NODE_MATCH) == 0))
transition1 = scalar_transition(flows.trans,
transition1, (uint8_t)input1);
input1 >>= CHAR_BIT;
}
if ((transition0 | transition1) & RTE_ACL_NODE_MATCH) {
transition0 = acl_match_check(transition0,
0, ctx, parms, &flows, resolve_priority_scalar);
transition1 = acl_match_check(transition1,
1, ctx, parms, &flows, resolve_priority_scalar);
}
}
return 0;
}