/* SPDX-License-Identifier: BSD-3-Clause * Copyright(C) 2019 Marvell International Ltd. */ #include "otx2_ethdev.h" #include "otx2_flow.h" static int flow_mcam_alloc_counter(struct otx2_mbox *mbox, uint16_t *ctr) { struct npc_mcam_alloc_counter_req *req; struct npc_mcam_alloc_counter_rsp *rsp; int rc; req = otx2_mbox_alloc_msg_npc_mcam_alloc_counter(mbox); req->count = 1; otx2_mbox_msg_send(mbox, 0); rc = otx2_mbox_get_rsp(mbox, 0, (void *)&rsp); *ctr = rsp->cntr_list[0]; return rc; } int otx2_flow_mcam_free_counter(struct otx2_mbox *mbox, uint16_t ctr_id) { struct npc_mcam_oper_counter_req *req; int rc; req = otx2_mbox_alloc_msg_npc_mcam_free_counter(mbox); req->cntr = ctr_id; otx2_mbox_msg_send(mbox, 0); rc = otx2_mbox_get_rsp(mbox, 0, NULL); return rc; } int otx2_flow_mcam_read_counter(struct otx2_mbox *mbox, uint32_t ctr_id, uint64_t *count) { struct npc_mcam_oper_counter_req *req; struct npc_mcam_oper_counter_rsp *rsp; int rc; req = otx2_mbox_alloc_msg_npc_mcam_counter_stats(mbox); req->cntr = ctr_id; otx2_mbox_msg_send(mbox, 0); rc = otx2_mbox_get_rsp(mbox, 0, (void *)&rsp); *count = rsp->stat; return rc; } int otx2_flow_mcam_clear_counter(struct otx2_mbox *mbox, uint32_t ctr_id) { struct npc_mcam_oper_counter_req *req; int rc; req = otx2_mbox_alloc_msg_npc_mcam_clear_counter(mbox); req->cntr = ctr_id; otx2_mbox_msg_send(mbox, 0); rc = otx2_mbox_get_rsp(mbox, 0, NULL); return rc; } int otx2_flow_mcam_free_entry(struct otx2_mbox *mbox, uint32_t entry) { struct npc_mcam_free_entry_req *req; int rc; req = otx2_mbox_alloc_msg_npc_mcam_free_entry(mbox); req->entry = entry; otx2_mbox_msg_send(mbox, 0); rc = otx2_mbox_get_rsp(mbox, 0, NULL); return rc; } int otx2_flow_mcam_free_all_entries(struct otx2_mbox *mbox) { struct npc_mcam_free_entry_req *req; int rc; req = otx2_mbox_alloc_msg_npc_mcam_free_entry(mbox); req->all = 1; otx2_mbox_msg_send(mbox, 0); rc = otx2_mbox_get_rsp(mbox, 0, NULL); return rc; } static void flow_prep_mcam_ldata(uint8_t *ptr, const uint8_t *data, int len) { int idx; for (idx = 0; idx < len; idx++) ptr[idx] = data[len - 1 - idx]; } static int flow_check_copysz(size_t size, size_t len) { if (len <= size) return len; return -1; } static inline int flow_mem_is_zero(const void *mem, int len) { const char *m = mem; int i; for (i = 0; i < len; i++) { if (m[i] != 0) return 0; } return 1; } static void flow_set_hw_mask(struct otx2_flow_item_info *info, struct npc_xtract_info *xinfo, char *hw_mask) { int max_off, offset; int j; if (xinfo->enable == 0) return; if (xinfo->hdr_off < info->hw_hdr_len) return; max_off = xinfo->hdr_off + xinfo->len - info->hw_hdr_len; if (max_off > info->len) max_off = info->len; offset = xinfo->hdr_off - info->hw_hdr_len; for (j = offset; j < max_off; j++) hw_mask[j] = 0xff; } void otx2_flow_get_hw_supp_mask(struct otx2_parse_state *pst, struct otx2_flow_item_info *info, int lid, int lt) { struct npc_xtract_info *xinfo, *lfinfo; char *hw_mask = info->hw_mask; int lf_cfg; int i, j; int intf; intf = pst->flow->nix_intf; xinfo = pst->npc->prx_dxcfg[intf][lid][lt].xtract; memset(hw_mask, 0, info->len); for (i = 0; i < NPC_MAX_LD; i++) { flow_set_hw_mask(info, &xinfo[i], hw_mask); } for (i = 0; i < NPC_MAX_LD; i++) { if (xinfo[i].flags_enable == 0) continue; lf_cfg = pst->npc->prx_lfcfg[i].i; if (lf_cfg == lid) { for (j = 0; j < NPC_MAX_LFL; j++) { lfinfo = pst->npc->prx_fxcfg[intf] [i][j].xtract; flow_set_hw_mask(info, &lfinfo[0], hw_mask); } } } } static int flow_update_extraction_data(struct otx2_parse_state *pst, struct otx2_flow_item_info *info, struct npc_xtract_info *xinfo) { uint8_t int_info_mask[NPC_MAX_EXTRACT_DATA_LEN]; uint8_t int_info[NPC_MAX_EXTRACT_DATA_LEN]; struct npc_xtract_info *x; int k, idx, hdr_off; int len = 0; x = xinfo; len = x->len; hdr_off = x->hdr_off; if (hdr_off < info->hw_hdr_len) return 0; if (x->enable == 0) return 0; otx2_npc_dbg("x->hdr_off = %d, len = %d, info->len = %d," "x->key_off = %d", x->hdr_off, len, info->len, x->key_off); hdr_off -= info->hw_hdr_len; if (hdr_off + len > info->len) len = info->len - hdr_off; /* Check for over-write of previous layer */ if (!flow_mem_is_zero(pst->mcam_mask + x->key_off, len)) { /* Cannot support this data match */ rte_flow_error_set(pst->error, ENOTSUP, RTE_FLOW_ERROR_TYPE_ITEM, pst->pattern, "Extraction unsupported"); return -rte_errno; } len = flow_check_copysz((OTX2_MAX_MCAM_WIDTH_DWORDS * 8) - x->key_off, len); if (len < 0) { rte_flow_error_set(pst->error, ENOTSUP, RTE_FLOW_ERROR_TYPE_ITEM, pst->pattern, "Internal Error"); return -rte_errno; } /* Need to reverse complete structure so that dest addr is at * MSB so as to program the MCAM using mcam_data & mcam_mask * arrays */ flow_prep_mcam_ldata(int_info, (const uint8_t *)info->spec + hdr_off, x->len); flow_prep_mcam_ldata(int_info_mask, (const uint8_t *)info->mask + hdr_off, x->len); otx2_npc_dbg("Spec: "); for (k = 0; k < info->len; k++) otx2_npc_dbg("0x%.2x ", ((const uint8_t *)info->spec)[k]); otx2_npc_dbg("Int_info: "); for (k = 0; k < info->len; k++) otx2_npc_dbg("0x%.2x ", int_info[k]); memcpy(pst->mcam_mask + x->key_off, int_info_mask, len); memcpy(pst->mcam_data + x->key_off, int_info, len); otx2_npc_dbg("Parse state mcam data & mask"); for (idx = 0; idx < len ; idx++) otx2_npc_dbg("data[%d]: 0x%x, mask[%d]: 0x%x", idx, *(pst->mcam_data + idx + x->key_off), idx, *(pst->mcam_mask + idx + x->key_off)); return 0; } int otx2_flow_update_parse_state(struct otx2_parse_state *pst, struct otx2_flow_item_info *info, int lid, int lt, uint8_t flags) { struct npc_lid_lt_xtract_info *xinfo; struct npc_xtract_info *lfinfo; int intf, lf_cfg; int i, j, rc = 0; otx2_npc_dbg("Parse state function info mask total %s", (const uint8_t *)info->mask); pst->layer_mask |= lid; pst->lt[lid] = lt; pst->flags[lid] = flags; intf = pst->flow->nix_intf; xinfo = &pst->npc->prx_dxcfg[intf][lid][lt]; otx2_npc_dbg("Is_terminating = %d", xinfo->is_terminating); if (xinfo->is_terminating) pst->terminate = 1; if (info->spec == NULL) { otx2_npc_dbg("Info spec NULL"); goto done; } for (i = 0; i < NPC_MAX_LD; i++) { rc = flow_update_extraction_data(pst, info, &xinfo->xtract[i]); if (rc != 0) return rc; } for (i = 0; i < NPC_MAX_LD; i++) { if (xinfo->xtract[i].flags_enable == 0) continue; lf_cfg = pst->npc->prx_lfcfg[i].i; if (lf_cfg == lid) { for (j = 0; j < NPC_MAX_LFL; j++) { lfinfo = pst->npc->prx_fxcfg[intf] [i][j].xtract; rc = flow_update_extraction_data(pst, info, &lfinfo[0]); if (rc != 0) return rc; if (lfinfo[0].enable) pst->flags[lid] = j; } } } done: /* Next pattern to parse by subsequent layers */ pst->pattern++; return 0; } static inline int flow_range_is_valid(const char *spec, const char *last, const char *mask, int len) { /* Mask must be zero or equal to spec as we do not support * non-contiguous ranges. */ while (len--) { if (last[len] && (spec[len] & mask[len]) != (last[len] & mask[len])) return 0; /* False */ } return 1; } static inline int flow_mask_is_supported(const char *mask, const char *hw_mask, int len) { /* * If no hw_mask, assume nothing is supported. * mask is never NULL */ if (hw_mask == NULL) return flow_mem_is_zero(mask, len); while (len--) { if ((mask[len] | hw_mask[len]) != hw_mask[len]) return 0; /* False */ } return 1; } int otx2_flow_parse_item_basic(const struct rte_flow_item *item, struct otx2_flow_item_info *info, struct rte_flow_error *error) { /* Item must not be NULL */ if (item == NULL) { rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_ITEM, NULL, "Item is NULL"); return -rte_errno; } /* If spec is NULL, both mask and last must be NULL, this * makes it to match ANY value (eq to mask = 0). * Setting either mask or last without spec is an error */ if (item->spec == NULL) { if (item->last == NULL && item->mask == NULL) { info->spec = NULL; return 0; } rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_ITEM, item, "mask or last set without spec"); return -rte_errno; } /* We have valid spec */ info->spec = item->spec; /* If mask is not set, use default mask, err if default mask is * also NULL. */ if (item->mask == NULL) { otx2_npc_dbg("Item mask null, using default mask"); if (info->def_mask == NULL) { rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_ITEM, item, "No mask or default mask given"); return -rte_errno; } info->mask = info->def_mask; } else { info->mask = item->mask; } /* mask specified must be subset of hw supported mask * mask | hw_mask == hw_mask */ if (!flow_mask_is_supported(info->mask, info->hw_mask, info->len)) { rte_flow_error_set(error, ENOTSUP, RTE_FLOW_ERROR_TYPE_ITEM, item, "Unsupported field in the mask"); return -rte_errno; } /* Now we have spec and mask. OTX2 does not support non-contiguous * range. We should have either: * - spec & mask == last & mask or, * - last == 0 or, * - last == NULL */ if (item->last != NULL && !flow_mem_is_zero(item->last, info->len)) { if (!flow_range_is_valid(item->spec, item->last, info->mask, info->len)) { rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_ITEM, item, "Unsupported range for match"); return -rte_errno; } } return 0; } void otx2_flow_keyx_compress(uint64_t *data, uint32_t nibble_mask) { uint64_t cdata[2] = {0ULL, 0ULL}, nibble; int i, j = 0; for (i = 0; i < NPC_MAX_KEY_NIBBLES; i++) { if (nibble_mask & (1 << i)) { nibble = (data[i / 16] >> ((i & 0xf) * 4)) & 0xf; cdata[j / 16] |= (nibble << ((j & 0xf) * 4)); j += 1; } } data[0] = cdata[0]; data[1] = cdata[1]; } static int flow_first_set_bit(uint64_t slab) { int num = 0; if ((slab & 0xffffffff) == 0) { num += 32; slab >>= 32; } if ((slab & 0xffff) == 0) { num += 16; slab >>= 16; } if ((slab & 0xff) == 0) { num += 8; slab >>= 8; } if ((slab & 0xf) == 0) { num += 4; slab >>= 4; } if ((slab & 0x3) == 0) { num += 2; slab >>= 2; } if ((slab & 0x1) == 0) num += 1; return num; } static int flow_shift_lv_ent(struct otx2_mbox *mbox, struct rte_flow *flow, struct otx2_npc_flow_info *flow_info, uint32_t old_ent, uint32_t new_ent) { struct npc_mcam_shift_entry_req *req; struct npc_mcam_shift_entry_rsp *rsp; struct otx2_flow_list *list; struct rte_flow *flow_iter; int rc = 0; otx2_npc_dbg("Old ent:%u new ent:%u priority:%u", old_ent, new_ent, flow->priority); list = &flow_info->flow_list[flow->priority]; /* Old entry is disabled & it's contents are moved to new_entry, * new entry is enabled finally. */ req = otx2_mbox_alloc_msg_npc_mcam_shift_entry(mbox); req->curr_entry[0] = old_ent; req->new_entry[0] = new_ent; req->shift_count = 1; otx2_mbox_msg_send(mbox, 0); rc = otx2_mbox_get_rsp(mbox, 0, (void *)&rsp); if (rc) return rc; /* Remove old node from list */ TAILQ_FOREACH(flow_iter, list, next) { if (flow_iter->mcam_id == old_ent) TAILQ_REMOVE(list, flow_iter, next); } /* Insert node with new mcam id at right place */ TAILQ_FOREACH(flow_iter, list, next) { if (flow_iter->mcam_id > new_ent) TAILQ_INSERT_BEFORE(flow_iter, flow, next); } return rc; } /* Exchange all required entries with a given priority level */ static int flow_shift_ent(struct otx2_mbox *mbox, struct rte_flow *flow, struct otx2_npc_flow_info *flow_info, struct npc_mcam_alloc_entry_rsp *rsp, int dir, int prio_lvl) { struct rte_bitmap *fr_bmp, *fr_bmp_rev, *lv_bmp, *lv_bmp_rev, *bmp; uint32_t e_fr = 0, e_lv = 0, e, e_id = 0, mcam_entries; uint64_t fr_bit_pos = 0, lv_bit_pos = 0, bit_pos = 0; /* Bit position within the slab */ uint32_t sl_fr_bit_off = 0, sl_lv_bit_off = 0; /* Overall bit position of the start of slab */ /* free & live entry index */ int rc_fr = 0, rc_lv = 0, rc = 0, idx = 0; struct otx2_mcam_ents_info *ent_info; /* free & live bitmap slab */ uint64_t sl_fr = 0, sl_lv = 0, *sl; fr_bmp = flow_info->free_entries[prio_lvl]; fr_bmp_rev = flow_info->free_entries_rev[prio_lvl]; lv_bmp = flow_info->live_entries[prio_lvl]; lv_bmp_rev = flow_info->live_entries_rev[prio_lvl]; ent_info = &flow_info->flow_entry_info[prio_lvl]; mcam_entries = flow_info->mcam_entries; /* New entries allocated are always contiguous, but older entries * already in free/live bitmap can be non-contiguous: so return * shifted entries should be in non-contiguous format. */ while (idx <= rsp->count) { if (!sl_fr && !sl_lv) { /* Lower index elements to be exchanged */ if (dir < 0) { rc_fr = rte_bitmap_scan(fr_bmp, &e_fr, &sl_fr); rc_lv = rte_bitmap_scan(lv_bmp, &e_lv, &sl_lv); otx2_npc_dbg("Fwd slab rc fr %u rc lv %u " "e_fr %u e_lv %u", rc_fr, rc_lv, e_fr, e_lv); } else { rc_fr = rte_bitmap_scan(fr_bmp_rev, &sl_fr_bit_off, &sl_fr); rc_lv = rte_bitmap_scan(lv_bmp_rev, &sl_lv_bit_off, &sl_lv); otx2_npc_dbg("Rev slab rc fr %u rc lv %u " "e_fr %u e_lv %u", rc_fr, rc_lv, e_fr, e_lv); } } if (rc_fr) { fr_bit_pos = flow_first_set_bit(sl_fr); e_fr = sl_fr_bit_off + fr_bit_pos; otx2_npc_dbg("Fr_bit_pos 0x%" PRIx64, fr_bit_pos); } else { e_fr = ~(0); } if (rc_lv) { lv_bit_pos = flow_first_set_bit(sl_lv); e_lv = sl_lv_bit_off + lv_bit_pos; otx2_npc_dbg("Lv_bit_pos 0x%" PRIx64, lv_bit_pos); } else { e_lv = ~(0); } /* First entry is from free_bmap */ if (e_fr < e_lv) { bmp = fr_bmp; e = e_fr; sl = &sl_fr; bit_pos = fr_bit_pos; if (dir > 0) e_id = mcam_entries - e - 1; else e_id = e; otx2_npc_dbg("Fr e %u e_id %u", e, e_id); } else { bmp = lv_bmp; e = e_lv; sl = &sl_lv; bit_pos = lv_bit_pos; if (dir > 0) e_id = mcam_entries - e - 1; else e_id = e; otx2_npc_dbg("Lv e %u e_id %u", e, e_id); if (idx < rsp->count) rc = flow_shift_lv_ent(mbox, flow, flow_info, e_id, rsp->entry + idx); } rte_bitmap_clear(bmp, e); rte_bitmap_set(bmp, rsp->entry + idx); /* Update entry list, use non-contiguous * list now. */ rsp->entry_list[idx] = e_id; *sl &= ~(1 << bit_pos); /* Update min & max entry identifiers in current * priority level. */ if (dir < 0) { ent_info->max_id = rsp->entry + idx; ent_info->min_id = e_id; } else { ent_info->max_id = e_id; ent_info->min_id = rsp->entry; } idx++; } return rc; } /* Validate if newly allocated entries lie in the correct priority zone * since NPC_MCAM_LOWER_PRIO & NPC_MCAM_HIGHER_PRIO don't ensure zone accuracy. * If not properly aligned, shift entries to do so */ static int flow_validate_and_shift_prio_ent(struct otx2_mbox *mbox, struct rte_flow *flow, struct otx2_npc_flow_info *flow_info, struct npc_mcam_alloc_entry_rsp *rsp, int req_prio) { int prio_idx = 0, rc = 0, needs_shift = 0, idx, prio = flow->priority; struct otx2_mcam_ents_info *info = flow_info->flow_entry_info; int dir = (req_prio == NPC_MCAM_HIGHER_PRIO) ? 1 : -1; uint32_t tot_ent = 0; otx2_npc_dbg("Dir %d, priority = %d", dir, prio); if (dir < 0) prio_idx = flow_info->flow_max_priority - 1; /* Only live entries needs to be shifted, free entries can just be * moved by bits manipulation. */ /* For dir = -1(NPC_MCAM_LOWER_PRIO), when shifting, * NPC_MAX_PREALLOC_ENT are exchanged with adjoining higher priority * level entries(lower indexes). * * For dir = +1(NPC_MCAM_HIGHER_PRIO), during shift, * NPC_MAX_PREALLOC_ENT are exchanged with adjoining lower priority * level entries(higher indexes) with highest indexes. */ do { tot_ent = info[prio_idx].free_ent + info[prio_idx].live_ent; if (dir < 0 && prio_idx != prio && rsp->entry > info[prio_idx].max_id && tot_ent) { otx2_npc_dbg("Rsp entry %u prio idx %u " "max id %u", rsp->entry, prio_idx, info[prio_idx].max_id); needs_shift = 1; } else if ((dir > 0) && (prio_idx != prio) && (rsp->entry < info[prio_idx].min_id) && tot_ent) { otx2_npc_dbg("Rsp entry %u prio idx %u " "min id %u", rsp->entry, prio_idx, info[prio_idx].min_id); needs_shift = 1; } otx2_npc_dbg("Needs_shift = %d", needs_shift); if (needs_shift) { needs_shift = 0; rc = flow_shift_ent(mbox, flow, flow_info, rsp, dir, prio_idx); } else { for (idx = 0; idx < rsp->count; idx++) rsp->entry_list[idx] = rsp->entry + idx; } } while ((prio_idx != prio) && (prio_idx += dir)); return rc; } static int flow_find_ref_entry(struct otx2_npc_flow_info *flow_info, int *prio, int prio_lvl) { struct otx2_mcam_ents_info *info = flow_info->flow_entry_info; int step = 1; while (step < flow_info->flow_max_priority) { if (((prio_lvl + step) < flow_info->flow_max_priority) && info[prio_lvl + step].live_ent) { *prio = NPC_MCAM_HIGHER_PRIO; return info[prio_lvl + step].min_id; } if (((prio_lvl - step) >= 0) && info[prio_lvl - step].live_ent) { otx2_npc_dbg("Prio_lvl %u live %u", prio_lvl - step, info[prio_lvl - step].live_ent); *prio = NPC_MCAM_LOWER_PRIO; return info[prio_lvl - step].max_id; } step++; } *prio = NPC_MCAM_ANY_PRIO; return 0; } static int flow_fill_entry_cache(struct otx2_mbox *mbox, struct rte_flow *flow, struct otx2_npc_flow_info *flow_info, uint32_t *free_ent) { struct rte_bitmap *free_bmp, *free_bmp_rev, *live_bmp, *live_bmp_rev; struct npc_mcam_alloc_entry_rsp rsp_local; struct npc_mcam_alloc_entry_rsp *rsp_cmd; struct npc_mcam_alloc_entry_req *req; struct npc_mcam_alloc_entry_rsp *rsp; struct otx2_mcam_ents_info *info; uint16_t ref_ent, idx; int rc, prio; info = &flow_info->flow_entry_info[flow->priority]; free_bmp = flow_info->free_entries[flow->priority]; free_bmp_rev = flow_info->free_entries_rev[flow->priority]; live_bmp = flow_info->live_entries[flow->priority]; live_bmp_rev = flow_info->live_entries_rev[flow->priority]; ref_ent = flow_find_ref_entry(flow_info, &prio, flow->priority); req = otx2_mbox_alloc_msg_npc_mcam_alloc_entry(mbox); req->contig = 1; req->count = flow_info->flow_prealloc_size; req->priority = prio; req->ref_entry = ref_ent; otx2_npc_dbg("Fill cache ref entry %u prio %u", ref_ent, prio); otx2_mbox_msg_send(mbox, 0); rc = otx2_mbox_get_rsp(mbox, 0, (void *)&rsp_cmd); if (rc) return rc; rsp = &rsp_local; memcpy(rsp, rsp_cmd, sizeof(*rsp)); otx2_npc_dbg("Alloc entry %u count %u , prio = %d", rsp->entry, rsp->count, prio); /* Non-first ent cache fill */ if (prio != NPC_MCAM_ANY_PRIO) { flow_validate_and_shift_prio_ent(mbox, flow, flow_info, rsp, prio); } else { /* Copy into response entry list */ for (idx = 0; idx < rsp->count; idx++) rsp->entry_list[idx] = rsp->entry + idx; } otx2_npc_dbg("Fill entry cache rsp count %u", rsp->count); /* Update free entries, reverse free entries list, * min & max entry ids. */ for (idx = 0; idx < rsp->count; idx++) { if (unlikely(rsp->entry_list[idx] < info->min_id)) info->min_id = rsp->entry_list[idx]; if (unlikely(rsp->entry_list[idx] > info->max_id)) info->max_id = rsp->entry_list[idx]; /* Skip entry to be returned, not to be part of free * list. */ if (prio == NPC_MCAM_HIGHER_PRIO) { if (unlikely(idx == (rsp->count - 1))) { *free_ent = rsp->entry_list[idx]; continue; } } else { if (unlikely(!idx)) { *free_ent = rsp->entry_list[idx]; continue; } } info->free_ent++; rte_bitmap_set(free_bmp, rsp->entry_list[idx]); rte_bitmap_set(free_bmp_rev, flow_info->mcam_entries - rsp->entry_list[idx] - 1); otx2_npc_dbg("Final rsp entry %u rsp entry rev %u", rsp->entry_list[idx], flow_info->mcam_entries - rsp->entry_list[idx] - 1); } otx2_npc_dbg("Cache free entry %u, rev = %u", *free_ent, flow_info->mcam_entries - *free_ent - 1); info->live_ent++; rte_bitmap_set(live_bmp, *free_ent); rte_bitmap_set(live_bmp_rev, flow_info->mcam_entries - *free_ent - 1); return 0; } static int flow_check_preallocated_entry_cache(struct otx2_mbox *mbox, struct rte_flow *flow, struct otx2_npc_flow_info *flow_info) { struct rte_bitmap *free, *free_rev, *live, *live_rev; uint32_t pos = 0, free_ent = 0, mcam_entries; struct otx2_mcam_ents_info *info; uint64_t slab = 0; int rc; otx2_npc_dbg("Flow priority %u", flow->priority); info = &flow_info->flow_entry_info[flow->priority]; free_rev = flow_info->free_entries_rev[flow->priority]; free = flow_info->free_entries[flow->priority]; live_rev = flow_info->live_entries_rev[flow->priority]; live = flow_info->live_entries[flow->priority]; mcam_entries = flow_info->mcam_entries; if (info->free_ent) { rc = rte_bitmap_scan(free, &pos, &slab); if (rc) { /* Get free_ent from free entry bitmap */ free_ent = pos + __builtin_ctzll(slab); otx2_npc_dbg("Allocated from cache entry %u", free_ent); /* Remove from free bitmaps and add to live ones */ rte_bitmap_clear(free, free_ent); rte_bitmap_set(live, free_ent); rte_bitmap_clear(free_rev, mcam_entries - free_ent - 1); rte_bitmap_set(live_rev, mcam_entries - free_ent - 1); info->free_ent--; info->live_ent++; return free_ent; } otx2_npc_dbg("No free entry:its a mess"); return -1; } rc = flow_fill_entry_cache(mbox, flow, flow_info, &free_ent); if (rc) return rc; return free_ent; } int otx2_flow_mcam_alloc_and_write(struct rte_flow *flow, struct otx2_mbox *mbox, __rte_unused struct otx2_parse_state *pst, struct otx2_npc_flow_info *flow_info) { int use_ctr = (flow->ctr_id == NPC_COUNTER_NONE ? 0 : 1); struct npc_mcam_write_entry_req *req; struct mbox_msghdr *rsp; uint16_t ctr = ~(0); int rc, idx; int entry; if (use_ctr) { rc = flow_mcam_alloc_counter(mbox, &ctr); if (rc) return rc; } entry = flow_check_preallocated_entry_cache(mbox, flow, flow_info); if (entry < 0) { otx2_err("Prealloc failed"); otx2_flow_mcam_free_counter(mbox, ctr); return NPC_MCAM_ALLOC_FAILED; } req = otx2_mbox_alloc_msg_npc_mcam_write_entry(mbox); req->set_cntr = use_ctr; req->cntr = ctr; req->entry = entry; otx2_npc_dbg("Alloc & write entry %u", entry); req->intf = (flow->nix_intf == OTX2_INTF_RX) ? NPC_MCAM_RX : NPC_MCAM_TX; req->enable_entry = 1; req->entry_data.action = flow->npc_action; /* * DPDK sets vtag action on per interface basis, not * per flow basis. It is a matter of how we decide to support * this pmd specific behavior. There are two ways: * 1. Inherit the vtag action from the one configured * for this interface. This can be read from the * vtag_action configured for default mcam entry of * this pf_func. * 2. Do not support vtag action with rte_flow. * * Second approach is used now. */ req->entry_data.vtag_action = 0ULL; for (idx = 0; idx < OTX2_MAX_MCAM_WIDTH_DWORDS; idx++) { req->entry_data.kw[idx] = flow->mcam_data[idx]; req->entry_data.kw_mask[idx] = flow->mcam_mask[idx]; } if (flow->nix_intf == OTX2_INTF_RX) { req->entry_data.kw[0] |= flow_info->channel; req->entry_data.kw_mask[0] |= (BIT_ULL(12) - 1); } else { uint16_t pf_func = (flow->npc_action >> 4) & 0xffff; pf_func = htons(pf_func); req->entry_data.kw[0] |= ((uint64_t)pf_func << 32); req->entry_data.kw_mask[0] |= ((uint64_t)0xffff << 32); } otx2_mbox_msg_send(mbox, 0); rc = otx2_mbox_get_rsp(mbox, 0, (void *)&rsp); if (rc != 0) return rc; flow->mcam_id = entry; if (use_ctr) flow->ctr_id = ctr; return 0; }