/* SPDX-License-Identifier: BSD-3-Clause * Copyright 2018 NXP */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include struct dpaa2_dpdmux_dev { TAILQ_ENTRY(dpaa2_dpdmux_dev) next; /**< Pointer to Next device instance */ struct fsl_mc_io dpdmux; /** handle to DPDMUX portal object */ uint16_t token; uint32_t dpdmux_id; /*HW ID for DPDMUX object */ uint8_t num_ifs; /* Number of interfaces in DPDMUX */ }; struct rte_flow { struct dpdmux_rule_cfg rule; }; TAILQ_HEAD(dpdmux_dev_list, dpaa2_dpdmux_dev); static struct dpdmux_dev_list dpdmux_dev_list = TAILQ_HEAD_INITIALIZER(dpdmux_dev_list); /*!< DPDMUX device list */ static struct dpaa2_dpdmux_dev *get_dpdmux_from_id(uint32_t dpdmux_id) { struct dpaa2_dpdmux_dev *dpdmux_dev = NULL; /* Get DPBP dev handle from list using index */ TAILQ_FOREACH(dpdmux_dev, &dpdmux_dev_list, next) { if (dpdmux_dev->dpdmux_id == dpdmux_id) break; } return dpdmux_dev; } struct rte_flow * rte_pmd_dpaa2_mux_flow_create(uint32_t dpdmux_id, struct rte_flow_item *pattern[], struct rte_flow_action *actions[]) { struct dpaa2_dpdmux_dev *dpdmux_dev; struct dpkg_profile_cfg kg_cfg; const struct rte_flow_item_ipv4 *spec; const struct rte_flow_action_vf *vf_conf; struct dpdmux_cls_action dpdmux_action; struct rte_flow *flow = NULL; void *key_iova, *mask_iova, *key_cfg_iova = NULL; int ret; if (pattern[0]->type != RTE_FLOW_ITEM_TYPE_IPV4) { DPAA2_PMD_ERR("Not supported pattern type: %d", pattern[0]->type); return NULL; } /* Find the DPDMUX from dpdmux_id in our list */ dpdmux_dev = get_dpdmux_from_id(dpdmux_id); if (!dpdmux_dev) { DPAA2_PMD_ERR("Invalid dpdmux_id: %d", dpdmux_id); return NULL; } key_cfg_iova = rte_zmalloc(NULL, DIST_PARAM_IOVA_SIZE, RTE_CACHE_LINE_SIZE); if (!key_cfg_iova) { DPAA2_PMD_ERR("Unable to allocate flow-dist parameters"); return NULL; } /* Currently taking only IP protocol as an extract type. * This can be exended to other fields using pattern->type. */ memset(&kg_cfg, 0, sizeof(struct dpkg_profile_cfg)); kg_cfg.extracts[0].extract.from_hdr.prot = NET_PROT_IP; kg_cfg.extracts[0].extract.from_hdr.field = NH_FLD_IP_PROTO; kg_cfg.extracts[0].type = DPKG_EXTRACT_FROM_HDR; kg_cfg.extracts[0].extract.from_hdr.type = DPKG_FULL_FIELD; kg_cfg.num_extracts = 1; ret = dpkg_prepare_key_cfg(&kg_cfg, key_cfg_iova); if (ret) { DPAA2_PMD_ERR("dpkg_prepare_key_cfg failed: err(%d)", ret); goto creation_error; } ret = dpdmux_set_custom_key(&dpdmux_dev->dpdmux, CMD_PRI_LOW, dpdmux_dev->token, (uint64_t)(DPAA2_VADDR_TO_IOVA(key_cfg_iova))); if (ret) { DPAA2_PMD_ERR("dpdmux_set_custom_key failed: err(%d)", ret); goto creation_error; } /* As now our key extract parameters are set, let us configure * the rule. */ flow = rte_zmalloc(NULL, sizeof(struct rte_flow) + (2 * DIST_PARAM_IOVA_SIZE), RTE_CACHE_LINE_SIZE); if (!flow) { DPAA2_PMD_ERR( "Memory allocation failure for rule configration\n"); goto creation_error; } key_iova = (void *)((size_t)flow + sizeof(struct rte_flow)); mask_iova = (void *)((size_t)key_iova + DIST_PARAM_IOVA_SIZE); spec = (const struct rte_flow_item_ipv4 *)pattern[0]->spec; memcpy(key_iova, (const void *)&spec->hdr.next_proto_id, sizeof(uint8_t)); memcpy(mask_iova, pattern[0]->mask, sizeof(uint8_t)); flow->rule.key_iova = (uint64_t)(DPAA2_VADDR_TO_IOVA(key_iova)); flow->rule.mask_iova = (uint64_t)(DPAA2_VADDR_TO_IOVA(mask_iova)); flow->rule.key_size = sizeof(uint8_t); vf_conf = (const struct rte_flow_action_vf *)(actions[0]->conf); if (vf_conf->id == 0 || vf_conf->id > dpdmux_dev->num_ifs) { DPAA2_PMD_ERR("Invalid destination id\n"); goto creation_error; } dpdmux_action.dest_if = vf_conf->id; ret = dpdmux_add_custom_cls_entry(&dpdmux_dev->dpdmux, CMD_PRI_LOW, dpdmux_dev->token, &flow->rule, &dpdmux_action); if (ret) { DPAA2_PMD_ERR("dpdmux_add_custom_cls_entry failed: err(%d)", ret); goto creation_error; } return flow; creation_error: rte_free((void *)key_cfg_iova); rte_free((void *)flow); return NULL; } static int dpaa2_create_dpdmux_device(int vdev_fd __rte_unused, struct vfio_device_info *obj_info __rte_unused, int dpdmux_id) { struct dpaa2_dpdmux_dev *dpdmux_dev; struct dpdmux_attr attr; int ret; PMD_INIT_FUNC_TRACE(); /* Allocate DPAA2 dpdmux handle */ dpdmux_dev = rte_malloc(NULL, sizeof(struct dpaa2_dpdmux_dev), 0); if (!dpdmux_dev) { DPAA2_PMD_ERR("Memory allocation failed for DPDMUX Device"); return -1; } /* Open the dpdmux object */ dpdmux_dev->dpdmux.regs = rte_mcp_ptr_list[MC_PORTAL_INDEX]; ret = dpdmux_open(&dpdmux_dev->dpdmux, CMD_PRI_LOW, dpdmux_id, &dpdmux_dev->token); if (ret) { DPAA2_PMD_ERR("Unable to open dpdmux object: err(%d)", ret); goto init_err; } ret = dpdmux_get_attributes(&dpdmux_dev->dpdmux, CMD_PRI_LOW, dpdmux_dev->token, &attr); if (ret) { DPAA2_PMD_ERR("Unable to get dpdmux attr: err(%d)", ret); goto init_err; } ret = dpdmux_if_set_default(&dpdmux_dev->dpdmux, CMD_PRI_LOW, dpdmux_dev->token, 1); if (ret) { DPAA2_PMD_ERR("setting default interface failed in %s", __func__); goto init_err; } dpdmux_dev->dpdmux_id = dpdmux_id; dpdmux_dev->num_ifs = attr.num_ifs; TAILQ_INSERT_TAIL(&dpdmux_dev_list, dpdmux_dev, next); return 0; init_err: if (dpdmux_dev) rte_free(dpdmux_dev); return -1; } static struct rte_dpaa2_object rte_dpaa2_dpdmux_obj = { .dev_type = DPAA2_MUX, .create = dpaa2_create_dpdmux_device, }; RTE_PMD_REGISTER_DPAA2_OBJECT(dpdmux, rte_dpaa2_dpdmux_obj);