/* SPDX-License-Identifier: BSD-3-Clause * Copyright(c) 2014-2018 Broadcom * All rights reserved. */ #include #include #include #include "bnxt.h" #include "bnxt_hwrm.h" #include "bnxt_ring.h" #include "hsi_struct_def_dpdk.h" void bnxt_wait_for_device_shutdown(struct bnxt *bp) { uint32_t val, timeout; /* if HWRM_FUNC_QCAPS_OUTPUT_FLAGS_ERR_RECOVER_RELOAD is set * in HWRM_FUNC_QCAPS command, wait for FW_STATUS to set * the SHUTDOWN bit in health register */ if (!(bp->recovery_info && (bp->fw_cap & BNXT_FW_CAP_ERR_RECOVER_RELOAD))) return; /* Driver has to wait for fw_reset_max_msecs or shutdown bit which comes * first for FW to collect crash dump. */ timeout = bp->fw_reset_max_msecs; /* Driver has to poll for shutdown bit in fw_status register * * 1. in case of hot fw upgrade, this bit will be set after all * function drivers unregistered with fw. * 2. in case of fw initiated error recovery, this bit will be * set after fw has collected the core dump */ do { val = bnxt_read_fw_status_reg(bp, BNXT_FW_STATUS_REG); if (val & BNXT_FW_STATUS_SHUTDOWN) return; rte_delay_ms(100); timeout -= 100; } while (timeout); } static void bnxt_process_default_vnic_change(struct bnxt *bp, struct hwrm_async_event_cmpl *async_cmp) { uint16_t fid, vnic_state, parent_id, vf_fid, vf_id; struct bnxt_vf_representor *vf_rep_bp; struct rte_eth_dev *eth_dev; bool vfr_found = false; uint32_t event_data; if (!BNXT_TRUFLOW_EN(bp)) return; PMD_DRV_LOG(INFO, "Default vnic change async event received\n"); event_data = rte_le_to_cpu_32(async_cmp->event_data1); vnic_state = (event_data & BNXT_DEFAULT_VNIC_STATE_MASK) >> BNXT_DEFAULT_VNIC_STATE_SFT; if (vnic_state != BNXT_DEFAULT_VNIC_ALLOC) return; parent_id = (event_data & BNXT_DEFAULT_VNIC_CHANGE_PF_ID_MASK) >> BNXT_DEFAULT_VNIC_CHANGE_PF_ID_SFT; fid = BNXT_PF(bp) ? bp->fw_fid : bp->parent->fid; if (parent_id != fid || !bp->rep_info) return; vf_fid = (event_data & BNXT_DEFAULT_VNIC_CHANGE_VF_ID_MASK) >> BNXT_DEFAULT_VNIC_CHANGE_VF_ID_SFT; PMD_DRV_LOG(INFO, "async event received vf_id 0x%x\n", vf_fid); for (vf_id = 0; vf_id < BNXT_MAX_VF_REPS; vf_id++) { eth_dev = bp->rep_info[vf_id].vfr_eth_dev; if (!eth_dev) continue; vf_rep_bp = eth_dev->data->dev_private; if (vf_rep_bp && vf_rep_bp->fw_fid == vf_fid) { vfr_found = true; break; } } if (!vfr_found) return; bnxt_vf_rep_dev_start_op(eth_dev); } /* * Async event handling */ void bnxt_handle_async_event(struct bnxt *bp, struct cmpl_base *cmp) { struct hwrm_async_event_cmpl *async_cmp = (struct hwrm_async_event_cmpl *)cmp; uint16_t event_id = rte_le_to_cpu_16(async_cmp->event_id); struct bnxt_error_recovery_info *info; uint32_t event_data; switch (event_id) { case HWRM_ASYNC_EVENT_CMPL_EVENT_ID_LINK_STATUS_CHANGE: case HWRM_ASYNC_EVENT_CMPL_EVENT_ID_LINK_SPEED_CHANGE: case HWRM_ASYNC_EVENT_CMPL_EVENT_ID_LINK_SPEED_CFG_CHANGE: /* FALLTHROUGH */ bnxt_link_update(bp->eth_dev, 0, ETH_LINK_UP); break; case HWRM_ASYNC_EVENT_CMPL_EVENT_ID_PF_DRVR_UNLOAD: PMD_DRV_LOG(INFO, "Async event: PF driver unloaded\n"); break; case HWRM_ASYNC_EVENT_CMPL_EVENT_ID_VF_CFG_CHANGE: PMD_DRV_LOG(INFO, "Async event: VF config changed\n"); bnxt_hwrm_func_qcfg(bp, NULL); break; case HWRM_ASYNC_EVENT_CMPL_EVENT_ID_PORT_CONN_NOT_ALLOWED: PMD_DRV_LOG(INFO, "Port conn async event\n"); break; case HWRM_ASYNC_EVENT_CMPL_EVENT_ID_RESET_NOTIFY: /* Ignore reset notify async events when stopping the port */ if (!bp->eth_dev->data->dev_started) { bp->flags |= BNXT_FLAG_FATAL_ERROR; return; } event_data = rte_le_to_cpu_32(async_cmp->event_data1); /* timestamp_lo/hi values are in units of 100ms */ bp->fw_reset_max_msecs = async_cmp->timestamp_hi ? rte_le_to_cpu_16(async_cmp->timestamp_hi) * 100 : BNXT_MAX_FW_RESET_TIMEOUT; bp->fw_reset_min_msecs = async_cmp->timestamp_lo ? async_cmp->timestamp_lo * 100 : BNXT_MIN_FW_READY_TIMEOUT; if ((event_data & EVENT_DATA1_REASON_CODE_MASK) == EVENT_DATA1_REASON_CODE_FW_EXCEPTION_FATAL) { PMD_DRV_LOG(INFO, "Firmware fatal reset event received\n"); bp->flags |= BNXT_FLAG_FATAL_ERROR; } else { PMD_DRV_LOG(INFO, "Firmware non-fatal reset event received\n"); } bp->flags |= BNXT_FLAG_FW_RESET; rte_eal_alarm_set(US_PER_MS, bnxt_dev_reset_and_resume, (void *)bp); break; case HWRM_ASYNC_EVENT_CMPL_EVENT_ID_ERROR_RECOVERY: info = bp->recovery_info; if (!info) return; PMD_DRV_LOG(INFO, "Error recovery async event received\n"); event_data = rte_le_to_cpu_32(async_cmp->event_data1) & EVENT_DATA1_FLAGS_MASK; if (event_data & EVENT_DATA1_FLAGS_MASTER_FUNC) info->flags |= BNXT_FLAG_MASTER_FUNC; else info->flags &= ~BNXT_FLAG_MASTER_FUNC; if (event_data & EVENT_DATA1_FLAGS_RECOVERY_ENABLED) info->flags |= BNXT_FLAG_RECOVERY_ENABLED; else info->flags &= ~BNXT_FLAG_RECOVERY_ENABLED; PMD_DRV_LOG(INFO, "recovery enabled(%d), master function(%d)\n", bnxt_is_recovery_enabled(bp), bnxt_is_master_func(bp)); if (bp->flags & BNXT_FLAG_FW_HEALTH_CHECK_SCHEDULED) return; info->last_heart_beat = bnxt_read_fw_status_reg(bp, BNXT_FW_HEARTBEAT_CNT_REG); info->last_reset_counter = bnxt_read_fw_status_reg(bp, BNXT_FW_RECOVERY_CNT_REG); bnxt_schedule_fw_health_check(bp); break; case HWRM_ASYNC_EVENT_CMPL_EVENT_ID_DEBUG_NOTIFICATION: PMD_DRV_LOG(INFO, "DNC event: evt_data1 %#x evt_data2 %#x\n", rte_le_to_cpu_32(async_cmp->event_data1), rte_le_to_cpu_32(async_cmp->event_data2)); break; case HWRM_ASYNC_EVENT_CMPL_EVENT_ID_DEFAULT_VNIC_CHANGE: bnxt_process_default_vnic_change(bp, async_cmp); break; default: PMD_DRV_LOG(DEBUG, "handle_async_event id = 0x%x\n", event_id); break; } } void bnxt_handle_fwd_req(struct bnxt *bp, struct cmpl_base *cmpl) { struct hwrm_exec_fwd_resp_input *fwreq; struct hwrm_fwd_req_cmpl *fwd_cmpl = (struct hwrm_fwd_req_cmpl *)cmpl; struct input *fwd_cmd; uint16_t fw_vf_id; uint16_t vf_id; uint16_t req_len; int rc; if (bp->pf->active_vfs <= 0) { PMD_DRV_LOG(ERR, "Forwarded VF with no active VFs\n"); return; } /* Qualify the fwd request */ fw_vf_id = rte_le_to_cpu_16(fwd_cmpl->source_id); vf_id = fw_vf_id - bp->pf->first_vf_id; req_len = (rte_le_to_cpu_16(fwd_cmpl->req_len_type) & HWRM_FWD_REQ_CMPL_REQ_LEN_MASK) >> HWRM_FWD_REQ_CMPL_REQ_LEN_SFT; if (req_len > sizeof(fwreq->encap_request)) req_len = sizeof(fwreq->encap_request); /* Locate VF's forwarded command */ fwd_cmd = (struct input *)bp->pf->vf_info[vf_id].req_buf; if (fw_vf_id < bp->pf->first_vf_id || fw_vf_id >= bp->pf->first_vf_id + bp->pf->active_vfs) { PMD_DRV_LOG(ERR, "FWD req's source_id 0x%x out of range 0x%x - 0x%x (%d %d)\n", fw_vf_id, bp->pf->first_vf_id, (bp->pf->first_vf_id) + bp->pf->active_vfs - 1, bp->pf->first_vf_id, bp->pf->active_vfs); goto reject; } if (bnxt_rcv_msg_from_vf(bp, vf_id, fwd_cmd) == true) { /* * In older firmware versions, the MAC had to be all zeros for * the VF to set it's MAC via hwrm_func_vf_cfg. Set to all * zeros if it's being configured and has been ok'd by caller. */ if (fwd_cmd->req_type == HWRM_FUNC_VF_CFG) { struct hwrm_func_vf_cfg_input *vfc = (void *)fwd_cmd; if (vfc->enables & HWRM_FUNC_VF_CFG_INPUT_ENABLES_DFLT_MAC_ADDR) { bnxt_hwrm_func_vf_mac(bp, vf_id, (const uint8_t *)"\x00\x00\x00\x00\x00"); } } if (fwd_cmd->req_type == HWRM_CFA_L2_SET_RX_MASK) { struct hwrm_cfa_l2_set_rx_mask_input *srm = (void *)fwd_cmd; srm->vlan_tag_tbl_addr = rte_cpu_to_le_64(0); srm->num_vlan_tags = rte_cpu_to_le_32(0); srm->mask &= ~rte_cpu_to_le_32( HWRM_CFA_L2_SET_RX_MASK_INPUT_MASK_VLANONLY | HWRM_CFA_L2_SET_RX_MASK_INPUT_MASK_VLAN_NONVLAN | HWRM_CFA_L2_SET_RX_MASK_INPUT_MASK_ANYVLAN_NONVLAN); } /* Forward */ rc = bnxt_hwrm_exec_fwd_resp(bp, fw_vf_id, fwd_cmd, req_len); if (rc) { PMD_DRV_LOG(ERR, "Failed to send FWD req VF 0x%x, type 0x%x.\n", fw_vf_id - bp->pf->first_vf_id, rte_le_to_cpu_16(fwd_cmd->req_type)); } return; } reject: rc = bnxt_hwrm_reject_fwd_resp(bp, fw_vf_id, fwd_cmd, req_len); if (rc) { PMD_DRV_LOG(ERR, "Failed to send REJECT req VF 0x%x, type 0x%x.\n", fw_vf_id - bp->pf->first_vf_id, rte_le_to_cpu_16(fwd_cmd->req_type)); } return; } int bnxt_event_hwrm_resp_handler(struct bnxt *bp, struct cmpl_base *cmp) { bool evt = 0; if (bp == NULL || cmp == NULL) { PMD_DRV_LOG(ERR, "invalid NULL argument\n"); return evt; } if (unlikely(is_bnxt_in_error(bp))) return 0; switch (CMP_TYPE(cmp)) { case CMPL_BASE_TYPE_HWRM_ASYNC_EVENT: /* Handle any async event */ bnxt_handle_async_event(bp, cmp); evt = 1; break; case CMPL_BASE_TYPE_HWRM_FWD_RESP: /* Handle HWRM forwarded responses */ bnxt_handle_fwd_req(bp, cmp); evt = 1; break; default: /* Ignore any other events */ PMD_DRV_LOG(DEBUG, "Ignoring %02x completion\n", CMP_TYPE(cmp)); break; } return evt; } bool bnxt_is_master_func(struct bnxt *bp) { if (bp->recovery_info->flags & BNXT_FLAG_MASTER_FUNC) return true; return false; } bool bnxt_is_recovery_enabled(struct bnxt *bp) { struct bnxt_error_recovery_info *info; info = bp->recovery_info; if (info && (info->flags & BNXT_FLAG_RECOVERY_ENABLED)) return true; return false; }