1e66f787c8
- Microsemi SCSI driver for PQI controllers. - Found on newer model HP servers. - Restrict to AMD64 only as per developer request. The driver provides support for the new generation of PQI controllers from Microsemi. This driver is the first SCSI driver to implement the PQI queuing model and it will replace the aacraid driver for Adaptec Series 9 controllers. HARDWARE Controllers supported by the driver include: HPE Gen10 Smart Array Controller Family OEM Controllers based on the Microsemi Chipset. Submitted by: deepak.ukey@microsemi.com Relnotes: yes Sponsored by: Microsemi Differential Revision: https://reviews.freebsd.org/D14514
237 lines
6.1 KiB
C
237 lines
6.1 KiB
C
/*-
|
|
* Copyright (c) 2018 Microsemi Corporation.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
/* $FreeBSD$ */
|
|
|
|
#include "smartpqi_includes.h"
|
|
|
|
/*
|
|
* Process internal RAID response in the case of success.
|
|
*/
|
|
void pqisrc_process_internal_raid_response_success(pqisrc_softstate_t *softs,
|
|
rcb_t *rcb)
|
|
{
|
|
DBG_FUNC("IN");
|
|
|
|
rcb->status = REQUEST_SUCCESS;
|
|
rcb->req_pending = false;
|
|
|
|
DBG_FUNC("OUT");
|
|
}
|
|
|
|
/*
|
|
* Process internal RAID response in the case of failure.
|
|
*/
|
|
void pqisrc_process_internal_raid_response_error(pqisrc_softstate_t *softs,
|
|
rcb_t *rcb, uint16_t err_idx)
|
|
{
|
|
raid_path_error_info_elem_t error_info;
|
|
|
|
DBG_FUNC("IN");
|
|
|
|
rcb->error_info = (char *) (softs->err_buf_dma_mem.virt_addr) +
|
|
(err_idx * PQI_ERROR_BUFFER_ELEMENT_LENGTH);
|
|
rcb->status = REQUEST_SUCCESS;
|
|
memcpy(&error_info, rcb->error_info, sizeof(error_info));
|
|
|
|
DBG_INFO("error_status 0x%x data_in_result 0x%x data_out_result 0x%x\n",
|
|
error_info.status, error_info.data_in_result, error_info.data_out_result);
|
|
|
|
if (error_info.status != 0)
|
|
rcb->status = REQUEST_FAILED;
|
|
if (error_info.data_in_result != PQI_RAID_DATA_IN_OUT_GOOD)
|
|
rcb->status = REQUEST_FAILED;
|
|
if (error_info.data_out_result != PQI_RAID_DATA_IN_OUT_GOOD)
|
|
rcb->status = REQUEST_FAILED;
|
|
|
|
rcb->req_pending = false;
|
|
|
|
DBG_FUNC("OUT");
|
|
}
|
|
|
|
/*
|
|
* Process the AIO/RAID IO in the case of success.
|
|
*/
|
|
void pqisrc_process_io_response_success(pqisrc_softstate_t *softs,
|
|
rcb_t *rcb)
|
|
{
|
|
DBG_FUNC("IN");
|
|
|
|
os_io_response_success(rcb);
|
|
|
|
DBG_FUNC("OUT");
|
|
}
|
|
|
|
/*
|
|
* Process the error info for AIO in the case of failure.
|
|
*/
|
|
void pqisrc_process_aio_response_error(pqisrc_softstate_t *softs,
|
|
rcb_t *rcb, uint16_t err_idx)
|
|
{
|
|
aio_path_error_info_elem_t *err_info = NULL;
|
|
|
|
DBG_FUNC("IN");
|
|
|
|
err_info = (aio_path_error_info_elem_t*)
|
|
softs->err_buf_dma_mem.virt_addr +
|
|
err_idx;
|
|
|
|
if(err_info == NULL) {
|
|
DBG_ERR("err_info structure is NULL err_idx :%x", err_idx);
|
|
return;
|
|
}
|
|
|
|
os_aio_response_error(rcb, err_info);
|
|
|
|
DBG_FUNC("OUT");
|
|
}
|
|
|
|
/*
|
|
* Process the error info for RAID IO in the case of failure.
|
|
*/
|
|
void pqisrc_process_raid_response_error(pqisrc_softstate_t *softs,
|
|
rcb_t *rcb, uint16_t err_idx)
|
|
{
|
|
raid_path_error_info_elem_t *err_info = NULL;
|
|
|
|
DBG_FUNC("IN");
|
|
|
|
err_info = (raid_path_error_info_elem_t*)
|
|
softs->err_buf_dma_mem.virt_addr +
|
|
err_idx;
|
|
|
|
if(err_info == NULL) {
|
|
DBG_ERR("err_info structure is NULL err_idx :%x", err_idx);
|
|
return;
|
|
}
|
|
|
|
os_raid_response_error(rcb, err_info);
|
|
|
|
DBG_FUNC("OUT");
|
|
}
|
|
|
|
/*
|
|
* Process the Task Management function response.
|
|
*/
|
|
int pqisrc_process_task_management_response(pqisrc_softstate_t *softs,
|
|
pqi_tmf_resp_t *tmf_resp)
|
|
{
|
|
int ret = REQUEST_SUCCESS;
|
|
uint32_t tag = (uint32_t)tmf_resp->req_id;
|
|
rcb_t *rcb = &softs->rcb[tag];
|
|
|
|
ASSERT(rcb->tag == tag);
|
|
|
|
DBG_FUNC("IN\n");
|
|
|
|
switch (tmf_resp->resp_code) {
|
|
case SOP_TASK_MANAGEMENT_FUNCTION_COMPLETE:
|
|
case SOP_TASK_MANAGEMENT_FUNCTION_SUCCEEDED:
|
|
ret = REQUEST_SUCCESS;
|
|
break;
|
|
default:
|
|
DBG_ERR("TMF Failed, Response code : 0x%x\n", tmf_resp->resp_code);
|
|
ret = REQUEST_FAILED;
|
|
break;
|
|
}
|
|
|
|
rcb->status = ret;
|
|
rcb->req_pending = false;
|
|
|
|
DBG_FUNC("OUT");
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Function used to process the response from the adapter
|
|
* which is invoked by IRQ handler.
|
|
*/
|
|
void
|
|
pqisrc_process_response_queue(pqisrc_softstate_t *softs, int oq_id)
|
|
{
|
|
ob_queue_t *ob_q;
|
|
struct pqi_io_response *response;
|
|
uint32_t oq_pi, oq_ci;
|
|
|
|
DBG_FUNC("IN");
|
|
|
|
OS_ATOMIC64_INC(softs, num_intrs);
|
|
|
|
ob_q = &softs->op_ob_q[oq_id - 1]; /* zero for event Q */
|
|
oq_ci = ob_q->ci_local;
|
|
oq_pi = *(ob_q->pi_virt_addr);
|
|
|
|
DBG_INFO("ci : %d pi : %d qid : %d\n", oq_ci, oq_pi, ob_q->q_id);
|
|
|
|
while (1) {
|
|
rcb_t *rcb = NULL;
|
|
uint32_t tag = 0;
|
|
uint32_t offset;
|
|
|
|
if (oq_pi == oq_ci)
|
|
break;
|
|
/* Get the response */
|
|
offset = oq_ci * ob_q->elem_size;
|
|
response = (struct pqi_io_response *)(ob_q->array_virt_addr +
|
|
offset);
|
|
tag = response->request_id;
|
|
rcb = &softs->rcb[tag];
|
|
/* Make sure we are processing a valid response. */
|
|
ASSERT(rcb->tag == tag && rcb->req_pending);
|
|
rcb->req_pending = false;
|
|
|
|
DBG_INFO("response.header.iu_type : %x \n", response->header.iu_type);
|
|
|
|
switch (response->header.iu_type) {
|
|
case PQI_RESPONSE_IU_RAID_PATH_IO_SUCCESS:
|
|
case PQI_RESPONSE_IU_AIO_PATH_IO_SUCCESS:
|
|
rcb->success_cmp_callback(softs, rcb);
|
|
break;
|
|
case PQI_RESPONSE_IU_RAID_PATH_IO_ERROR:
|
|
case PQI_RESPONSE_IU_AIO_PATH_IO_ERROR:
|
|
rcb->error_cmp_callback(softs, rcb, LE_16(response->error_index));
|
|
break;
|
|
case PQI_RESPONSE_IU_GENERAL_MANAGEMENT:
|
|
rcb->req_pending = false;
|
|
break;
|
|
case PQI_RESPONSE_IU_TASK_MANAGEMENT:
|
|
rcb->status = pqisrc_process_task_management_response(softs, (void *)response);
|
|
break;
|
|
|
|
default:
|
|
DBG_ERR("Invalid Response IU 0x%x\n",response->header.iu_type);
|
|
break;
|
|
}
|
|
|
|
oq_ci = (oq_ci + 1) % ob_q->num_elem;
|
|
}
|
|
|
|
ob_q->ci_local = oq_ci;
|
|
PCI_MEM_PUT32(softs, ob_q->ci_register_abs,
|
|
ob_q->ci_register_offset, ob_q->ci_local );
|
|
DBG_FUNC("OUT");
|
|
}
|