/*- * Copyright 2016-2023 Microchip Technology, Inc. and/or its subsidiaries. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ #include "smartpqi_includes.h" /* * Process internal RAID response in the case of success. */ void pqisrc_process_internal_raid_response_success(pqisrc_softstate_t *softs,rcb_t *rcb) { DBG_FUNC("IN\n"); rcb->status = PQI_STATUS_SUCCESS; rcb->req_pending = false; DBG_FUNC("OUT\n"); } /* Safely determines if cdb is available and if so, will return SCSI opcode or BMIC cmd if BMIC op code is detected */ uint8_t pqisrc_get_cmd_from_rcb(rcb_t *rcb) { uint8_t opcode = 0xFF; if (rcb && rcb->cdbp) { opcode = rcb->cdbp[0]; if (IS_BMIC_OPCODE(opcode)) return rcb->cdbp[6]; } return opcode; } /* * Process internal RAID response in the case of failure. */ void pqisrc_process_internal_raid_response_error(pqisrc_softstate_t *softs, rcb_t *rcb, uint16_t err_idx) { raid_path_error_info_elem_t error_info; DBG_FUNC("IN\n"); rcb->error_info = (char *) (softs->err_buf_dma_mem.virt_addr) + (err_idx * PQI_ERROR_BUFFER_ELEMENT_LENGTH); memcpy(&error_info, rcb->error_info, sizeof(error_info)); rcb->status = PQI_STATUS_TIMEOUT; switch (error_info.data_out_result) { case PQI_RAID_DATA_IN_OUT_GOOD: if (error_info.status == PQI_RAID_DATA_IN_OUT_GOOD) rcb->status = PQI_STATUS_SUCCESS; break; case PQI_RAID_DATA_IN_OUT_UNDERFLOW: if (error_info.status == PQI_RAID_DATA_IN_OUT_GOOD || error_info.status == PQI_RAID_STATUS_CHECK_CONDITION) rcb->status = PQI_STATUS_SUCCESS; break; default: DBG_WARN("error_status 0x%x data_in_result 0x%x data_out_result 0x%x cmd rcb tag 0x%x\n", error_info.status, error_info.data_in_result, error_info.data_out_result, rcb->tag); } if (rcb->status != PQI_STATUS_SUCCESS) { DBG_INFO("error_status=0x%x data_in=0x%x data_out=0x%x detail=0x%x\n", error_info.status, error_info.data_in_result, error_info.data_out_result, pqisrc_get_cmd_from_rcb(rcb)); } rcb->req_pending = false; DBG_FUNC("OUT\n"); } /* * Process the AIO/RAID IO in the case of success. */ void pqisrc_process_io_response_success(pqisrc_softstate_t *softs, rcb_t *rcb) { DBG_FUNC("IN\n"); os_io_response_success(rcb); DBG_FUNC("OUT\n"); } static void pqisrc_extract_sense_data(sense_data_u_t *sense_data, uint8_t *key, uint8_t *asc, uint8_t *ascq) { if (sense_data->fixed_format.response_code == SCSI_SENSE_RESPONSE_70 || sense_data->fixed_format.response_code == SCSI_SENSE_RESPONSE_71) { sense_data_fixed_t *fixed = &sense_data->fixed_format; *key = fixed->sense_key; *asc = fixed->sense_code; *ascq = fixed->sense_qual; } else if (sense_data->descriptor_format.response_code == SCSI_SENSE_RESPONSE_72 || sense_data->descriptor_format.response_code == SCSI_SENSE_RESPONSE_73) { sense_data_descriptor_t *desc = &sense_data->descriptor_format; *key = desc->sense_key; *asc = desc->sense_code; *ascq = desc->sense_qual; } else { *key = 0xFF; *asc = 0xFF; *ascq = 0xFF; } } /* Suppress common errors unless verbose debug flag is on */ boolean_t suppress_innocuous_error_prints(pqisrc_softstate_t *softs, rcb_t *rcb) { uint8_t opcode = rcb->cdbp ? rcb->cdbp[0] : 0xFF; if ((opcode == SCSI_INQUIRY || /* 0x12 */ opcode == SCSI_MODE_SENSE || /* 0x1a */ opcode == SCSI_REPORT_LUNS || /* 0xa0 */ opcode == SCSI_LOG_SENSE || /* 0x4d */ opcode == SCSI_ATA_PASSTHRU16) /* 0x85 */ && (softs->err_resp_verbose == false)) return true; return false; } static void pqisrc_show_sense_data_simple(pqisrc_softstate_t *softs, rcb_t *rcb, sense_data_u_t *sense_data) { uint8_t opcode = rcb->cdbp ? rcb->cdbp[0] : 0xFF; char *path = io_path_to_ascii(rcb->path); uint8_t key, asc, ascq; pqisrc_extract_sense_data(sense_data, &key, &asc, &ascq); DBG_NOTE("[ERR INFO] BTL: %d:%d:%d op=0x%x path=%s K:C:Q: %x:%x:%x\n", rcb->dvp->bus, rcb->dvp->target, rcb->dvp->lun, opcode, path, key, asc, ascq); } void pqisrc_show_sense_data_full(pqisrc_softstate_t *softs, rcb_t *rcb, sense_data_u_t *sense_data) { if (suppress_innocuous_error_prints(softs, rcb)) return; pqisrc_print_buffer(softs, "sense data", sense_data, 32, 0); pqisrc_show_sense_data_simple(softs, rcb, sense_data); /* add more detail here as needed */ } /* dumps the aio error info and sense data then breaks down the output */ void pqisrc_show_aio_error_info(pqisrc_softstate_t *softs, rcb_t *rcb, aio_path_error_info_elem_t *aio_err) { DBG_NOTE("\n"); DBG_NOTE("aio err: status=0x%x serv_resp=0x%x data_pres=0x%x data_len=0x%x\n", aio_err->status, aio_err->service_resp, aio_err->data_pres, aio_err->data_len); pqisrc_print_buffer(softs, "aio err info", aio_err, offsetof(aio_path_error_info_elem_t, data), PRINT_FLAG_HDR_COLUMN); pqisrc_show_sense_data_full(softs, rcb, &aio_err->sense_data); } /* dumps the raid error info and sense data then breaks down the output */ void pqisrc_show_raid_error_info(pqisrc_softstate_t *softs, rcb_t *rcb, raid_path_error_info_elem_t *raid_err) { DBG_NOTE("\n"); DBG_NOTE("raid err: data_in=0x%x out=0x%x status=0x%x sense_len=0x%x resp_len=0x%x\n", raid_err->data_in_result, raid_err->data_in_result, raid_err->status, raid_err->sense_data_len, raid_err->resp_data_len); pqisrc_print_buffer(softs, "raid err info", raid_err, offsetof(raid_path_error_info_elem_t, data), PRINT_FLAG_HDR_COLUMN); pqisrc_show_sense_data_full(softs, rcb, &raid_err->sense_data); } /* return true if this an innocuous error */ boolean_t pqisrc_is_innocuous_error(pqisrc_softstate_t *softs, rcb_t *rcb, void *err_info) { uint8_t opcode = rcb->cdbp ? rcb->cdbp[0] : 0xFF; /* These SCSI cmds are frequently cause "underrun" and other minor "error" conditions while determining log page length, support, etc. */ if (opcode != SCSI_INQUIRY && /* 0x12 */ opcode != SCSI_MODE_SENSE && /* 0x1a */ opcode != SCSI_REPORT_LUNS && /* 0xa0 */ opcode != SCSI_LOG_SENSE && /* 0x4d */ opcode != SCSI_ATA_PASSTHRU16) /* 0x85 */ { return false; } /* treat all cmds above as innocuous unless verbose flag is set. */ if (softs->err_resp_verbose == false) return true; if (rcb->path == AIO_PATH) { aio_path_error_info_elem_t *aio_err = err_info; uint8_t key, asc, ascq; /* Byte[0]=Status=0x51, Byte[1]=service_resp=0x01 */ if (aio_err->status == PQI_AIO_STATUS_UNDERRUN && aio_err->service_resp == PQI_AIO_SERV_RESPONSE_FAILURE) { return true; } /* get the key info so we can apply more filters... */ pqisrc_extract_sense_data(&aio_err->sense_data, &key, &asc, &ascq); /* Seeing a lot of invalid field in CDB for REPORT LUNs on AIO path. Example CDB = a0 00 11 00 00 00 00 00 20 08 00 00 So filter out the full dump info for now. Also wonder if we should just send REPORT LUNS to raid path? */ if (opcode == SCSI_REPORT_LUNS && key == 5 && asc == 0x24) { pqisrc_show_sense_data_simple(softs, rcb, &aio_err->sense_data); return true; } /* may want to return true here eventually? */ } else { raid_path_error_info_elem_t *raid_err = err_info; /* Byte[1]=data_out=0x01 */ if (raid_err->data_out_result == PQI_RAID_DATA_IN_OUT_UNDERFLOW) return true; /* We get these a alot: leave a tiny breadcrumb about the error, but don't do full spew about it */ if (raid_err->status == PQI_AIO_STATUS_CHECK_CONDITION) { pqisrc_show_sense_data_simple(softs, rcb, &raid_err->sense_data); return true; } } return false; } /* * Process the error info for AIO in the case of failure. */ void pqisrc_process_aio_response_error(pqisrc_softstate_t *softs, rcb_t *rcb, uint16_t err_idx) { aio_path_error_info_elem_t *err_info = NULL; DBG_FUNC("IN\n"); ASSERT(rcb->path == AIO_PATH); err_info = (aio_path_error_info_elem_t*) softs->err_buf_dma_mem.virt_addr + err_idx; if(err_info == NULL) { DBG_ERR("err_info structure is NULL err_idx :%x\n", err_idx); return; } /* filter out certain underrun/success "errors" from printing */ if (!pqisrc_is_innocuous_error(softs, rcb, err_info)) { if (softs->err_resp_verbose == true) pqisrc_show_rcb_details(softs, rcb, "aio error", err_info); } os_aio_response_error(rcb, err_info); DBG_FUNC("OUT\n"); } /* * Process the error info for RAID IO in the case of failure. */ void pqisrc_process_raid_response_error(pqisrc_softstate_t *softs, rcb_t *rcb, uint16_t err_idx) { raid_path_error_info_elem_t *err_info = NULL; DBG_FUNC("IN\n"); ASSERT(rcb->path == RAID_PATH); err_info = (raid_path_error_info_elem_t*) softs->err_buf_dma_mem.virt_addr + err_idx; if(err_info == NULL) { DBG_ERR("err_info structure is NULL err_idx :%x\n", err_idx); return; } /* filter out certain underrun/success "errors" from printing */ if (!pqisrc_is_innocuous_error(softs, rcb, err_info)) { if( softs->err_resp_verbose == true ) pqisrc_show_rcb_details(softs, rcb, "raid error", err_info); } os_raid_response_error(rcb, err_info); DBG_FUNC("OUT\n"); } /* * Process the Task Management function response. */ int pqisrc_process_task_management_response(pqisrc_softstate_t *softs, pqi_tmf_resp_t *tmf_resp) { int ret = PQI_STATUS_SUCCESS; uint32_t tag = (uint32_t)tmf_resp->req_id; rcb_t *rcb = &softs->rcb[tag]; ASSERT(rcb->tag == tag); DBG_FUNC("IN\n"); switch (tmf_resp->resp_code) { case SOP_TASK_MANAGEMENT_FUNCTION_COMPLETE: case SOP_TASK_MANAGEMENT_FUNCTION_SUCCEEDED: ret = PQI_STATUS_SUCCESS; break; default: DBG_ERR("Tag #0x%08x TMF Failed, Response code : 0x%x\n", rcb->tag, tmf_resp->resp_code); ret = PQI_STATUS_TIMEOUT; break; } rcb->status = ret; rcb->req_pending = false; DBG_FUNC("OUT\n"); return ret; } static int pqisrc_process_vendor_general_response(pqi_vendor_general_response_t *response) { int ret = PQI_STATUS_SUCCESS; switch(response->status) { case PQI_VENDOR_RESPONSE_IU_SUCCESS: break; case PQI_VENDOR_RESPONSE_IU_UNSUCCESS: case PQI_VENDOR_RESPONSE_IU_INVALID_PARAM: case PQI_VENDOR_RESPONSE_IU_INSUFF_RESRC: ret = PQI_STATUS_TIMEOUT; break; } return ret; } /* * Function used to process the response from the adapter * which is invoked by IRQ handler. */ void pqisrc_process_response_queue(pqisrc_softstate_t *softs, int oq_id) { ob_queue_t *ob_q; struct pqi_io_response *response; uint32_t oq_pi, oq_ci; pqi_scsi_dev_t *dvp = NULL; DBG_FUNC("IN\n"); ob_q = &softs->op_ob_q[oq_id - 1]; /* zero for event Q */ oq_ci = ob_q->ci_local; oq_pi = *(ob_q->pi_virt_addr); DBG_IO("ci : %u pi : %u qid : %u\n", oq_ci, oq_pi, ob_q->q_id); while (1) { boolean_t os_scsi_cmd = false; rcb_t *rcb = NULL; uint32_t tag = 0; uint32_t offset; if (oq_pi == oq_ci) break; /* Get the response */ offset = oq_ci * ob_q->elem_size; response = (struct pqi_io_response *)(ob_q->array_virt_addr + offset); tag = response->request_id; rcb = &softs->rcb[tag]; /* Make sure we are processing a valid response. */ if ((rcb->tag != tag) || (rcb->req_pending == false)) { DBG_ERR("No such request pending with tag : %x rcb->tag : %x", tag, rcb->tag); oq_ci = (oq_ci + 1) % ob_q->num_elem; break; } /* Timedout request has been completed. This should not hit, * if timeout is set as TIMEOUT_INFINITE while calling * pqisrc_wait_on_condition(softs,rcb,timeout). */ if (rcb->timedout) { DBG_WARN("timed out request completing from firmware, driver already completed it with failure , free the tag 0x%x\n", tag); oq_ci = (oq_ci + 1) % ob_q->num_elem; os_reset_rcb(rcb); pqisrc_put_tag(&softs->taglist, tag); break; } if (rcb->host_wants_to_abort_this) { DBG_INFO("cmd that was aborted came back. tag=%u\n", rcb->tag); } if (rcb->is_abort_cmd_from_host) { DBG_INFO("abort cmd came back. tag=%u\n", rcb->tag); } if (IS_OS_SCSICMD(rcb)) { dvp = rcb->dvp; if (dvp) os_scsi_cmd = true; else DBG_WARN("Received IO completion for the Null device!!!\n"); } DBG_IO("response.header.iu_type : %x \n", response->header.iu_type); switch (response->header.iu_type) { case PQI_RESPONSE_IU_RAID_PATH_IO_SUCCESS: case PQI_RESPONSE_IU_AIO_PATH_IO_SUCCESS: rcb->success_cmp_callback(softs, rcb); if (os_scsi_cmd) pqisrc_decrement_device_active_io(softs, dvp); break; case PQI_RESPONSE_IU_RAID_PATH_IO_ERROR: case PQI_RESPONSE_IU_AIO_PATH_IO_ERROR: rcb->error_cmp_callback(softs, rcb, LE_16(response->error_index)); if (os_scsi_cmd) pqisrc_decrement_device_active_io(softs, dvp); break; case PQI_RESPONSE_IU_GENERAL_MANAGEMENT: rcb->req_pending = false; break; case PQI_RESPONSE_IU_VENDOR_GENERAL: rcb->req_pending = false; rcb->status = pqisrc_process_vendor_general_response( (pqi_vendor_general_response_t *)response); break; case PQI_RESPONSE_IU_TASK_MANAGEMENT: rcb->status = pqisrc_process_task_management_response(softs, (void *)response); break; default: DBG_ERR("Invalid Response IU 0x%x\n",response->header.iu_type); break; } oq_ci = (oq_ci + 1) % ob_q->num_elem; } ob_q->ci_local = oq_ci; PCI_MEM_PUT32(softs, ob_q->ci_register_abs, ob_q->ci_register_offset, ob_q->ci_local ); DBG_FUNC("OUT\n"); }