freebsd-nq/sys/dev/buslogic/bt.c
Steven Hartland 85c9dd9d89 Prevent overflow issues in timeout processing
Previously, any timeout value for which (timeout * hz) will overflow the
signed integer, will give weird results, since callout(9) routines will
convert negative values of ticks to '1'. For unsigned integer overflow we
will get sufficiently smaller timeout values than expected.

Switch from callout_reset, which requires conversion to int based ticks
to callout_reset_sbt to avoid this.

Also correct isci to correctly resolve ccb timeout.

This was based on the original work done by Eygene Ryabinkin
<rea@freebsd.org> back in 5 Aug 2011 which used a macro to help avoid
the overlow.

Differential Revision:	https://reviews.freebsd.org/D1157
Reviewed by:	mav, davide
MFC after:	1 month
Sponsored by:	Multiplay
2014-11-21 21:01:24 +00:00

2394 lines
61 KiB
C

/*-
* Generic driver for the BusLogic MultiMaster SCSI host adapters
* Product specific probe and attach routines can be found in:
* sys/dev/buslogic/bt_isa.c BT-54X, BT-445 cards
* sys/dev/buslogic/bt_mca.c BT-64X, SDC3211B, SDC3211F
* sys/dev/buslogic/bt_eisa.c BT-74X, BT-75x cards, SDC3222F
* sys/dev/buslogic/bt_pci.c BT-946, BT-948, BT-956, BT-958 cards
*
* Copyright (c) 1998, 1999 Justin T. Gibbs.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions, and the following disclaimer,
* without modification, immediately at the beginning of the file.
* 2. The name of the author may not be used to endorse or promote products
* derived from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
* ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*/
#include <sys/cdefs.h>
__FBSDID("$FreeBSD$");
/*
* Special thanks to Leonard N. Zubkoff for writing such a complete and
* well documented Mylex/BusLogic MultiMaster driver for Linux. Support
* in this driver for the wide range of MultiMaster controllers and
* firmware revisions, with their otherwise undocumented quirks, would not
* have been possible without his efforts.
*/
#include <sys/param.h>
#include <sys/conf.h>
#include <sys/systm.h>
#include <sys/malloc.h>
#include <sys/kernel.h>
#include <sys/lock.h>
#include <sys/module.h>
#include <sys/mutex.h>
#include <sys/sysctl.h>
#include <sys/bus.h>
#include <machine/bus.h>
#include <sys/rman.h>
#include <cam/cam.h>
#include <cam/cam_ccb.h>
#include <cam/cam_sim.h>
#include <cam/cam_xpt_sim.h>
#include <cam/cam_debug.h>
#include <cam/scsi/scsi_message.h>
#include <vm/vm.h>
#include <vm/pmap.h>
#include <dev/buslogic/btreg.h>
/* MailBox Management functions */
static __inline void btnextinbox(struct bt_softc *bt);
static __inline void btnextoutbox(struct bt_softc *bt);
static __inline void
btnextinbox(struct bt_softc *bt)
{
if (bt->cur_inbox == bt->last_inbox)
bt->cur_inbox = bt->in_boxes;
else
bt->cur_inbox++;
}
static __inline void
btnextoutbox(struct bt_softc *bt)
{
if (bt->cur_outbox == bt->last_outbox)
bt->cur_outbox = bt->out_boxes;
else
bt->cur_outbox++;
}
/* CCB Mangement functions */
static __inline u_int32_t btccbvtop(struct bt_softc *bt,
struct bt_ccb *bccb);
static __inline struct bt_ccb* btccbptov(struct bt_softc *bt,
u_int32_t ccb_addr);
static __inline u_int32_t btsensepaddr(struct bt_softc *bt,
struct bt_ccb *bccb);
static __inline struct scsi_sense_data* btsensevaddr(struct bt_softc *bt,
struct bt_ccb *bccb);
static __inline u_int32_t
btccbvtop(struct bt_softc *bt, struct bt_ccb *bccb)
{
return (bt->bt_ccb_physbase
+ (u_int32_t)((caddr_t)bccb - (caddr_t)bt->bt_ccb_array));
}
static __inline struct bt_ccb *
btccbptov(struct bt_softc *bt, u_int32_t ccb_addr)
{
return (bt->bt_ccb_array +
((struct bt_ccb*)(uintptr_t)ccb_addr - (struct bt_ccb*)(uintptr_t)bt->bt_ccb_physbase));
}
static __inline u_int32_t
btsensepaddr(struct bt_softc *bt, struct bt_ccb *bccb)
{
u_int index;
index = (u_int)(bccb - bt->bt_ccb_array);
return (bt->sense_buffers_physbase
+ (index * sizeof(struct scsi_sense_data)));
}
static __inline struct scsi_sense_data *
btsensevaddr(struct bt_softc *bt, struct bt_ccb *bccb)
{
u_int index;
index = (u_int)(bccb - bt->bt_ccb_array);
return (bt->sense_buffers + index);
}
static __inline struct bt_ccb* btgetccb(struct bt_softc *bt);
static __inline void btfreeccb(struct bt_softc *bt,
struct bt_ccb *bccb);
static void btallocccbs(struct bt_softc *bt);
static bus_dmamap_callback_t btexecuteccb;
static void btdone(struct bt_softc *bt, struct bt_ccb *bccb,
bt_mbi_comp_code_t comp_code);
static void bt_intr_locked(struct bt_softc *bt);
/* Host adapter command functions */
static int btreset(struct bt_softc* bt, int hard_reset);
/* Initialization functions */
static int btinitmboxes(struct bt_softc *bt);
static bus_dmamap_callback_t btmapmboxes;
static bus_dmamap_callback_t btmapccbs;
static bus_dmamap_callback_t btmapsgs;
/* Transfer Negotiation Functions */
static void btfetchtransinfo(struct bt_softc *bt,
struct ccb_trans_settings *cts);
/* CAM SIM entry points */
#define ccb_bccb_ptr spriv_ptr0
#define ccb_bt_ptr spriv_ptr1
static void btaction(struct cam_sim *sim, union ccb *ccb);
static void btpoll(struct cam_sim *sim);
/* Our timeout handler */
static void bttimeout(void *arg);
/*
* XXX
* Do our own re-probe protection until a configuration
* manager can do it for us. This ensures that we don't
* reprobe a card already found by the EISA or PCI probes.
*/
struct bt_isa_port bt_isa_ports[] =
{
{ 0x130, 0, 4 },
{ 0x134, 0, 5 },
{ 0x230, 0, 2 },
{ 0x234, 0, 3 },
{ 0x330, 0, 0 },
{ 0x334, 0, 1 }
};
/*
* I/O ports listed in the order enumerated by the
* card for certain op codes.
*/
u_int16_t bt_board_ports[] =
{
0x330,
0x334,
0x230,
0x234,
0x130,
0x134
};
/* Exported functions */
void
bt_init_softc(device_t dev, struct resource *port,
struct resource *irq, struct resource *drq)
{
struct bt_softc *bt = device_get_softc(dev);
SLIST_INIT(&bt->free_bt_ccbs);
LIST_INIT(&bt->pending_ccbs);
SLIST_INIT(&bt->sg_maps);
bt->dev = dev;
bt->port = port;
bt->irq = irq;
bt->drq = drq;
mtx_init(&bt->lock, "bt", NULL, MTX_DEF);
}
void
bt_free_softc(device_t dev)
{
struct bt_softc *bt = device_get_softc(dev);
switch (bt->init_level) {
default:
case 11:
bus_dmamap_unload(bt->sense_dmat, bt->sense_dmamap);
case 10:
bus_dmamem_free(bt->sense_dmat, bt->sense_buffers,
bt->sense_dmamap);
case 9:
bus_dma_tag_destroy(bt->sense_dmat);
case 8:
{
struct sg_map_node *sg_map;
while ((sg_map = SLIST_FIRST(&bt->sg_maps))!= NULL) {
SLIST_REMOVE_HEAD(&bt->sg_maps, links);
bus_dmamap_unload(bt->sg_dmat,
sg_map->sg_dmamap);
bus_dmamem_free(bt->sg_dmat, sg_map->sg_vaddr,
sg_map->sg_dmamap);
free(sg_map, M_DEVBUF);
}
bus_dma_tag_destroy(bt->sg_dmat);
}
case 7:
bus_dmamap_unload(bt->ccb_dmat, bt->ccb_dmamap);
/* FALLTHROUGH */
case 6:
bus_dmamem_free(bt->ccb_dmat, bt->bt_ccb_array,
bt->ccb_dmamap);
/* FALLTHROUGH */
case 5:
bus_dma_tag_destroy(bt->ccb_dmat);
/* FALLTHROUGH */
case 4:
bus_dmamap_unload(bt->mailbox_dmat, bt->mailbox_dmamap);
/* FALLTHROUGH */
case 3:
bus_dmamem_free(bt->mailbox_dmat, bt->in_boxes,
bt->mailbox_dmamap);
/* FALLTHROUGH */
case 2:
bus_dma_tag_destroy(bt->buffer_dmat);
/* FALLTHROUGH */
case 1:
bus_dma_tag_destroy(bt->mailbox_dmat);
/* FALLTHROUGH */
case 0:
break;
}
mtx_destroy(&bt->lock);
}
int
bt_port_probe(device_t dev, struct bt_probe_info *info)
{
struct bt_softc *bt = device_get_softc(dev);
config_data_t config_data;
int error;
/* See if there is really a card present */
if (bt_probe(dev) || bt_fetch_adapter_info(dev))
return(1);
/*
* Determine our IRQ, and DMA settings and
* export them to the configuration system.
*/
mtx_lock(&bt->lock);
error = bt_cmd(bt, BOP_INQUIRE_CONFIG, NULL, /*parmlen*/0,
(u_int8_t*)&config_data, sizeof(config_data),
DEFAULT_CMD_TIMEOUT);
mtx_unlock(&bt->lock);
if (error != 0) {
printf("bt_port_probe: Could not determine IRQ or DMA "
"settings for adapter.\n");
return (1);
}
if (bt->model[0] == '5') {
/* DMA settings only make sense for ISA cards */
switch (config_data.dma_chan) {
case DMA_CHAN_5:
info->drq = 5;
break;
case DMA_CHAN_6:
info->drq = 6;
break;
case DMA_CHAN_7:
info->drq = 7;
break;
default:
printf("bt_port_probe: Invalid DMA setting "
"detected for adapter.\n");
return (1);
}
} else {
/* VL/EISA/PCI DMA */
info->drq = -1;
}
switch (config_data.irq) {
case IRQ_9:
case IRQ_10:
case IRQ_11:
case IRQ_12:
case IRQ_14:
case IRQ_15:
info->irq = ffs(config_data.irq) + 8;
break;
default:
printf("bt_port_probe: Invalid IRQ setting %x"
"detected for adapter.\n", config_data.irq);
return (1);
}
return (0);
}
/*
* Probe the adapter and verify that the card is a BusLogic.
*/
int
bt_probe(device_t dev)
{
struct bt_softc *bt = device_get_softc(dev);
esetup_info_data_t esetup_info;
u_int status;
u_int intstat;
u_int geometry;
int error;
u_int8_t param;
/*
* See if the three I/O ports look reasonable.
* Touch the minimal number of registers in the
* failure case.
*/
status = bt_inb(bt, STATUS_REG);
if ((status == 0)
|| (status & (DIAG_ACTIVE|CMD_REG_BUSY|
STATUS_REG_RSVD|CMD_INVALID)) != 0) {
if (bootverbose)
device_printf(dev, "Failed Status Reg Test - %x\n",
status);
return (ENXIO);
}
intstat = bt_inb(bt, INTSTAT_REG);
if ((intstat & INTSTAT_REG_RSVD) != 0) {
device_printf(dev, "Failed Intstat Reg Test\n");
return (ENXIO);
}
geometry = bt_inb(bt, GEOMETRY_REG);
if (geometry == 0xFF) {
if (bootverbose)
device_printf(dev, "Failed Geometry Reg Test\n");
return (ENXIO);
}
/*
* Looking good so far. Final test is to reset the
* adapter and attempt to fetch the extended setup
* information. This should filter out all 1542 cards.
*/
mtx_lock(&bt->lock);
if ((error = btreset(bt, /*hard_reset*/TRUE)) != 0) {
mtx_unlock(&bt->lock);
if (bootverbose)
device_printf(dev, "Failed Reset\n");
return (ENXIO);
}
param = sizeof(esetup_info);
error = bt_cmd(bt, BOP_INQUIRE_ESETUP_INFO, &param, /*parmlen*/1,
(u_int8_t*)&esetup_info, sizeof(esetup_info),
DEFAULT_CMD_TIMEOUT);
mtx_unlock(&bt->lock);
if (error != 0) {
return (ENXIO);
}
return (0);
}
/*
* Pull the boards setup information and record it in our softc.
*/
int
bt_fetch_adapter_info(device_t dev)
{
struct bt_softc *bt = device_get_softc(dev);
board_id_data_t board_id;
esetup_info_data_t esetup_info;
config_data_t config_data;
int error;
u_int8_t length_param;
/* First record the firmware version */
mtx_lock(&bt->lock);
error = bt_cmd(bt, BOP_INQUIRE_BOARD_ID, NULL, /*parmlen*/0,
(u_int8_t*)&board_id, sizeof(board_id),
DEFAULT_CMD_TIMEOUT);
if (error != 0) {
mtx_unlock(&bt->lock);
device_printf(dev, "bt_fetch_adapter_info - Failed Get Board Info\n");
return (error);
}
bt->firmware_ver[0] = board_id.firmware_rev_major;
bt->firmware_ver[1] = '.';
bt->firmware_ver[2] = board_id.firmware_rev_minor;
bt->firmware_ver[3] = '\0';
/*
* Depending on the firmware major and minor version,
* we may be able to fetch additional minor version info.
*/
if (bt->firmware_ver[0] > '0') {
error = bt_cmd(bt, BOP_INQUIRE_FW_VER_3DIG, NULL, /*parmlen*/0,
(u_int8_t*)&bt->firmware_ver[3], 1,
DEFAULT_CMD_TIMEOUT);
if (error != 0) {
mtx_unlock(&bt->lock);
device_printf(dev,
"bt_fetch_adapter_info - Failed Get "
"Firmware 3rd Digit\n");
return (error);
}
if (bt->firmware_ver[3] == ' ')
bt->firmware_ver[3] = '\0';
bt->firmware_ver[4] = '\0';
}
if (strcmp(bt->firmware_ver, "3.3") >= 0) {
error = bt_cmd(bt, BOP_INQUIRE_FW_VER_4DIG, NULL, /*parmlen*/0,
(u_int8_t*)&bt->firmware_ver[4], 1,
DEFAULT_CMD_TIMEOUT);
if (error != 0) {
mtx_unlock(&bt->lock);
device_printf(dev,
"bt_fetch_adapter_info - Failed Get "
"Firmware 4th Digit\n");
return (error);
}
if (bt->firmware_ver[4] == ' ')
bt->firmware_ver[4] = '\0';
bt->firmware_ver[5] = '\0';
}
/*
* Some boards do not handle the "recently documented"
* Inquire Board Model Number command correctly or do not give
* exact information. Use the Firmware and Extended Setup
* information in these cases to come up with the right answer.
* The major firmware revision number indicates:
*
* 5.xx BusLogic "W" Series Host Adapters:
* BT-948/958/958D
* 4.xx BusLogic "C" Series Host Adapters:
* BT-946C/956C/956CD/747C/757C/757CD/445C/545C/540CF
* 3.xx BusLogic "S" Series Host Adapters:
* BT-747S/747D/757S/757D/445S/545S/542D
* BT-542B/742A (revision H)
* 2.xx BusLogic "A" Series Host Adapters:
* BT-542B/742A (revision G and below)
* 0.xx AMI FastDisk VLB/EISA BusLogic Clone Host Adapter
*/
length_param = sizeof(esetup_info);
error = bt_cmd(bt, BOP_INQUIRE_ESETUP_INFO, &length_param, /*parmlen*/1,
(u_int8_t*)&esetup_info, sizeof(esetup_info),
DEFAULT_CMD_TIMEOUT);
if (error != 0) {
mtx_unlock(&bt->lock);
return (error);
}
bt->bios_addr = esetup_info.bios_addr << 12;
bt->mailbox_addrlimit = BUS_SPACE_MAXADDR;
if (esetup_info.bus_type == 'A'
&& bt->firmware_ver[0] == '2') {
snprintf(bt->model, sizeof(bt->model), "542B");
} else if (esetup_info.bus_type == 'E'
&& bt->firmware_ver[0] == '2') {
/*
* The 742A seems to object if its mailboxes are
* allocated above the 16MB mark.
*/
bt->mailbox_addrlimit = BUS_SPACE_MAXADDR_24BIT;
snprintf(bt->model, sizeof(bt->model), "742A");
} else if (esetup_info.bus_type == 'E'
&& bt->firmware_ver[0] == '0') {
/* AMI FastDisk EISA Series 441 0.x */
snprintf(bt->model, sizeof(bt->model), "747A");
} else {
ha_model_data_t model_data;
int i;
length_param = sizeof(model_data);
error = bt_cmd(bt, BOP_INQUIRE_MODEL, &length_param, 1,
(u_int8_t*)&model_data, sizeof(model_data),
DEFAULT_CMD_TIMEOUT);
if (error != 0) {
mtx_unlock(&bt->lock);
device_printf(dev,
"bt_fetch_adapter_info - Failed Inquire "
"Model Number\n");
return (error);
}
for (i = 0; i < sizeof(model_data.ascii_model); i++) {
bt->model[i] = model_data.ascii_model[i];
if (bt->model[i] == ' ')
break;
}
bt->model[i] = '\0';
}
bt->level_trigger_ints = esetup_info.level_trigger_ints ? 1 : 0;
/* SG element limits */
bt->max_sg = esetup_info.max_sg;
/* Set feature flags */
bt->wide_bus = esetup_info.wide_bus;
bt->diff_bus = esetup_info.diff_bus;
bt->ultra_scsi = esetup_info.ultra_scsi;
if ((bt->firmware_ver[0] == '5')
|| (bt->firmware_ver[0] == '4' && bt->wide_bus))
bt->extended_lun = TRUE;
bt->strict_rr = (strcmp(bt->firmware_ver, "3.31") >= 0);
bt->extended_trans =
((bt_inb(bt, GEOMETRY_REG) & EXTENDED_TRANSLATION) != 0);
/*
* Determine max CCB count and whether tagged queuing is
* available based on controller type. Tagged queuing
* only works on 'W' series adapters, 'C' series adapters
* with firmware of rev 4.42 and higher, and 'S' series
* adapters with firmware of rev 3.35 and higher. The
* maximum CCB counts are as follows:
*
* 192 BT-948/958/958D
* 100 BT-946C/956C/956CD/747C/757C/757CD/445C
* 50 BT-545C/540CF
* 30 BT-747S/747D/757S/757D/445S/545S/542D/542B/742A
*/
if (bt->firmware_ver[0] == '5') {
bt->max_ccbs = 192;
bt->tag_capable = TRUE;
} else if (bt->firmware_ver[0] == '4') {
if (bt->model[0] == '5')
bt->max_ccbs = 50;
else
bt->max_ccbs = 100;
bt->tag_capable = (strcmp(bt->firmware_ver, "4.22") >= 0);
} else {
bt->max_ccbs = 30;
if (bt->firmware_ver[0] == '3'
&& (strcmp(bt->firmware_ver, "3.35") >= 0))
bt->tag_capable = TRUE;
else
bt->tag_capable = FALSE;
}
if (bt->tag_capable != FALSE)
bt->tags_permitted = ALL_TARGETS;
/* Determine Sync/Wide/Disc settings */
if (bt->firmware_ver[0] >= '4') {
auto_scsi_data_t auto_scsi_data;
fetch_lram_params_t fetch_lram_params;
int error;
/*
* These settings are stored in the
* AutoSCSI data in LRAM of 'W' and 'C'
* adapters.
*/
fetch_lram_params.offset = AUTO_SCSI_BYTE_OFFSET;
fetch_lram_params.response_len = sizeof(auto_scsi_data);
error = bt_cmd(bt, BOP_FETCH_LRAM,
(u_int8_t*)&fetch_lram_params,
sizeof(fetch_lram_params),
(u_int8_t*)&auto_scsi_data,
sizeof(auto_scsi_data), DEFAULT_CMD_TIMEOUT);
if (error != 0) {
mtx_unlock(&bt->lock);
device_printf(dev,
"bt_fetch_adapter_info - Failed "
"Get Auto SCSI Info\n");
return (error);
}
bt->disc_permitted = auto_scsi_data.low_disc_permitted
| (auto_scsi_data.high_disc_permitted << 8);
bt->sync_permitted = auto_scsi_data.low_sync_permitted
| (auto_scsi_data.high_sync_permitted << 8);
bt->fast_permitted = auto_scsi_data.low_fast_permitted
| (auto_scsi_data.high_fast_permitted << 8);
bt->ultra_permitted = auto_scsi_data.low_ultra_permitted
| (auto_scsi_data.high_ultra_permitted << 8);
bt->wide_permitted = auto_scsi_data.low_wide_permitted
| (auto_scsi_data.high_wide_permitted << 8);
if (bt->ultra_scsi == FALSE)
bt->ultra_permitted = 0;
if (bt->wide_bus == FALSE)
bt->wide_permitted = 0;
} else {
/*
* 'S' and 'A' series have this information in the setup
* information structure.
*/
setup_data_t setup_info;
length_param = sizeof(setup_info);
error = bt_cmd(bt, BOP_INQUIRE_SETUP_INFO, &length_param,
/*paramlen*/1, (u_int8_t*)&setup_info,
sizeof(setup_info), DEFAULT_CMD_TIMEOUT);
if (error != 0) {
mtx_unlock(&bt->lock);
device_printf(dev,
"bt_fetch_adapter_info - Failed "
"Get Setup Info\n");
return (error);
}
if (setup_info.initiate_sync != 0) {
bt->sync_permitted = ALL_TARGETS;
if (bt->model[0] == '7') {
if (esetup_info.sync_neg10MB != 0)
bt->fast_permitted = ALL_TARGETS;
if (strcmp(bt->model, "757") == 0)
bt->wide_permitted = ALL_TARGETS;
}
}
bt->disc_permitted = ALL_TARGETS;
}
/* We need as many mailboxes as we can have ccbs */
bt->num_boxes = bt->max_ccbs;
/* Determine our SCSI ID */
error = bt_cmd(bt, BOP_INQUIRE_CONFIG, NULL, /*parmlen*/0,
(u_int8_t*)&config_data, sizeof(config_data),
DEFAULT_CMD_TIMEOUT);
mtx_unlock(&bt->lock);
if (error != 0) {
device_printf(dev,
"bt_fetch_adapter_info - Failed Get Config\n");
return (error);
}
bt->scsi_id = config_data.scsi_id;
return (0);
}
/*
* Start the board, ready for normal operation
*/
int
bt_init(device_t dev)
{
struct bt_softc *bt = device_get_softc(dev);
/* Announce the Adapter */
device_printf(dev, "BT-%s FW Rev. %s ", bt->model, bt->firmware_ver);
if (bt->ultra_scsi != 0)
printf("Ultra ");
if (bt->wide_bus != 0)
printf("Wide ");
else
printf("Narrow ");
if (bt->diff_bus != 0)
printf("Diff ");
printf("SCSI Host Adapter, SCSI ID %d, %d CCBs\n", bt->scsi_id,
bt->max_ccbs);
/*
* Create our DMA tags. These tags define the kinds of device
* accessible memory allocations and memory mappings we will
* need to perform during normal operation.
*
* Unless we need to further restrict the allocation, we rely
* on the restrictions of the parent dmat, hence the common
* use of MAXADDR and MAXSIZE.
*/
/* DMA tag for mapping buffers into device visible space. */
if (bus_dma_tag_create( /* parent */ bt->parent_dmat,
/* alignment */ 1,
/* boundary */ 0,
/* lowaddr */ BUS_SPACE_MAXADDR,
/* highaddr */ BUS_SPACE_MAXADDR,
/* filter */ NULL,
/* filterarg */ NULL,
/* maxsize */ MAXBSIZE,
/* nsegments */ BT_NSEG,
/* maxsegsz */ BUS_SPACE_MAXSIZE_32BIT,
/* flags */ BUS_DMA_ALLOCNOW,
/* lockfunc */ busdma_lock_mutex,
/* lockarg */ &bt->lock,
&bt->buffer_dmat) != 0) {
goto error_exit;
}
bt->init_level++;
/* DMA tag for our mailboxes */
if (bus_dma_tag_create( /* parent */ bt->parent_dmat,
/* alignment */ 1,
/* boundary */ 0,
/* lowaddr */ bt->mailbox_addrlimit,
/* highaddr */ BUS_SPACE_MAXADDR,
/* filter */ NULL,
/* filterarg */ NULL,
/* maxsize */ bt->num_boxes *
(sizeof(bt_mbox_in_t) +
sizeof(bt_mbox_out_t)),
/* nsegments */ 1,
/* maxsegsz */ BUS_SPACE_MAXSIZE_32BIT,
/* flags */ 0,
/* lockfunc */ NULL,
/* lockarg */ NULL,
&bt->mailbox_dmat) != 0) {
goto error_exit;
}
bt->init_level++;
/* Allocation for our mailboxes */
if (bus_dmamem_alloc(bt->mailbox_dmat, (void **)&bt->out_boxes,
BUS_DMA_NOWAIT, &bt->mailbox_dmamap) != 0) {
goto error_exit;
}
bt->init_level++;
/* And permanently map them */
bus_dmamap_load(bt->mailbox_dmat, bt->mailbox_dmamap,
bt->out_boxes,
bt->num_boxes * (sizeof(bt_mbox_in_t)
+ sizeof(bt_mbox_out_t)),
btmapmboxes, bt, /*flags*/0);
bt->init_level++;
bt->in_boxes = (bt_mbox_in_t *)&bt->out_boxes[bt->num_boxes];
mtx_lock(&bt->lock);
btinitmboxes(bt);
mtx_unlock(&bt->lock);
/* DMA tag for our ccb structures */
if (bus_dma_tag_create( /* parent */ bt->parent_dmat,
/* alignment */ 1,
/* boundary */ 0,
/* lowaddr */ BUS_SPACE_MAXADDR,
/* highaddr */ BUS_SPACE_MAXADDR,
/* filter */ NULL,
/* filterarg */ NULL,
/* maxsize */ bt->max_ccbs *
sizeof(struct bt_ccb),
/* nsegments */ 1,
/* maxsegsz */ BUS_SPACE_MAXSIZE_32BIT,
/* flags */ 0,
/* lockfunc */ NULL,
/* lockarg */ NULL,
&bt->ccb_dmat) != 0) {
goto error_exit;
}
bt->init_level++;
/* Allocation for our ccbs */
if (bus_dmamem_alloc(bt->ccb_dmat, (void **)&bt->bt_ccb_array,
BUS_DMA_NOWAIT, &bt->ccb_dmamap) != 0) {
goto error_exit;
}
bt->init_level++;
/* And permanently map them */
bus_dmamap_load(bt->ccb_dmat, bt->ccb_dmamap,
bt->bt_ccb_array,
bt->max_ccbs * sizeof(struct bt_ccb),
btmapccbs, bt, /*flags*/0);
bt->init_level++;
/* DMA tag for our S/G structures. We allocate in page sized chunks */
if (bus_dma_tag_create( /* parent */ bt->parent_dmat,
/* alignment */ 1,
/* boundary */ 0,
/* lowaddr */ BUS_SPACE_MAXADDR,
/* highaddr */ BUS_SPACE_MAXADDR,
/* filter */ NULL,
/* filterarg */ NULL,
/* maxsize */ PAGE_SIZE,
/* nsegments */ 1,
/* maxsegsz */ BUS_SPACE_MAXSIZE_32BIT,
/* flags */ 0,
/* lockfunc */ NULL,
/* lockarg */ NULL,
&bt->sg_dmat) != 0) {
goto error_exit;
}
bt->init_level++;
/* Perform initial CCB allocation */
bzero(bt->bt_ccb_array, bt->max_ccbs * sizeof(struct bt_ccb));
btallocccbs(bt);
if (bt->num_ccbs == 0) {
device_printf(dev,
"bt_init - Unable to allocate initial ccbs\n");
goto error_exit;
}
/*
* Note that we are going and return (to attach)
*/
return 0;
error_exit:
return (ENXIO);
}
int
bt_attach(device_t dev)
{
struct bt_softc *bt = device_get_softc(dev);
int tagged_dev_openings;
struct cam_devq *devq;
int error;
/*
* We reserve 1 ccb for error recovery, so don't
* tell the XPT about it.
*/
if (bt->tag_capable != 0)
tagged_dev_openings = bt->max_ccbs - 1;
else
tagged_dev_openings = 0;
/*
* Create the device queue for our SIM.
*/
devq = cam_simq_alloc(bt->max_ccbs - 1);
if (devq == NULL)
return (ENOMEM);
/*
* Construct our SIM entry
*/
bt->sim = cam_sim_alloc(btaction, btpoll, "bt", bt,
device_get_unit(bt->dev), &bt->lock, 2, tagged_dev_openings, devq);
if (bt->sim == NULL) {
cam_simq_free(devq);
return (ENOMEM);
}
mtx_lock(&bt->lock);
if (xpt_bus_register(bt->sim, dev, 0) != CAM_SUCCESS) {
cam_sim_free(bt->sim, /*free_devq*/TRUE);
mtx_unlock(&bt->lock);
return (ENXIO);
}
if (xpt_create_path(&bt->path, /*periph*/NULL,
cam_sim_path(bt->sim), CAM_TARGET_WILDCARD,
CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
xpt_bus_deregister(cam_sim_path(bt->sim));
cam_sim_free(bt->sim, /*free_devq*/TRUE);
mtx_unlock(&bt->lock);
return (ENXIO);
}
mtx_unlock(&bt->lock);
/*
* Setup interrupt.
*/
error = bus_setup_intr(dev, bt->irq, INTR_TYPE_CAM | INTR_ENTROPY |
INTR_MPSAFE, NULL, bt_intr, bt, &bt->ih);
if (error) {
device_printf(dev, "bus_setup_intr() failed: %d\n", error);
return (error);
}
return (0);
}
int
bt_check_probed_iop(u_int ioport)
{
u_int i;
for (i = 0; i < BT_NUM_ISAPORTS; i++) {
if (bt_isa_ports[i].addr == ioport) {
if (bt_isa_ports[i].probed != 0)
return (1);
else {
return (0);
}
}
}
return (1);
}
void
bt_mark_probed_bio(isa_compat_io_t port)
{
if (port < BIO_DISABLED)
bt_mark_probed_iop(bt_board_ports[port]);
}
void
bt_mark_probed_iop(u_int ioport)
{
u_int i;
for (i = 0; i < BT_NUM_ISAPORTS; i++) {
if (ioport == bt_isa_ports[i].addr) {
bt_isa_ports[i].probed = 1;
break;
}
}
}
void
bt_find_probe_range(int ioport, int *port_index, int *max_port_index)
{
if (ioport > 0) {
int i;
for (i = 0;i < BT_NUM_ISAPORTS; i++)
if (ioport <= bt_isa_ports[i].addr)
break;
if ((i >= BT_NUM_ISAPORTS)
|| (ioport != bt_isa_ports[i].addr)) {
printf(
"bt_find_probe_range: Invalid baseport of 0x%x specified.\n"
"bt_find_probe_range: Nearest valid baseport is 0x%x.\n"
"bt_find_probe_range: Failing probe.\n",
ioport,
(i < BT_NUM_ISAPORTS)
? bt_isa_ports[i].addr
: bt_isa_ports[BT_NUM_ISAPORTS - 1].addr);
*port_index = *max_port_index = -1;
return;
}
*port_index = *max_port_index = bt_isa_ports[i].bio;
} else {
*port_index = 0;
*max_port_index = BT_NUM_ISAPORTS - 1;
}
}
int
bt_iop_from_bio(isa_compat_io_t bio_index)
{
if (bio_index < BT_NUM_ISAPORTS)
return (bt_board_ports[bio_index]);
return (-1);
}
static void
btallocccbs(struct bt_softc *bt)
{
struct bt_ccb *next_ccb;
struct sg_map_node *sg_map;
bus_addr_t physaddr;
bt_sg_t *segs;
int newcount;
int i;
if (bt->num_ccbs >= bt->max_ccbs)
/* Can't allocate any more */
return;
next_ccb = &bt->bt_ccb_array[bt->num_ccbs];
sg_map = malloc(sizeof(*sg_map), M_DEVBUF, M_NOWAIT);
if (sg_map == NULL)
goto error_exit;
/* Allocate S/G space for the next batch of CCBS */
if (bus_dmamem_alloc(bt->sg_dmat, (void **)&sg_map->sg_vaddr,
BUS_DMA_NOWAIT, &sg_map->sg_dmamap) != 0) {
free(sg_map, M_DEVBUF);
goto error_exit;
}
SLIST_INSERT_HEAD(&bt->sg_maps, sg_map, links);
bus_dmamap_load(bt->sg_dmat, sg_map->sg_dmamap, sg_map->sg_vaddr,
PAGE_SIZE, btmapsgs, bt, /*flags*/0);
segs = sg_map->sg_vaddr;
physaddr = sg_map->sg_physaddr;
newcount = (PAGE_SIZE / (BT_NSEG * sizeof(bt_sg_t)));
for (i = 0; bt->num_ccbs < bt->max_ccbs && i < newcount; i++) {
int error;
next_ccb->sg_list = segs;
next_ccb->sg_list_phys = physaddr;
next_ccb->flags = BCCB_FREE;
callout_init_mtx(&next_ccb->timer, &bt->lock, 0);
error = bus_dmamap_create(bt->buffer_dmat, /*flags*/0,
&next_ccb->dmamap);
if (error != 0)
break;
SLIST_INSERT_HEAD(&bt->free_bt_ccbs, next_ccb, links);
segs += BT_NSEG;
physaddr += (BT_NSEG * sizeof(bt_sg_t));
next_ccb++;
bt->num_ccbs++;
}
/* Reserve a CCB for error recovery */
if (bt->recovery_bccb == NULL) {
bt->recovery_bccb = SLIST_FIRST(&bt->free_bt_ccbs);
SLIST_REMOVE_HEAD(&bt->free_bt_ccbs, links);
}
if (SLIST_FIRST(&bt->free_bt_ccbs) != NULL)
return;
error_exit:
device_printf(bt->dev, "Can't malloc BCCBs\n");
}
static __inline void
btfreeccb(struct bt_softc *bt, struct bt_ccb *bccb)
{
if (!dumping)
mtx_assert(&bt->lock, MA_OWNED);
if ((bccb->flags & BCCB_ACTIVE) != 0)
LIST_REMOVE(&bccb->ccb->ccb_h, sim_links.le);
if (bt->resource_shortage != 0
&& (bccb->ccb->ccb_h.status & CAM_RELEASE_SIMQ) == 0) {
bccb->ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
bt->resource_shortage = FALSE;
}
bccb->flags = BCCB_FREE;
SLIST_INSERT_HEAD(&bt->free_bt_ccbs, bccb, links);
bt->active_ccbs--;
}
static __inline struct bt_ccb*
btgetccb(struct bt_softc *bt)
{
struct bt_ccb* bccb;
if (!dumping)
mtx_assert(&bt->lock, MA_OWNED);
if ((bccb = SLIST_FIRST(&bt->free_bt_ccbs)) != NULL) {
SLIST_REMOVE_HEAD(&bt->free_bt_ccbs, links);
bt->active_ccbs++;
} else {
btallocccbs(bt);
bccb = SLIST_FIRST(&bt->free_bt_ccbs);
if (bccb != NULL) {
SLIST_REMOVE_HEAD(&bt->free_bt_ccbs, links);
bt->active_ccbs++;
}
}
return (bccb);
}
static void
btaction(struct cam_sim *sim, union ccb *ccb)
{
struct bt_softc *bt;
CAM_DEBUG(ccb->ccb_h.path, CAM_DEBUG_TRACE, ("btaction\n"));
bt = (struct bt_softc *)cam_sim_softc(sim);
mtx_assert(&bt->lock, MA_OWNED);
switch (ccb->ccb_h.func_code) {
/* Common cases first */
case XPT_SCSI_IO: /* Execute the requested I/O operation */
case XPT_RESET_DEV: /* Bus Device Reset the specified SCSI device */
{
struct bt_ccb *bccb;
struct bt_hccb *hccb;
/*
* get a bccb to use.
*/
if ((bccb = btgetccb(bt)) == NULL) {
bt->resource_shortage = TRUE;
xpt_freeze_simq(bt->sim, /*count*/1);
ccb->ccb_h.status = CAM_REQUEUE_REQ;
xpt_done(ccb);
return;
}
hccb = &bccb->hccb;
/*
* So we can find the BCCB when an abort is requested
*/
bccb->ccb = ccb;
ccb->ccb_h.ccb_bccb_ptr = bccb;
ccb->ccb_h.ccb_bt_ptr = bt;
/*
* Put all the arguments for the xfer in the bccb
*/
hccb->target_id = ccb->ccb_h.target_id;
hccb->target_lun = ccb->ccb_h.target_lun;
hccb->btstat = 0;
hccb->sdstat = 0;
if (ccb->ccb_h.func_code == XPT_SCSI_IO) {
struct ccb_scsiio *csio;
struct ccb_hdr *ccbh;
int error;
csio = &ccb->csio;
ccbh = &csio->ccb_h;
hccb->opcode = INITIATOR_CCB_WRESID;
hccb->datain = (ccb->ccb_h.flags & CAM_DIR_IN) ? 1 : 0;
hccb->dataout =(ccb->ccb_h.flags & CAM_DIR_OUT) ? 1 : 0;
hccb->cmd_len = csio->cdb_len;
if (hccb->cmd_len > sizeof(hccb->scsi_cdb)) {
ccb->ccb_h.status = CAM_REQ_INVALID;
btfreeccb(bt, bccb);
xpt_done(ccb);
return;
}
hccb->sense_len = csio->sense_len;
if ((ccbh->flags & CAM_TAG_ACTION_VALID) != 0
&& ccb->csio.tag_action != CAM_TAG_ACTION_NONE) {
hccb->tag_enable = TRUE;
hccb->tag_type = (ccb->csio.tag_action & 0x3);
} else {
hccb->tag_enable = FALSE;
hccb->tag_type = 0;
}
if ((ccbh->flags & CAM_CDB_POINTER) != 0) {
if ((ccbh->flags & CAM_CDB_PHYS) == 0) {
bcopy(csio->cdb_io.cdb_ptr,
hccb->scsi_cdb, hccb->cmd_len);
} else {
/* I guess I could map it in... */
ccbh->status = CAM_REQ_INVALID;
btfreeccb(bt, bccb);
xpt_done(ccb);
return;
}
} else {
bcopy(csio->cdb_io.cdb_bytes,
hccb->scsi_cdb, hccb->cmd_len);
}
/* If need be, bounce our sense buffer */
if (bt->sense_buffers != NULL) {
hccb->sense_addr = btsensepaddr(bt, bccb);
} else {
hccb->sense_addr = vtophys(&csio->sense_data);
}
/*
* If we have any data to send with this command,
* map it into bus space.
*/
error = bus_dmamap_load_ccb(
bt->buffer_dmat,
bccb->dmamap,
ccb,
btexecuteccb,
bccb,
/*flags*/0);
if (error == EINPROGRESS) {
/*
* So as to maintain ordering, freeze the
* controller queue until our mapping is
* returned.
*/
xpt_freeze_simq(bt->sim, 1);
csio->ccb_h.status |= CAM_RELEASE_SIMQ;
}
} else {
hccb->opcode = INITIATOR_BUS_DEV_RESET;
/* No data transfer */
hccb->datain = TRUE;
hccb->dataout = TRUE;
hccb->cmd_len = 0;
hccb->sense_len = 0;
hccb->tag_enable = FALSE;
hccb->tag_type = 0;
btexecuteccb(bccb, NULL, 0, 0);
}
break;
}
case XPT_EN_LUN: /* Enable LUN as a target */
case XPT_TARGET_IO: /* Execute target I/O request */
case XPT_ACCEPT_TARGET_IO: /* Accept Host Target Mode CDB */
case XPT_CONT_TARGET_IO: /* Continue Host Target I/O Connection*/
case XPT_ABORT: /* Abort the specified CCB */
/* XXX Implement */
ccb->ccb_h.status = CAM_REQ_INVALID;
xpt_done(ccb);
break;
case XPT_SET_TRAN_SETTINGS:
{
/* XXX Implement */
ccb->ccb_h.status = CAM_PROVIDE_FAIL;
xpt_done(ccb);
break;
}
case XPT_GET_TRAN_SETTINGS:
/* Get default/user set transfer settings for the target */
{
struct ccb_trans_settings *cts;
u_int target_mask;
cts = &ccb->cts;
target_mask = 0x01 << ccb->ccb_h.target_id;
if (cts->type == CTS_TYPE_CURRENT_SETTINGS) {
struct ccb_trans_settings_scsi *scsi =
&cts->proto_specific.scsi;
struct ccb_trans_settings_spi *spi =
&cts->xport_specific.spi;
cts->protocol = PROTO_SCSI;
cts->protocol_version = SCSI_REV_2;
cts->transport = XPORT_SPI;
cts->transport_version = 2;
scsi->flags &= ~CTS_SCSI_FLAGS_TAG_ENB;
spi->flags &= ~CTS_SPI_FLAGS_DISC_ENB;
if ((bt->disc_permitted & target_mask) != 0)
spi->flags |= CTS_SPI_FLAGS_DISC_ENB;
if ((bt->tags_permitted & target_mask) != 0)
scsi->flags |= CTS_SCSI_FLAGS_TAG_ENB;
if ((bt->ultra_permitted & target_mask) != 0)
spi->sync_period = 12;
else if ((bt->fast_permitted & target_mask) != 0)
spi->sync_period = 25;
else if ((bt->sync_permitted & target_mask) != 0)
spi->sync_period = 50;
else
spi->sync_period = 0;
if (spi->sync_period != 0)
spi->sync_offset = 15;
spi->valid |= CTS_SPI_VALID_SYNC_RATE;
spi->valid |= CTS_SPI_VALID_SYNC_OFFSET;
spi->valid |= CTS_SPI_VALID_BUS_WIDTH;
if ((bt->wide_permitted & target_mask) != 0)
spi->bus_width = MSG_EXT_WDTR_BUS_16_BIT;
else
spi->bus_width = MSG_EXT_WDTR_BUS_8_BIT;
if (cts->ccb_h.target_lun != CAM_LUN_WILDCARD) {
scsi->valid = CTS_SCSI_VALID_TQ;
spi->valid |= CTS_SPI_VALID_DISC;
} else
scsi->valid = 0;
} else {
btfetchtransinfo(bt, cts);
}
ccb->ccb_h.status = CAM_REQ_CMP;
xpt_done(ccb);
break;
}
case XPT_CALC_GEOMETRY:
{
struct ccb_calc_geometry *ccg;
u_int32_t size_mb;
u_int32_t secs_per_cylinder;
ccg = &ccb->ccg;
size_mb = ccg->volume_size
/ ((1024L * 1024L) / ccg->block_size);
if (size_mb >= 1024 && (bt->extended_trans != 0)) {
if (size_mb >= 2048) {
ccg->heads = 255;
ccg->secs_per_track = 63;
} else {
ccg->heads = 128;
ccg->secs_per_track = 32;
}
} else {
ccg->heads = 64;
ccg->secs_per_track = 32;
}
secs_per_cylinder = ccg->heads * ccg->secs_per_track;
ccg->cylinders = ccg->volume_size / secs_per_cylinder;
ccb->ccb_h.status = CAM_REQ_CMP;
xpt_done(ccb);
break;
}
case XPT_RESET_BUS: /* Reset the specified SCSI bus */
{
btreset(bt, /*hardreset*/TRUE);
ccb->ccb_h.status = CAM_REQ_CMP;
xpt_done(ccb);
break;
}
case XPT_TERM_IO: /* Terminate the I/O process */
/* XXX Implement */
ccb->ccb_h.status = CAM_REQ_INVALID;
xpt_done(ccb);
break;
case XPT_PATH_INQ: /* Path routing inquiry */
{
struct ccb_pathinq *cpi = &ccb->cpi;
cpi->version_num = 1; /* XXX??? */
cpi->hba_inquiry = PI_SDTR_ABLE;
if (bt->tag_capable != 0)
cpi->hba_inquiry |= PI_TAG_ABLE;
if (bt->wide_bus != 0)
cpi->hba_inquiry |= PI_WIDE_16;
cpi->target_sprt = 0;
cpi->hba_misc = 0;
cpi->hba_eng_cnt = 0;
cpi->max_target = bt->wide_bus ? 15 : 7;
cpi->max_lun = 7;
cpi->initiator_id = bt->scsi_id;
cpi->bus_id = cam_sim_bus(sim);
cpi->base_transfer_speed = 3300;
strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
strncpy(cpi->hba_vid, "BusLogic", HBA_IDLEN);
strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
cpi->unit_number = cam_sim_unit(sim);
cpi->ccb_h.status = CAM_REQ_CMP;
cpi->transport = XPORT_SPI;
cpi->transport_version = 2;
cpi->protocol = PROTO_SCSI;
cpi->protocol_version = SCSI_REV_2;
xpt_done(ccb);
break;
}
default:
ccb->ccb_h.status = CAM_REQ_INVALID;
xpt_done(ccb);
break;
}
}
static void
btexecuteccb(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error)
{
struct bt_ccb *bccb;
union ccb *ccb;
struct bt_softc *bt;
bccb = (struct bt_ccb *)arg;
ccb = bccb->ccb;
bt = (struct bt_softc *)ccb->ccb_h.ccb_bt_ptr;
if (error != 0) {
if (error != EFBIG)
device_printf(bt->dev,
"Unexepected error 0x%x returned from "
"bus_dmamap_load\n", error);
if (ccb->ccb_h.status == CAM_REQ_INPROG) {
xpt_freeze_devq(ccb->ccb_h.path, /*count*/1);
ccb->ccb_h.status = CAM_REQ_TOO_BIG|CAM_DEV_QFRZN;
}
btfreeccb(bt, bccb);
xpt_done(ccb);
return;
}
if (nseg != 0) {
bt_sg_t *sg;
bus_dma_segment_t *end_seg;
bus_dmasync_op_t op;
end_seg = dm_segs + nseg;
/* Copy the segments into our SG list */
sg = bccb->sg_list;
while (dm_segs < end_seg) {
sg->len = dm_segs->ds_len;
sg->addr = dm_segs->ds_addr;
sg++;
dm_segs++;
}
if (nseg > 1) {
bccb->hccb.opcode = INITIATOR_SG_CCB_WRESID;
bccb->hccb.data_len = sizeof(bt_sg_t) * nseg;
bccb->hccb.data_addr = bccb->sg_list_phys;
} else {
bccb->hccb.data_len = bccb->sg_list->len;
bccb->hccb.data_addr = bccb->sg_list->addr;
}
if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN)
op = BUS_DMASYNC_PREREAD;
else
op = BUS_DMASYNC_PREWRITE;
bus_dmamap_sync(bt->buffer_dmat, bccb->dmamap, op);
} else {
bccb->hccb.opcode = INITIATOR_CCB;
bccb->hccb.data_len = 0;
bccb->hccb.data_addr = 0;
}
/*
* Last time we need to check if this CCB needs to
* be aborted.
*/
if (ccb->ccb_h.status != CAM_REQ_INPROG) {
if (nseg != 0)
bus_dmamap_unload(bt->buffer_dmat, bccb->dmamap);
btfreeccb(bt, bccb);
xpt_done(ccb);
return;
}
bccb->flags = BCCB_ACTIVE;
ccb->ccb_h.status |= CAM_SIM_QUEUED;
LIST_INSERT_HEAD(&bt->pending_ccbs, &ccb->ccb_h, sim_links.le);
callout_reset_sbt(&bccb->timer, SBT_1MS * ccb->ccb_h.timeout, 0,
bttimeout, bccb, 0);
/* Tell the adapter about this command */
bt->cur_outbox->ccb_addr = btccbvtop(bt, bccb);
if (bt->cur_outbox->action_code != BMBO_FREE) {
/*
* We should never encounter a busy mailbox.
* If we do, warn the user, and treat it as
* a resource shortage. If the controller is
* hung, one of the pending transactions will
* timeout causing us to start recovery operations.
*/
device_printf(bt->dev,
"Encountered busy mailbox with %d out of %d "
"commands active!!!\n", bt->active_ccbs,
bt->max_ccbs);
callout_stop(&bccb->timer);
if (nseg != 0)
bus_dmamap_unload(bt->buffer_dmat, bccb->dmamap);
btfreeccb(bt, bccb);
bt->resource_shortage = TRUE;
xpt_freeze_simq(bt->sim, /*count*/1);
ccb->ccb_h.status = CAM_REQUEUE_REQ;
xpt_done(ccb);
return;
}
bt->cur_outbox->action_code = BMBO_START;
bt_outb(bt, COMMAND_REG, BOP_START_MBOX);
btnextoutbox(bt);
}
void
bt_intr(void *arg)
{
struct bt_softc *bt;
bt = arg;
mtx_lock(&bt->lock);
bt_intr_locked(bt);
mtx_unlock(&bt->lock);
}
void
bt_intr_locked(struct bt_softc *bt)
{
u_int intstat;
while (((intstat = bt_inb(bt, INTSTAT_REG)) & INTR_PENDING) != 0) {
if ((intstat & CMD_COMPLETE) != 0) {
bt->latched_status = bt_inb(bt, STATUS_REG);
bt->command_cmp = TRUE;
}
bt_outb(bt, CONTROL_REG, RESET_INTR);
if ((intstat & IMB_LOADED) != 0) {
while (bt->cur_inbox->comp_code != BMBI_FREE) {
btdone(bt,
btccbptov(bt, bt->cur_inbox->ccb_addr),
bt->cur_inbox->comp_code);
bt->cur_inbox->comp_code = BMBI_FREE;
btnextinbox(bt);
}
}
if ((intstat & SCSI_BUS_RESET) != 0) {
btreset(bt, /*hardreset*/FALSE);
}
}
}
static void
btdone(struct bt_softc *bt, struct bt_ccb *bccb, bt_mbi_comp_code_t comp_code)
{
union ccb *ccb;
struct ccb_scsiio *csio;
ccb = bccb->ccb;
csio = &bccb->ccb->csio;
if ((bccb->flags & BCCB_ACTIVE) == 0) {
device_printf(bt->dev,
"btdone - Attempt to free non-active BCCB %p\n",
(void *)bccb);
return;
}
if ((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE) {
bus_dmasync_op_t op;
if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN)
op = BUS_DMASYNC_POSTREAD;
else
op = BUS_DMASYNC_POSTWRITE;
bus_dmamap_sync(bt->buffer_dmat, bccb->dmamap, op);
bus_dmamap_unload(bt->buffer_dmat, bccb->dmamap);
}
if (bccb == bt->recovery_bccb) {
/*
* The recovery BCCB does not have a CCB associated
* with it, so short circuit the normal error handling.
* We now traverse our list of pending CCBs and process
* any that were terminated by the recovery CCBs action.
* We also reinstate timeouts for all remaining, pending,
* CCBs.
*/
struct cam_path *path;
struct ccb_hdr *ccb_h;
cam_status error;
/* Notify all clients that a BDR occured */
error = xpt_create_path(&path, /*periph*/NULL,
cam_sim_path(bt->sim),
bccb->hccb.target_id,
CAM_LUN_WILDCARD);
if (error == CAM_REQ_CMP) {
xpt_async(AC_SENT_BDR, path, NULL);
xpt_free_path(path);
}
ccb_h = LIST_FIRST(&bt->pending_ccbs);
while (ccb_h != NULL) {
struct bt_ccb *pending_bccb;
pending_bccb = (struct bt_ccb *)ccb_h->ccb_bccb_ptr;
if (pending_bccb->hccb.target_id
== bccb->hccb.target_id) {
pending_bccb->hccb.btstat = BTSTAT_HA_BDR;
ccb_h = LIST_NEXT(ccb_h, sim_links.le);
btdone(bt, pending_bccb, BMBI_ERROR);
} else {
callout_reset_sbt(&pending_bccb->timer,
SBT_1MS * ccb_h->timeout, 0, bttimeout,
pending_bccb, 0);
ccb_h = LIST_NEXT(ccb_h, sim_links.le);
}
}
device_printf(bt->dev, "No longer in timeout\n");
return;
}
callout_stop(&bccb->timer);
switch (comp_code) {
case BMBI_FREE:
device_printf(bt->dev,
"btdone - CCB completed with free status!\n");
break;
case BMBI_NOT_FOUND:
device_printf(bt->dev,
"btdone - CCB Abort failed to find CCB\n");
break;
case BMBI_ABORT:
case BMBI_ERROR:
if (bootverbose) {
printf("bt: ccb %p - error %x occured. "
"btstat = %x, sdstat = %x\n",
(void *)bccb, comp_code, bccb->hccb.btstat,
bccb->hccb.sdstat);
}
/* An error occured */
switch(bccb->hccb.btstat) {
case BTSTAT_DATARUN_ERROR:
if (bccb->hccb.data_len == 0) {
/*
* At least firmware 4.22, does this
* for a QUEUE FULL condition.
*/
bccb->hccb.sdstat = SCSI_STATUS_QUEUE_FULL;
} else if (bccb->hccb.data_len < 0) {
csio->ccb_h.status = CAM_DATA_RUN_ERR;
break;
}
/* FALLTHROUGH */
case BTSTAT_NOERROR:
case BTSTAT_LINKED_CMD_COMPLETE:
case BTSTAT_LINKED_CMD_FLAG_COMPLETE:
case BTSTAT_DATAUNDERUN_ERROR:
csio->scsi_status = bccb->hccb.sdstat;
csio->ccb_h.status |= CAM_SCSI_STATUS_ERROR;
switch(csio->scsi_status) {
case SCSI_STATUS_CHECK_COND:
case SCSI_STATUS_CMD_TERMINATED:
csio->ccb_h.status |= CAM_AUTOSNS_VALID;
/* Bounce sense back if necessary */
if (bt->sense_buffers != NULL) {
csio->sense_data =
*btsensevaddr(bt, bccb);
}
break;
default:
break;
case SCSI_STATUS_OK:
csio->ccb_h.status = CAM_REQ_CMP;
break;
}
csio->resid = bccb->hccb.data_len;
break;
case BTSTAT_SELTIMEOUT:
csio->ccb_h.status = CAM_SEL_TIMEOUT;
break;
case BTSTAT_UNEXPECTED_BUSFREE:
csio->ccb_h.status = CAM_UNEXP_BUSFREE;
break;
case BTSTAT_INVALID_PHASE:
csio->ccb_h.status = CAM_SEQUENCE_FAIL;
break;
case BTSTAT_INVALID_ACTION_CODE:
panic("%s: Inavlid Action code", bt_name(bt));
break;
case BTSTAT_INVALID_OPCODE:
panic("%s: Inavlid CCB Opcode code", bt_name(bt));
break;
case BTSTAT_LINKED_CCB_LUN_MISMATCH:
/* We don't even support linked commands... */
panic("%s: Linked CCB Lun Mismatch", bt_name(bt));
break;
case BTSTAT_INVALID_CCB_OR_SG_PARAM:
panic("%s: Invalid CCB or SG list", bt_name(bt));
break;
case BTSTAT_AUTOSENSE_FAILED:
csio->ccb_h.status = CAM_AUTOSENSE_FAIL;
break;
case BTSTAT_TAGGED_MSG_REJECTED:
{
struct ccb_trans_settings neg;
struct ccb_trans_settings_scsi *scsi =
&neg.proto_specific.scsi;
neg.protocol = PROTO_SCSI;
neg.protocol_version = SCSI_REV_2;
neg.transport = XPORT_SPI;
neg.transport_version = 2;
scsi->valid = CTS_SCSI_VALID_TQ;
scsi->flags = 0;
xpt_print_path(csio->ccb_h.path);
printf("refuses tagged commands. Performing "
"non-tagged I/O\n");
xpt_setup_ccb(&neg.ccb_h, csio->ccb_h.path,
/*priority*/1);
xpt_async(AC_TRANSFER_NEG, csio->ccb_h.path, &neg);
bt->tags_permitted &= ~(0x01 << csio->ccb_h.target_id);
csio->ccb_h.status = CAM_MSG_REJECT_REC;
break;
}
case BTSTAT_UNSUPPORTED_MSG_RECEIVED:
/*
* XXX You would think that this is
* a recoverable error... Hmmm.
*/
csio->ccb_h.status = CAM_REQ_CMP_ERR;
break;
case BTSTAT_HA_SOFTWARE_ERROR:
case BTSTAT_HA_WATCHDOG_ERROR:
case BTSTAT_HARDWARE_FAILURE:
/* Hardware reset ??? Can we recover ??? */
csio->ccb_h.status = CAM_NO_HBA;
break;
case BTSTAT_TARGET_IGNORED_ATN:
case BTSTAT_OTHER_SCSI_BUS_RESET:
case BTSTAT_HA_SCSI_BUS_RESET:
if ((csio->ccb_h.status & CAM_STATUS_MASK)
!= CAM_CMD_TIMEOUT)
csio->ccb_h.status = CAM_SCSI_BUS_RESET;
break;
case BTSTAT_HA_BDR:
if ((bccb->flags & BCCB_DEVICE_RESET) == 0)
csio->ccb_h.status = CAM_BDR_SENT;
else
csio->ccb_h.status = CAM_CMD_TIMEOUT;
break;
case BTSTAT_INVALID_RECONNECT:
case BTSTAT_ABORT_QUEUE_GENERATED:
csio->ccb_h.status = CAM_REQ_TERMIO;
break;
case BTSTAT_SCSI_PERROR_DETECTED:
csio->ccb_h.status = CAM_UNCOR_PARITY;
break;
}
if (csio->ccb_h.status != CAM_REQ_CMP) {
xpt_freeze_devq(csio->ccb_h.path, /*count*/1);
csio->ccb_h.status |= CAM_DEV_QFRZN;
}
if ((bccb->flags & BCCB_RELEASE_SIMQ) != 0)
ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
btfreeccb(bt, bccb);
xpt_done(ccb);
break;
case BMBI_OK:
/* All completed without incident */
ccb->ccb_h.status |= CAM_REQ_CMP;
if ((bccb->flags & BCCB_RELEASE_SIMQ) != 0)
ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
btfreeccb(bt, bccb);
xpt_done(ccb);
break;
}
}
static int
btreset(struct bt_softc* bt, int hard_reset)
{
struct ccb_hdr *ccb_h;
u_int status;
u_int timeout;
u_int8_t reset_type;
if (hard_reset != 0)
reset_type = HARD_RESET;
else
reset_type = SOFT_RESET;
bt_outb(bt, CONTROL_REG, reset_type);
/* Wait 5sec. for Diagnostic start */
timeout = 5 * 10000;
while (--timeout) {
status = bt_inb(bt, STATUS_REG);
if ((status & DIAG_ACTIVE) != 0)
break;
DELAY(100);
}
if (timeout == 0) {
if (bootverbose)
device_printf(bt->dev,
"btreset - Diagnostic Active failed to "
"assert. status = 0x%x\n", status);
return (ETIMEDOUT);
}
/* Wait 10sec. for Diagnostic end */
timeout = 10 * 10000;
while (--timeout) {
status = bt_inb(bt, STATUS_REG);
if ((status & DIAG_ACTIVE) == 0)
break;
DELAY(100);
}
if (timeout == 0) {
panic("%s: btreset - Diagnostic Active failed to drop. "
"status = 0x%x\n", bt_name(bt), status);
return (ETIMEDOUT);
}
/* Wait for the host adapter to become ready or report a failure */
timeout = 10000;
while (--timeout) {
status = bt_inb(bt, STATUS_REG);
if ((status & (DIAG_FAIL|HA_READY|DATAIN_REG_READY)) != 0)
break;
DELAY(100);
}
if (timeout == 0) {
device_printf(bt->dev,
"btreset - Host adapter failed to come ready. "
"status = 0x%x\n", status);
return (ETIMEDOUT);
}
/* If the diagnostics failed, tell the user */
if ((status & DIAG_FAIL) != 0
|| (status & HA_READY) == 0) {
device_printf(bt->dev,
"btreset - Adapter failed diagnostics\n");
if ((status & DATAIN_REG_READY) != 0)
device_printf(bt->dev,
"btreset - Host Adapter Error code = 0x%x\n",
bt_inb(bt, DATAIN_REG));
return (ENXIO);
}
/* If we've allocated mailboxes, initialize them */
if (bt->init_level > 4)
btinitmboxes(bt);
/* If we've attached to the XPT, tell it about the event */
if (bt->path != NULL)
xpt_async(AC_BUS_RESET, bt->path, NULL);
/*
* Perform completion processing for all outstanding CCBs.
*/
while ((ccb_h = LIST_FIRST(&bt->pending_ccbs)) != NULL) {
struct bt_ccb *pending_bccb;
pending_bccb = (struct bt_ccb *)ccb_h->ccb_bccb_ptr;
pending_bccb->hccb.btstat = BTSTAT_HA_SCSI_BUS_RESET;
btdone(bt, pending_bccb, BMBI_ERROR);
}
return (0);
}
/*
* Send a command to the adapter.
*/
int
bt_cmd(struct bt_softc *bt, bt_op_t opcode, u_int8_t *params, u_int param_len,
u_int8_t *reply_data, u_int reply_len, u_int cmd_timeout)
{
u_int timeout;
u_int status;
u_int saved_status;
u_int intstat;
u_int reply_buf_size;
int cmd_complete;
int error;
/* No data returned to start */
reply_buf_size = reply_len;
reply_len = 0;
intstat = 0;
cmd_complete = 0;
saved_status = 0;
error = 0;
bt->command_cmp = 0;
/*
* Wait up to 10 sec. for the adapter to become
* ready to accept commands.
*/
timeout = 100000;
while (--timeout) {
status = bt_inb(bt, STATUS_REG);
if ((status & HA_READY) != 0
&& (status & CMD_REG_BUSY) == 0)
break;
/*
* Throw away any pending data which may be
* left over from earlier commands that we
* timedout on.
*/
if ((status & DATAIN_REG_READY) != 0)
(void)bt_inb(bt, DATAIN_REG);
DELAY(100);
}
if (timeout == 0) {
device_printf(bt->dev,
"bt_cmd: Timeout waiting for adapter ready, "
"status = 0x%x\n", status);
return (ETIMEDOUT);
}
/*
* Send the opcode followed by any necessary parameter bytes.
*/
bt_outb(bt, COMMAND_REG, opcode);
/*
* Wait for up to 1sec for each byte of the
* parameter list sent to be sent.
*/
timeout = 10000;
while (param_len && --timeout) {
DELAY(100);
status = bt_inb(bt, STATUS_REG);
intstat = bt_inb(bt, INTSTAT_REG);
if ((intstat & (INTR_PENDING|CMD_COMPLETE))
== (INTR_PENDING|CMD_COMPLETE)) {
saved_status = status;
cmd_complete = 1;
break;
}
if (bt->command_cmp != 0) {
saved_status = bt->latched_status;
cmd_complete = 1;
break;
}
if ((status & DATAIN_REG_READY) != 0)
break;
if ((status & CMD_REG_BUSY) == 0) {
bt_outb(bt, COMMAND_REG, *params++);
param_len--;
timeout = 10000;
}
}
if (timeout == 0) {
device_printf(bt->dev, "bt_cmd: Timeout sending parameters, "
"status = 0x%x\n", status);
cmd_complete = 1;
saved_status = status;
error = ETIMEDOUT;
}
/*
* Wait for the command to complete.
*/
while (cmd_complete == 0 && --cmd_timeout) {
status = bt_inb(bt, STATUS_REG);
intstat = bt_inb(bt, INTSTAT_REG);
/*
* It may be that this command was issued with
* controller interrupts disabled. We'll never
* get to our command if an incoming mailbox
* interrupt is pending, so take care of completed
* mailbox commands by calling our interrupt handler.
*/
if ((intstat & (INTR_PENDING|IMB_LOADED))
== (INTR_PENDING|IMB_LOADED))
bt_intr_locked(bt);
if (bt->command_cmp != 0) {
/*
* Our interrupt handler saw CMD_COMPLETE
* status before we did.
*/
cmd_complete = 1;
saved_status = bt->latched_status;
} else if ((intstat & (INTR_PENDING|CMD_COMPLETE))
== (INTR_PENDING|CMD_COMPLETE)) {
/*
* Our poll (in case interrupts are blocked)
* saw the CMD_COMPLETE interrupt.
*/
cmd_complete = 1;
saved_status = status;
} else if (opcode == BOP_MODIFY_IO_ADDR
&& (status & CMD_REG_BUSY) == 0) {
/*
* The BOP_MODIFY_IO_ADDR does not issue a CMD_COMPLETE,
* but it should update the status register. So, we
* consider this command complete when the CMD_REG_BUSY
* status clears.
*/
saved_status = status;
cmd_complete = 1;
} else if ((status & DATAIN_REG_READY) != 0) {
u_int8_t data;
data = bt_inb(bt, DATAIN_REG);
if (reply_len < reply_buf_size) {
*reply_data++ = data;
} else {
device_printf(bt->dev,
"bt_cmd - Discarded reply data byte "
"for opcode 0x%x\n", opcode);
}
/*
* Reset timeout to ensure at least a second
* between response bytes.
*/
cmd_timeout = MAX(cmd_timeout, 10000);
reply_len++;
} else if ((opcode == BOP_FETCH_LRAM)
&& (status & HA_READY) != 0) {
saved_status = status;
cmd_complete = 1;
}
DELAY(100);
}
if (cmd_timeout == 0) {
device_printf(bt->dev,
"bt_cmd: Timeout waiting for command (%x) "
"to complete.\n", opcode);
device_printf(bt->dev, "status = 0x%x, intstat = 0x%x, "
"rlen %d\n", status, intstat, reply_len);
error = (ETIMEDOUT);
}
/*
* Clear any pending interrupts.
*/
bt_intr_locked(bt);
if (error != 0)
return (error);
/*
* If the command was rejected by the controller, tell the caller.
*/
if ((saved_status & CMD_INVALID) != 0) {
/*
* Some early adapters may not recover properly from
* an invalid command. If it appears that the controller
* has wedged (i.e. status was not cleared by our interrupt
* reset above), perform a soft reset.
*/
if (bootverbose)
device_printf(bt->dev, "Invalid Command 0x%x\n",
opcode);
DELAY(1000);
status = bt_inb(bt, STATUS_REG);
if ((status & (CMD_INVALID|STATUS_REG_RSVD|DATAIN_REG_READY|
CMD_REG_BUSY|DIAG_FAIL|DIAG_ACTIVE)) != 0
|| (status & (HA_READY|INIT_REQUIRED))
!= (HA_READY|INIT_REQUIRED)) {
btreset(bt, /*hard_reset*/FALSE);
}
return (EINVAL);
}
if (param_len > 0) {
/* The controller did not accept the full argument list */
return (E2BIG);
}
if (reply_len != reply_buf_size) {
/* Too much or too little data received */
return (EMSGSIZE);
}
/* We were successful */
return (0);
}
static int
btinitmboxes(struct bt_softc *bt) {
init_32b_mbox_params_t init_mbox;
int error;
bzero(bt->in_boxes, sizeof(bt_mbox_in_t) * bt->num_boxes);
bzero(bt->out_boxes, sizeof(bt_mbox_out_t) * bt->num_boxes);
bt->cur_inbox = bt->in_boxes;
bt->last_inbox = bt->in_boxes + bt->num_boxes - 1;
bt->cur_outbox = bt->out_boxes;
bt->last_outbox = bt->out_boxes + bt->num_boxes - 1;
/* Tell the adapter about them */
init_mbox.num_boxes = bt->num_boxes;
init_mbox.base_addr[0] = bt->mailbox_physbase & 0xFF;
init_mbox.base_addr[1] = (bt->mailbox_physbase >> 8) & 0xFF;
init_mbox.base_addr[2] = (bt->mailbox_physbase >> 16) & 0xFF;
init_mbox.base_addr[3] = (bt->mailbox_physbase >> 24) & 0xFF;
error = bt_cmd(bt, BOP_INITIALIZE_32BMBOX, (u_int8_t *)&init_mbox,
/*parmlen*/sizeof(init_mbox), /*reply_buf*/NULL,
/*reply_len*/0, DEFAULT_CMD_TIMEOUT);
if (error != 0)
printf("btinitmboxes: Initialization command failed\n");
else if (bt->strict_rr != 0) {
/*
* If the controller supports
* strict round robin mode,
* enable it
*/
u_int8_t param;
param = 0;
error = bt_cmd(bt, BOP_ENABLE_STRICT_RR, &param, 1,
/*reply_buf*/NULL, /*reply_len*/0,
DEFAULT_CMD_TIMEOUT);
if (error != 0) {
printf("btinitmboxes: Unable to enable strict RR\n");
error = 0;
} else if (bootverbose) {
device_printf(bt->dev,
"Using Strict Round Robin Mailbox Mode\n");
}
}
return (error);
}
/*
* Update the XPT's idea of the negotiated transfer
* parameters for a particular target.
*/
static void
btfetchtransinfo(struct bt_softc *bt, struct ccb_trans_settings *cts)
{
setup_data_t setup_info;
u_int target;
u_int targ_offset;
u_int targ_mask;
u_int sync_period;
u_int sync_offset;
u_int bus_width;
int error;
u_int8_t param;
targ_syncinfo_t sync_info;
struct ccb_trans_settings_scsi *scsi =
&cts->proto_specific.scsi;
struct ccb_trans_settings_spi *spi =
&cts->xport_specific.spi;
spi->valid = 0;
scsi->valid = 0;
target = cts->ccb_h.target_id;
targ_offset = (target & 0x7);
targ_mask = (0x01 << targ_offset);
/*
* Inquire Setup Information. This command retreives the
* Wide negotiation status for recent adapters as well as
* the sync info for older models.
*/
param = sizeof(setup_info);
error = bt_cmd(bt, BOP_INQUIRE_SETUP_INFO, &param, /*paramlen*/1,
(u_int8_t*)&setup_info, sizeof(setup_info),
DEFAULT_CMD_TIMEOUT);
if (error != 0) {
device_printf(bt->dev,
"btfetchtransinfo - Inquire Setup Info Failed %x\n",
error);
return;
}
sync_info = (target < 8) ? setup_info.low_syncinfo[targ_offset]
: setup_info.high_syncinfo[targ_offset];
if (sync_info.sync == 0)
sync_offset = 0;
else
sync_offset = sync_info.offset;
bus_width = MSG_EXT_WDTR_BUS_8_BIT;
if (strcmp(bt->firmware_ver, "5.06L") >= 0) {
u_int wide_active;
wide_active =
(target < 8) ? (setup_info.low_wide_active & targ_mask)
: (setup_info.high_wide_active & targ_mask);
if (wide_active)
bus_width = MSG_EXT_WDTR_BUS_16_BIT;
} else if ((bt->wide_permitted & targ_mask) != 0) {
struct ccb_getdev cgd;
/*
* Prior to rev 5.06L, wide status isn't provided,
* so we "guess" that wide transfers are in effect
* if the user settings allow for wide and the inquiry
* data for the device indicates that it can handle
* wide transfers.
*/
xpt_setup_ccb(&cgd.ccb_h, cts->ccb_h.path, /*priority*/1);
cgd.ccb_h.func_code = XPT_GDEV_TYPE;
xpt_action((union ccb *)&cgd);
if ((cgd.ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP
&& (cgd.inq_data.flags & SID_WBus16) != 0)
bus_width = MSG_EXT_WDTR_BUS_16_BIT;
}
if (bt->firmware_ver[0] >= '3') {
/*
* For adapters that can do fast or ultra speeds,
* use the more exact Target Sync Information command.
*/
target_sync_info_data_t sync_info;
param = sizeof(sync_info);
error = bt_cmd(bt, BOP_TARG_SYNC_INFO, &param, /*paramlen*/1,
(u_int8_t*)&sync_info, sizeof(sync_info),
DEFAULT_CMD_TIMEOUT);
if (error != 0) {
device_printf(bt->dev,
"btfetchtransinfo - Inquire Sync "
"Info Failed 0x%x\n", error);
return;
}
sync_period = sync_info.sync_rate[target] * 100;
} else {
sync_period = 2000 + (500 * sync_info.period);
}
cts->protocol = PROTO_SCSI;
cts->protocol_version = SCSI_REV_2;
cts->transport = XPORT_SPI;
cts->transport_version = 2;
spi->sync_period = sync_period;
spi->valid |= CTS_SPI_VALID_SYNC_RATE;
spi->sync_offset = sync_offset;
spi->valid |= CTS_SPI_VALID_SYNC_OFFSET;
spi->valid |= CTS_SPI_VALID_BUS_WIDTH;
spi->bus_width = bus_width;
if (cts->ccb_h.target_lun != CAM_LUN_WILDCARD) {
scsi->valid = CTS_SCSI_VALID_TQ;
spi->valid |= CTS_SPI_VALID_DISC;
} else
scsi->valid = 0;
xpt_async(AC_TRANSFER_NEG, cts->ccb_h.path, cts);
}
static void
btmapmboxes(void *arg, bus_dma_segment_t *segs, int nseg, int error)
{
struct bt_softc* bt;
bt = (struct bt_softc*)arg;
bt->mailbox_physbase = segs->ds_addr;
}
static void
btmapccbs(void *arg, bus_dma_segment_t *segs, int nseg, int error)
{
struct bt_softc* bt;
bt = (struct bt_softc*)arg;
bt->bt_ccb_physbase = segs->ds_addr;
}
static void
btmapsgs(void *arg, bus_dma_segment_t *segs, int nseg, int error)
{
struct bt_softc* bt;
bt = (struct bt_softc*)arg;
SLIST_FIRST(&bt->sg_maps)->sg_physaddr = segs->ds_addr;
}
static void
btpoll(struct cam_sim *sim)
{
bt_intr_locked(cam_sim_softc(sim));
}
void
bttimeout(void *arg)
{
struct bt_ccb *bccb;
union ccb *ccb;
struct bt_softc *bt;
bccb = (struct bt_ccb *)arg;
ccb = bccb->ccb;
bt = (struct bt_softc *)ccb->ccb_h.ccb_bt_ptr;
mtx_assert(&bt->lock, MA_OWNED);
xpt_print_path(ccb->ccb_h.path);
printf("CCB %p - timed out\n", (void *)bccb);
if ((bccb->flags & BCCB_ACTIVE) == 0) {
xpt_print_path(ccb->ccb_h.path);
printf("CCB %p - timed out CCB already completed\n",
(void *)bccb);
return;
}
/*
* In order to simplify the recovery process, we ask the XPT
* layer to halt the queue of new transactions and we traverse
* the list of pending CCBs and remove their timeouts. This
* means that the driver attempts to clear only one error
* condition at a time. In general, timeouts that occur
* close together are related anyway, so there is no benefit
* in attempting to handle errors in parrallel. Timeouts will
* be reinstated when the recovery process ends.
*/
if ((bccb->flags & BCCB_DEVICE_RESET) == 0) {
struct ccb_hdr *ccb_h;
if ((bccb->flags & BCCB_RELEASE_SIMQ) == 0) {
xpt_freeze_simq(bt->sim, /*count*/1);
bccb->flags |= BCCB_RELEASE_SIMQ;
}
ccb_h = LIST_FIRST(&bt->pending_ccbs);
while (ccb_h != NULL) {
struct bt_ccb *pending_bccb;
pending_bccb = (struct bt_ccb *)ccb_h->ccb_bccb_ptr;
callout_stop(&pending_bccb->timer);
ccb_h = LIST_NEXT(ccb_h, sim_links.le);
}
}
if ((bccb->flags & BCCB_DEVICE_RESET) != 0
|| bt->cur_outbox->action_code != BMBO_FREE
|| ((bccb->hccb.tag_enable == TRUE)
&& (bt->firmware_ver[0] < '5'))) {
/*
* Try a full host adapter/SCSI bus reset.
* We do this only if we have already attempted
* to clear the condition with a BDR, or we cannot
* attempt a BDR for lack of mailbox resources
* or because of faulty firmware. It turns out
* that firmware versions prior to 5.xx treat BDRs
* as untagged commands that cannot be sent until
* all outstanding tagged commands have been processed.
* This makes it somewhat difficult to use a BDR to
* clear up a problem with an uncompleted tagged command.
*/
ccb->ccb_h.status = CAM_CMD_TIMEOUT;
btreset(bt, /*hardreset*/TRUE);
device_printf(bt->dev, "No longer in timeout\n");
} else {
/*
* Send a Bus Device Reset message:
* The target that is holding up the bus may not
* be the same as the one that triggered this timeout
* (different commands have different timeout lengths),
* but we have no way of determining this from our
* timeout handler. Our strategy here is to queue a
* BDR message to the target of the timed out command.
* If this fails, we'll get another timeout 2 seconds
* later which will attempt a bus reset.
*/
bccb->flags |= BCCB_DEVICE_RESET;
callout_reset(&bccb->timer, 2 * hz, bttimeout, bccb);
bt->recovery_bccb->hccb.opcode = INITIATOR_BUS_DEV_RESET;
/* No Data Transfer */
bt->recovery_bccb->hccb.datain = TRUE;
bt->recovery_bccb->hccb.dataout = TRUE;
bt->recovery_bccb->hccb.btstat = 0;
bt->recovery_bccb->hccb.sdstat = 0;
bt->recovery_bccb->hccb.target_id = ccb->ccb_h.target_id;
/* Tell the adapter about this command */
bt->cur_outbox->ccb_addr = btccbvtop(bt, bt->recovery_bccb);
bt->cur_outbox->action_code = BMBO_START;
bt_outb(bt, COMMAND_REG, BOP_START_MBOX);
btnextoutbox(bt);
}
}
MODULE_VERSION(bt, 1);
MODULE_DEPEND(bt, cam, 1, 1, 1);