fdeb9f2f66
Remove sim queue freezes for resource shortages. I've had too many strange race conditions where I freeze on a resource shortage but never get unfrozen. Consolidate the remaining sim queue freeze condition (for loopdown) into an inline with debug messages that allows us to track problems at ISP_LOGDEBUG0 level easier. Change a bunch of debug messages about loop down/up conditions to ISP_LOGDEBUG0 level. Remove dead isp_relsim code. Change some internal flag stuff for efficiency. Complain vociferously if we try and use our FC scratch area while it's busy being used already (I mean, if we don't have solaris' ability to sleep as an interrupt thread which would allow us to just use a p/v semaphore, at least *say* when you've just borked yourself). Add infrastructure to allow overrides of hard loopid && initiator id from boot variables. Fix the usual quota of silly bugs: + 'ktmature' needs to be per-instance. Argh. + When entering isp_watchdog, set intsok to zero, preserving old value to restore later. It's not nice to try and sleep from splsoftclock. + Fix tick overflow buglet in checking timeout value. MFC after: 1 week
3075 lines
79 KiB
C
3075 lines
79 KiB
C
/* $FreeBSD$ */
|
|
/*
|
|
* Platform (FreeBSD) dependent common attachment code for Qlogic adapters.
|
|
*
|
|
* Copyright (c) 1997, 1998, 1999, 2000, 2001 by Matthew Jacob
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice immediately at the beginning of the file, without modification,
|
|
* this list of conditions, and the following disclaimer.
|
|
* 2. The name of the author may not be used to endorse or promote products
|
|
* derived from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
|
|
* ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
#include <dev/isp/isp_freebsd.h>
|
|
#include <sys/unistd.h>
|
|
#include <sys/kthread.h>
|
|
#include <machine/stdarg.h> /* for use by isp_prt below */
|
|
#include <sys/conf.h>
|
|
#include <sys/ioccom.h>
|
|
#include <dev/isp/isp_ioctl.h>
|
|
|
|
|
|
static d_ioctl_t ispioctl;
|
|
static void isp_intr_enable(void *);
|
|
static void isp_cam_async(void *, u_int32_t, struct cam_path *, void *);
|
|
static void isp_poll(struct cam_sim *);
|
|
static timeout_t isp_watchdog;
|
|
static void isp_kthread(void *);
|
|
static void isp_action(struct cam_sim *, union ccb *);
|
|
|
|
|
|
#define ISP_CDEV_MAJOR 248
|
|
static struct cdevsw isp_cdevsw = {
|
|
/* open */ nullopen,
|
|
/* close */ nullclose,
|
|
/* read */ noread,
|
|
/* write */ nowrite,
|
|
/* ioctl */ ispioctl,
|
|
/* poll */ nopoll,
|
|
/* mmap */ nommap,
|
|
/* strategy */ nostrategy,
|
|
/* name */ "isp",
|
|
/* maj */ ISP_CDEV_MAJOR,
|
|
/* dump */ nodump,
|
|
/* psize */ nopsize,
|
|
/* flags */ D_TAPE,
|
|
};
|
|
|
|
static struct ispsoftc *isplist = NULL;
|
|
|
|
void
|
|
isp_attach(struct ispsoftc *isp)
|
|
{
|
|
int primary, secondary;
|
|
struct ccb_setasync csa;
|
|
struct cam_devq *devq;
|
|
struct cam_sim *sim;
|
|
struct cam_path *path;
|
|
|
|
/*
|
|
* Establish (in case of 12X0) which bus is the primary.
|
|
*/
|
|
|
|
primary = 0;
|
|
secondary = 1;
|
|
|
|
/*
|
|
* Create the device queue for our SIM(s).
|
|
*/
|
|
devq = cam_simq_alloc(isp->isp_maxcmds);
|
|
if (devq == NULL) {
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Construct our SIM entry.
|
|
*/
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
sim = cam_sim_alloc(isp_action, isp_poll, "isp", isp,
|
|
device_get_unit(isp->isp_dev), 1, isp->isp_maxcmds, devq);
|
|
if (sim == NULL) {
|
|
cam_simq_free(devq);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
return;
|
|
}
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
|
|
isp->isp_osinfo.ehook.ich_func = isp_intr_enable;
|
|
isp->isp_osinfo.ehook.ich_arg = isp;
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
if (config_intrhook_establish(&isp->isp_osinfo.ehook) != 0) {
|
|
cam_sim_free(sim, TRUE);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"could not establish interrupt enable hook");
|
|
return;
|
|
}
|
|
|
|
if (xpt_bus_register(sim, primary) != CAM_SUCCESS) {
|
|
cam_sim_free(sim, TRUE);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
return;
|
|
}
|
|
|
|
if (xpt_create_path(&path, NULL, cam_sim_path(sim),
|
|
CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
|
|
xpt_bus_deregister(cam_sim_path(sim));
|
|
cam_sim_free(sim, TRUE);
|
|
config_intrhook_disestablish(&isp->isp_osinfo.ehook);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
return;
|
|
}
|
|
|
|
xpt_setup_ccb(&csa.ccb_h, path, 5);
|
|
csa.ccb_h.func_code = XPT_SASYNC_CB;
|
|
csa.event_enable = AC_LOST_DEVICE;
|
|
csa.callback = isp_cam_async;
|
|
csa.callback_arg = sim;
|
|
xpt_action((union ccb *)&csa);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
isp->isp_sim = sim;
|
|
isp->isp_path = path;
|
|
/*
|
|
* Create a kernel thread for fibre channel instances. We
|
|
* don't have dual channel FC cards.
|
|
*/
|
|
if (IS_FC(isp)) {
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
/* XXX: LOCK VIOLATION */
|
|
cv_init(&isp->isp_osinfo.kthread_cv, "isp_kthread_cv");
|
|
if (kthread_create(isp_kthread, isp, &isp->isp_osinfo.kproc,
|
|
RFHIGHPID, "%s: fc_thrd",
|
|
device_get_nameunit(isp->isp_dev))) {
|
|
xpt_bus_deregister(cam_sim_path(sim));
|
|
cam_sim_free(sim, TRUE);
|
|
config_intrhook_disestablish(&isp->isp_osinfo.ehook);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
isp_prt(isp, ISP_LOGERR, "could not create kthread");
|
|
return;
|
|
}
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
}
|
|
|
|
|
|
/*
|
|
* If we have a second channel, construct SIM entry for that.
|
|
*/
|
|
if (IS_DUALBUS(isp)) {
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
sim = cam_sim_alloc(isp_action, isp_poll, "isp", isp,
|
|
device_get_unit(isp->isp_dev), 1, isp->isp_maxcmds, devq);
|
|
if (sim == NULL) {
|
|
xpt_bus_deregister(cam_sim_path(isp->isp_sim));
|
|
xpt_free_path(isp->isp_path);
|
|
cam_simq_free(devq);
|
|
config_intrhook_disestablish(&isp->isp_osinfo.ehook);
|
|
return;
|
|
}
|
|
if (xpt_bus_register(sim, secondary) != CAM_SUCCESS) {
|
|
xpt_bus_deregister(cam_sim_path(isp->isp_sim));
|
|
xpt_free_path(isp->isp_path);
|
|
cam_sim_free(sim, TRUE);
|
|
config_intrhook_disestablish(&isp->isp_osinfo.ehook);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
return;
|
|
}
|
|
|
|
if (xpt_create_path(&path, NULL, cam_sim_path(sim),
|
|
CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
|
|
xpt_bus_deregister(cam_sim_path(isp->isp_sim));
|
|
xpt_free_path(isp->isp_path);
|
|
xpt_bus_deregister(cam_sim_path(sim));
|
|
cam_sim_free(sim, TRUE);
|
|
config_intrhook_disestablish(&isp->isp_osinfo.ehook);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
return;
|
|
}
|
|
|
|
xpt_setup_ccb(&csa.ccb_h, path, 5);
|
|
csa.ccb_h.func_code = XPT_SASYNC_CB;
|
|
csa.event_enable = AC_LOST_DEVICE;
|
|
csa.callback = isp_cam_async;
|
|
csa.callback_arg = sim;
|
|
xpt_action((union ccb *)&csa);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
isp->isp_sim2 = sim;
|
|
isp->isp_path2 = path;
|
|
}
|
|
|
|
#ifdef ISP_TARGET_MODE
|
|
cv_init(&isp->isp_osinfo.tgtcv0[0], "isp_tgcv0a");
|
|
cv_init(&isp->isp_osinfo.tgtcv0[1], "isp_tgcv0b");
|
|
cv_init(&isp->isp_osinfo.tgtcv1[0], "isp_tgcv1a");
|
|
cv_init(&isp->isp_osinfo.tgtcv1[1], "isp_tgcv1b");
|
|
#endif
|
|
/*
|
|
* Create device nodes
|
|
*/
|
|
(void) make_dev(&isp_cdevsw, device_get_unit(isp->isp_dev), UID_ROOT,
|
|
GID_OPERATOR, 0600, "%s", device_get_nameunit(isp->isp_dev));
|
|
|
|
if (isp->isp_role != ISP_ROLE_NONE) {
|
|
isp->isp_state = ISP_RUNSTATE;
|
|
ENABLE_INTS(isp);
|
|
}
|
|
if (isplist == NULL) {
|
|
isplist = isp;
|
|
} else {
|
|
struct ispsoftc *tmp = isplist;
|
|
while (tmp->isp_osinfo.next) {
|
|
tmp = tmp->isp_osinfo.next;
|
|
}
|
|
tmp->isp_osinfo.next = isp;
|
|
}
|
|
|
|
}
|
|
|
|
static __inline void
|
|
isp_freeze_loopdown(struct ispsoftc *isp, char *msg)
|
|
{
|
|
if (isp->isp_osinfo.simqfrozen == 0) {
|
|
isp_prt(isp, ISP_LOGDEBUG0, "%s: freeze simq (loopdown)", msg);
|
|
isp->isp_osinfo.simqfrozen |= SIMQFRZ_LOOPDOWN;
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_freeze_simq(isp->isp_sim, 1);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
} else {
|
|
isp_prt(isp, ISP_LOGDEBUG0, "%s: mark frozen (loopdown)", msg);
|
|
isp->isp_osinfo.simqfrozen |= SIMQFRZ_LOOPDOWN;
|
|
}
|
|
}
|
|
|
|
static int
|
|
ispioctl(dev_t dev, u_long cmd, caddr_t addr, int flags, struct thread *td)
|
|
{
|
|
struct ispsoftc *isp;
|
|
int retval = ENOTTY;
|
|
|
|
isp = isplist;
|
|
while (isp) {
|
|
if (minor(dev) == device_get_unit(isp->isp_dev)) {
|
|
break;
|
|
}
|
|
isp = isp->isp_osinfo.next;
|
|
}
|
|
if (isp == NULL)
|
|
return (ENXIO);
|
|
|
|
switch (cmd) {
|
|
#ifdef ISP_FW_CRASH_DUMP
|
|
case ISP_GET_FW_CRASH_DUMP:
|
|
{
|
|
u_int16_t *ptr = FCPARAM(isp)->isp_dump_data;
|
|
size_t sz;
|
|
|
|
retval = 0;
|
|
if (IS_2200(isp))
|
|
sz = QLA2200_RISC_IMAGE_DUMP_SIZE;
|
|
else
|
|
sz = QLA2300_RISC_IMAGE_DUMP_SIZE;
|
|
ISP_LOCK(isp);
|
|
if (ptr && *ptr) {
|
|
void *uaddr = *((void **) addr);
|
|
if (copyout(ptr, uaddr, sz)) {
|
|
retval = EFAULT;
|
|
} else {
|
|
*ptr = 0;
|
|
}
|
|
} else {
|
|
retval = ENXIO;
|
|
}
|
|
ISP_UNLOCK(isp);
|
|
break;
|
|
}
|
|
|
|
case ISP_FORCE_CRASH_DUMP:
|
|
ISP_LOCK(isp);
|
|
isp_freeze_loopdown(isp, "ispioctl(ISP_FORCE_CRASH_DUMP)");
|
|
isp_fw_dump(isp);
|
|
isp_reinit(isp);
|
|
ISP_UNLOCK(isp);
|
|
retval = 0;
|
|
break;
|
|
#endif
|
|
case ISP_SDBLEV:
|
|
{
|
|
int olddblev = isp->isp_dblev;
|
|
isp->isp_dblev = *(int *)addr;
|
|
*(int *)addr = olddblev;
|
|
retval = 0;
|
|
break;
|
|
}
|
|
case ISP_RESETHBA:
|
|
ISP_LOCK(isp);
|
|
isp_reinit(isp);
|
|
ISP_UNLOCK(isp);
|
|
retval = 0;
|
|
break;
|
|
case ISP_RESCAN:
|
|
if (IS_FC(isp)) {
|
|
ISP_LOCK(isp);
|
|
if (isp_fc_runstate(isp, 5 * 1000000)) {
|
|
retval = EIO;
|
|
} else {
|
|
retval = 0;
|
|
}
|
|
ISP_UNLOCK(isp);
|
|
}
|
|
break;
|
|
case ISP_FC_LIP:
|
|
if (IS_FC(isp)) {
|
|
ISP_LOCK(isp);
|
|
if (isp_control(isp, ISPCTL_SEND_LIP, 0)) {
|
|
retval = EIO;
|
|
} else {
|
|
retval = 0;
|
|
}
|
|
ISP_UNLOCK(isp);
|
|
}
|
|
break;
|
|
case ISP_FC_GETDINFO:
|
|
{
|
|
struct isp_fc_device *ifc = (struct isp_fc_device *) addr;
|
|
struct lportdb *lp;
|
|
|
|
if (ifc->loopid < 0 || ifc->loopid >= MAX_FC_TARG) {
|
|
retval = EINVAL;
|
|
break;
|
|
}
|
|
ISP_LOCK(isp);
|
|
lp = &FCPARAM(isp)->portdb[ifc->loopid];
|
|
if (lp->valid) {
|
|
ifc->loopid = lp->loopid;
|
|
ifc->portid = lp->portid;
|
|
ifc->node_wwn = lp->node_wwn;
|
|
ifc->port_wwn = lp->port_wwn;
|
|
retval = 0;
|
|
} else {
|
|
retval = ENODEV;
|
|
}
|
|
ISP_UNLOCK(isp);
|
|
break;
|
|
}
|
|
case ISP_GET_STATS:
|
|
{
|
|
isp_stats_t *sp = (isp_stats_t *) addr;
|
|
|
|
MEMZERO(sp, sizeof (*sp));
|
|
sp->isp_stat_version = ISP_STATS_VERSION;
|
|
sp->isp_type = isp->isp_type;
|
|
sp->isp_revision = isp->isp_revision;
|
|
ISP_LOCK(isp);
|
|
sp->isp_stats[ISP_INTCNT] = isp->isp_intcnt;
|
|
sp->isp_stats[ISP_INTBOGUS] = isp->isp_intbogus;
|
|
sp->isp_stats[ISP_INTMBOXC] = isp->isp_intmboxc;
|
|
sp->isp_stats[ISP_INGOASYNC] = isp->isp_intoasync;
|
|
sp->isp_stats[ISP_RSLTCCMPLT] = isp->isp_rsltccmplt;
|
|
sp->isp_stats[ISP_FPHCCMCPLT] = isp->isp_fphccmplt;
|
|
sp->isp_stats[ISP_RSCCHIWAT] = isp->isp_rscchiwater;
|
|
sp->isp_stats[ISP_FPCCHIWAT] = isp->isp_fpcchiwater;
|
|
ISP_UNLOCK(isp);
|
|
retval = 0;
|
|
break;
|
|
}
|
|
case ISP_CLR_STATS:
|
|
ISP_LOCK(isp);
|
|
isp->isp_intcnt = 0;
|
|
isp->isp_intbogus = 0;
|
|
isp->isp_intmboxc = 0;
|
|
isp->isp_intoasync = 0;
|
|
isp->isp_rsltccmplt = 0;
|
|
isp->isp_fphccmplt = 0;
|
|
isp->isp_rscchiwater = 0;
|
|
isp->isp_fpcchiwater = 0;
|
|
ISP_UNLOCK(isp);
|
|
retval = 0;
|
|
break;
|
|
case ISP_FC_GETHINFO:
|
|
{
|
|
struct isp_hba_device *hba = (struct isp_hba_device *) addr;
|
|
MEMZERO(hba, sizeof (*hba));
|
|
ISP_LOCK(isp);
|
|
hba->fc_speed = FCPARAM(isp)->isp_gbspeed;
|
|
hba->fc_scsi_supported = 1;
|
|
hba->fc_topology = FCPARAM(isp)->isp_topo + 1;
|
|
hba->fc_loopid = FCPARAM(isp)->isp_loopid;
|
|
hba->active_node_wwn = FCPARAM(isp)->isp_nodewwn;
|
|
hba->active_port_wwn = FCPARAM(isp)->isp_portwwn;
|
|
ISP_UNLOCK(isp);
|
|
retval = 0;
|
|
break;
|
|
}
|
|
case ISP_GET_FC_PARAM:
|
|
{
|
|
struct isp_fc_param *f = (struct isp_fc_param *) addr;
|
|
|
|
if (!IS_FC(isp)) {
|
|
retval = EINVAL;
|
|
break;
|
|
}
|
|
f->parameter = 0;
|
|
if (strcmp(f->param_name, "framelength") == 0) {
|
|
f->parameter = FCPARAM(isp)->isp_maxfrmlen;
|
|
retval = 0;
|
|
break;
|
|
}
|
|
if (strcmp(f->param_name, "exec_throttle") == 0) {
|
|
f->parameter = FCPARAM(isp)->isp_execthrottle;
|
|
retval = 0;
|
|
break;
|
|
}
|
|
if (strcmp(f->param_name, "fullduplex") == 0) {
|
|
if (FCPARAM(isp)->isp_fwoptions & ICBOPT_FULL_DUPLEX)
|
|
f->parameter = 1;
|
|
retval = 0;
|
|
break;
|
|
}
|
|
if (strcmp(f->param_name, "loopid") == 0) {
|
|
f->parameter = FCPARAM(isp)->isp_loopid;
|
|
retval = 0;
|
|
break;
|
|
}
|
|
retval = EINVAL;
|
|
break;
|
|
}
|
|
case ISP_SET_FC_PARAM:
|
|
{
|
|
struct isp_fc_param *f = (struct isp_fc_param *) addr;
|
|
u_int32_t param = f->parameter;
|
|
|
|
if (!IS_FC(isp)) {
|
|
retval = EINVAL;
|
|
break;
|
|
}
|
|
f->parameter = 0;
|
|
if (strcmp(f->param_name, "framelength") == 0) {
|
|
if (param != 512 && param != 1024 && param != 1024) {
|
|
retval = EINVAL;
|
|
break;
|
|
}
|
|
FCPARAM(isp)->isp_maxfrmlen = param;
|
|
retval = 0;
|
|
break;
|
|
}
|
|
if (strcmp(f->param_name, "exec_throttle") == 0) {
|
|
if (param < 16 || param > 255) {
|
|
retval = EINVAL;
|
|
break;
|
|
}
|
|
FCPARAM(isp)->isp_execthrottle = param;
|
|
retval = 0;
|
|
break;
|
|
}
|
|
if (strcmp(f->param_name, "fullduplex") == 0) {
|
|
if (param != 0 && param != 1) {
|
|
retval = EINVAL;
|
|
break;
|
|
}
|
|
if (param) {
|
|
FCPARAM(isp)->isp_fwoptions |=
|
|
ICBOPT_FULL_DUPLEX;
|
|
} else {
|
|
FCPARAM(isp)->isp_fwoptions &=
|
|
~ICBOPT_FULL_DUPLEX;
|
|
}
|
|
retval = 0;
|
|
break;
|
|
}
|
|
if (strcmp(f->param_name, "loopid") == 0) {
|
|
if (param < 0 || param > 125) {
|
|
retval = EINVAL;
|
|
break;
|
|
}
|
|
FCPARAM(isp)->isp_loopid = param;
|
|
retval = 0;
|
|
break;
|
|
}
|
|
retval = EINVAL;
|
|
break;
|
|
}
|
|
default:
|
|
break;
|
|
}
|
|
return (retval);
|
|
}
|
|
|
|
static void
|
|
isp_intr_enable(void *arg)
|
|
{
|
|
struct ispsoftc *isp = arg;
|
|
if (isp->isp_role != ISP_ROLE_NONE) {
|
|
ENABLE_INTS(isp);
|
|
isp->isp_osinfo.intsok = 1;
|
|
}
|
|
/* Release our hook so that the boot can continue. */
|
|
config_intrhook_disestablish(&isp->isp_osinfo.ehook);
|
|
}
|
|
|
|
/*
|
|
* Put the target mode functions here, because some are inlines
|
|
*/
|
|
|
|
#ifdef ISP_TARGET_MODE
|
|
|
|
static __inline int is_lun_enabled(struct ispsoftc *, int, lun_id_t);
|
|
static __inline int are_any_luns_enabled(struct ispsoftc *, int);
|
|
static __inline tstate_t *get_lun_statep(struct ispsoftc *, int, lun_id_t);
|
|
static __inline void rls_lun_statep(struct ispsoftc *, tstate_t *);
|
|
static __inline int isp_psema_sig_rqe(struct ispsoftc *, int);
|
|
static __inline int isp_cv_wait_timed_rqe(struct ispsoftc *, int, int);
|
|
static __inline void isp_cv_signal_rqe(struct ispsoftc *, int, int);
|
|
static __inline void isp_vsema_rqe(struct ispsoftc *, int);
|
|
static __inline atio_private_data_t *isp_get_atpd(struct ispsoftc *, int);
|
|
static cam_status
|
|
create_lun_state(struct ispsoftc *, int, struct cam_path *, tstate_t **);
|
|
static void destroy_lun_state(struct ispsoftc *, tstate_t *);
|
|
static void isp_en_lun(struct ispsoftc *, union ccb *);
|
|
static cam_status isp_abort_tgt_ccb(struct ispsoftc *, union ccb *);
|
|
static timeout_t isp_refire_putback_atio;
|
|
static void isp_complete_ctio(union ccb *);
|
|
static void isp_target_putback_atio(union ccb *);
|
|
static cam_status isp_target_start_ctio(struct ispsoftc *, union ccb *);
|
|
static int isp_handle_platform_atio(struct ispsoftc *, at_entry_t *);
|
|
static int isp_handle_platform_atio2(struct ispsoftc *, at2_entry_t *);
|
|
static int isp_handle_platform_ctio(struct ispsoftc *, void *);
|
|
static int isp_handle_platform_notify_scsi(struct ispsoftc *, in_entry_t *);
|
|
static int isp_handle_platform_notify_fc(struct ispsoftc *, in_fcentry_t *);
|
|
|
|
static __inline int
|
|
is_lun_enabled(struct ispsoftc *isp, int bus, lun_id_t lun)
|
|
{
|
|
tstate_t *tptr;
|
|
tptr = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(isp, bus, lun)];
|
|
if (tptr == NULL) {
|
|
return (0);
|
|
}
|
|
do {
|
|
if (tptr->lun == (lun_id_t) lun && tptr->bus == bus) {
|
|
return (1);
|
|
}
|
|
} while ((tptr = tptr->next) != NULL);
|
|
return (0);
|
|
}
|
|
|
|
static __inline int
|
|
are_any_luns_enabled(struct ispsoftc *isp, int port)
|
|
{
|
|
int lo, hi;
|
|
if (IS_DUALBUS(isp)) {
|
|
lo = (port * (LUN_HASH_SIZE >> 1));
|
|
hi = lo + (LUN_HASH_SIZE >> 1);
|
|
} else {
|
|
lo = 0;
|
|
hi = LUN_HASH_SIZE;
|
|
}
|
|
for (lo = 0; lo < hi; lo++) {
|
|
if (isp->isp_osinfo.lun_hash[lo]) {
|
|
return (1);
|
|
}
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
static __inline tstate_t *
|
|
get_lun_statep(struct ispsoftc *isp, int bus, lun_id_t lun)
|
|
{
|
|
tstate_t *tptr = NULL;
|
|
|
|
if (lun == CAM_LUN_WILDCARD) {
|
|
if (isp->isp_osinfo.tmflags[bus] & TM_WILDCARD_ENABLED) {
|
|
tptr = &isp->isp_osinfo.tsdflt[bus];
|
|
tptr->hold++;
|
|
return (tptr);
|
|
}
|
|
} else {
|
|
tptr = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(isp, bus, lun)];
|
|
if (tptr == NULL) {
|
|
return (NULL);
|
|
}
|
|
}
|
|
|
|
do {
|
|
if (tptr->lun == lun && tptr->bus == bus) {
|
|
tptr->hold++;
|
|
return (tptr);
|
|
}
|
|
} while ((tptr = tptr->next) != NULL);
|
|
return (tptr);
|
|
}
|
|
|
|
static __inline void
|
|
rls_lun_statep(struct ispsoftc *isp, tstate_t *tptr)
|
|
{
|
|
if (tptr->hold)
|
|
tptr->hold--;
|
|
}
|
|
|
|
static __inline int
|
|
isp_psema_sig_rqe(struct ispsoftc *isp, int bus)
|
|
{
|
|
while (isp->isp_osinfo.tmflags[bus] & TM_BUSY) {
|
|
isp->isp_osinfo.tmflags[bus] |= TM_WANTED;
|
|
if (cv_wait_sig(&isp->isp_osinfo.tgtcv0[bus], &isp->isp_lock)) {
|
|
return (-1);
|
|
}
|
|
isp->isp_osinfo.tmflags[bus] |= TM_BUSY;
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
static __inline int
|
|
isp_cv_wait_timed_rqe(struct ispsoftc *isp, int bus, int timo)
|
|
{
|
|
if (cv_timedwait(&isp->isp_osinfo.tgtcv1[bus], &isp->isp_lock, timo)) {
|
|
return (-1);
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
static __inline void
|
|
isp_cv_signal_rqe(struct ispsoftc *isp, int bus, int status)
|
|
{
|
|
isp->isp_osinfo.rstatus[bus] = status;
|
|
cv_signal(&isp->isp_osinfo.tgtcv1[bus]);
|
|
}
|
|
|
|
static __inline void
|
|
isp_vsema_rqe(struct ispsoftc *isp, int bus)
|
|
{
|
|
if (isp->isp_osinfo.tmflags[bus] & TM_WANTED) {
|
|
isp->isp_osinfo.tmflags[bus] &= ~TM_WANTED;
|
|
cv_signal(&isp->isp_osinfo.tgtcv0[bus]);
|
|
}
|
|
isp->isp_osinfo.tmflags[bus] &= ~TM_BUSY;
|
|
}
|
|
|
|
static __inline atio_private_data_t *
|
|
isp_get_atpd(struct ispsoftc *isp, int tag)
|
|
{
|
|
atio_private_data_t *atp;
|
|
for (atp = isp->isp_osinfo.atpdp;
|
|
atp < &isp->isp_osinfo.atpdp[ATPDPSIZE]; atp++) {
|
|
if (atp->tag == tag)
|
|
return (atp);
|
|
}
|
|
return (NULL);
|
|
}
|
|
|
|
static cam_status
|
|
create_lun_state(struct ispsoftc *isp, int bus,
|
|
struct cam_path *path, tstate_t **rslt)
|
|
{
|
|
cam_status status;
|
|
lun_id_t lun;
|
|
int hfx;
|
|
tstate_t *tptr, *new;
|
|
|
|
lun = xpt_path_lun_id(path);
|
|
if (lun < 0) {
|
|
return (CAM_LUN_INVALID);
|
|
}
|
|
if (is_lun_enabled(isp, bus, lun)) {
|
|
return (CAM_LUN_ALRDY_ENA);
|
|
}
|
|
new = (tstate_t *) malloc(sizeof (tstate_t), M_DEVBUF, M_NOWAIT|M_ZERO);
|
|
if (new == NULL) {
|
|
return (CAM_RESRC_UNAVAIL);
|
|
}
|
|
|
|
status = xpt_create_path(&new->owner, NULL, xpt_path_path_id(path),
|
|
xpt_path_target_id(path), xpt_path_lun_id(path));
|
|
if (status != CAM_REQ_CMP) {
|
|
free(new, M_DEVBUF);
|
|
return (status);
|
|
}
|
|
new->bus = bus;
|
|
new->lun = lun;
|
|
SLIST_INIT(&new->atios);
|
|
SLIST_INIT(&new->inots);
|
|
new->hold = 1;
|
|
|
|
hfx = LUN_HASH_FUNC(isp, new->bus, new->lun);
|
|
tptr = isp->isp_osinfo.lun_hash[hfx];
|
|
if (tptr == NULL) {
|
|
isp->isp_osinfo.lun_hash[hfx] = new;
|
|
} else {
|
|
while (tptr->next)
|
|
tptr = tptr->next;
|
|
tptr->next = new;
|
|
}
|
|
*rslt = new;
|
|
return (CAM_REQ_CMP);
|
|
}
|
|
|
|
static __inline void
|
|
destroy_lun_state(struct ispsoftc *isp, tstate_t *tptr)
|
|
{
|
|
int hfx;
|
|
tstate_t *lw, *pw;
|
|
|
|
hfx = LUN_HASH_FUNC(isp, tptr->bus, tptr->lun);
|
|
if (tptr->hold) {
|
|
return;
|
|
}
|
|
pw = isp->isp_osinfo.lun_hash[hfx];
|
|
if (pw == NULL) {
|
|
return;
|
|
} else if (pw->lun == tptr->lun && pw->bus == tptr->bus) {
|
|
isp->isp_osinfo.lun_hash[hfx] = pw->next;
|
|
} else {
|
|
lw = pw;
|
|
pw = lw->next;
|
|
while (pw) {
|
|
if (pw->lun == tptr->lun && pw->bus == tptr->bus) {
|
|
lw->next = pw->next;
|
|
break;
|
|
}
|
|
lw = pw;
|
|
pw = pw->next;
|
|
}
|
|
if (pw == NULL) {
|
|
return;
|
|
}
|
|
}
|
|
free(tptr, M_DEVBUF);
|
|
}
|
|
|
|
/*
|
|
* we enter with our locks held.
|
|
*/
|
|
static void
|
|
isp_en_lun(struct ispsoftc *isp, union ccb *ccb)
|
|
{
|
|
const char lfmt[] = "Lun now %sabled for target mode on channel %d";
|
|
struct ccb_en_lun *cel = &ccb->cel;
|
|
tstate_t *tptr;
|
|
u_int16_t rstat;
|
|
int bus, cmd, av, wildcard;
|
|
lun_id_t lun;
|
|
target_id_t tgt;
|
|
|
|
|
|
bus = XS_CHANNEL(ccb) & 0x1;
|
|
tgt = ccb->ccb_h.target_id;
|
|
lun = ccb->ccb_h.target_lun;
|
|
|
|
/*
|
|
* Do some sanity checking first.
|
|
*/
|
|
|
|
if ((lun != CAM_LUN_WILDCARD) &&
|
|
(lun < 0 || lun >= (lun_id_t) isp->isp_maxluns)) {
|
|
ccb->ccb_h.status = CAM_LUN_INVALID;
|
|
return;
|
|
}
|
|
|
|
if (IS_SCSI(isp)) {
|
|
sdparam *sdp = isp->isp_param;
|
|
sdp += bus;
|
|
if (tgt != CAM_TARGET_WILDCARD &&
|
|
tgt != sdp->isp_initiator_id) {
|
|
ccb->ccb_h.status = CAM_TID_INVALID;
|
|
return;
|
|
}
|
|
} else {
|
|
if (tgt != CAM_TARGET_WILDCARD &&
|
|
tgt != FCPARAM(isp)->isp_iid) {
|
|
ccb->ccb_h.status = CAM_TID_INVALID;
|
|
return;
|
|
}
|
|
/*
|
|
* This is as a good a place as any to check f/w capabilities.
|
|
*/
|
|
if ((FCPARAM(isp)->isp_fwattr & ISP_FW_ATTR_TMODE) == 0) {
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"firmware does not support target mode");
|
|
ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
|
|
return;
|
|
}
|
|
/*
|
|
* XXX: We *could* handle non-SCCLUN f/w, but we'd have to
|
|
* XXX: dorks with our already fragile enable/disable code.
|
|
*/
|
|
if ((FCPARAM(isp)->isp_fwattr & ISP_FW_ATTR_SCCLUN) == 0) {
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"firmware not SCCLUN capable");
|
|
}
|
|
}
|
|
|
|
if (tgt == CAM_TARGET_WILDCARD) {
|
|
if (lun == CAM_LUN_WILDCARD) {
|
|
wildcard = 1;
|
|
} else {
|
|
ccb->ccb_h.status = CAM_LUN_INVALID;
|
|
return;
|
|
}
|
|
} else {
|
|
wildcard = 0;
|
|
}
|
|
|
|
/*
|
|
* Next check to see whether this is a target/lun wildcard action.
|
|
*
|
|
* If so, we know that we can accept commands for luns that haven't
|
|
* been enabled yet and send them upstream. Otherwise, we have to
|
|
* handle them locally (if we see them at all).
|
|
*/
|
|
|
|
if (wildcard) {
|
|
tptr = &isp->isp_osinfo.tsdflt[bus];
|
|
if (cel->enable) {
|
|
if (isp->isp_osinfo.tmflags[bus] &
|
|
TM_WILDCARD_ENABLED) {
|
|
ccb->ccb_h.status = CAM_LUN_ALRDY_ENA;
|
|
return;
|
|
}
|
|
ccb->ccb_h.status =
|
|
xpt_create_path(&tptr->owner, NULL,
|
|
xpt_path_path_id(ccb->ccb_h.path),
|
|
xpt_path_target_id(ccb->ccb_h.path),
|
|
xpt_path_lun_id(ccb->ccb_h.path));
|
|
if (ccb->ccb_h.status != CAM_REQ_CMP) {
|
|
return;
|
|
}
|
|
SLIST_INIT(&tptr->atios);
|
|
SLIST_INIT(&tptr->inots);
|
|
isp->isp_osinfo.tmflags[bus] |= TM_WILDCARD_ENABLED;
|
|
} else {
|
|
if ((isp->isp_osinfo.tmflags[bus] &
|
|
TM_WILDCARD_ENABLED) == 0) {
|
|
ccb->ccb_h.status = CAM_REQ_CMP;
|
|
return;
|
|
}
|
|
if (tptr->hold) {
|
|
ccb->ccb_h.status = CAM_SCSI_BUSY;
|
|
return;
|
|
}
|
|
xpt_free_path(tptr->owner);
|
|
isp->isp_osinfo.tmflags[bus] &= ~TM_WILDCARD_ENABLED;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Now check to see whether this bus needs to be
|
|
* enabled/disabled with respect to target mode.
|
|
*/
|
|
av = bus << 31;
|
|
if (cel->enable && !(isp->isp_osinfo.tmflags[bus] & TM_TMODE_ENABLED)) {
|
|
av |= ENABLE_TARGET_FLAG;
|
|
av = isp_control(isp, ISPCTL_TOGGLE_TMODE, &av);
|
|
if (av) {
|
|
ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
|
|
if (wildcard) {
|
|
isp->isp_osinfo.tmflags[bus] &=
|
|
~TM_WILDCARD_ENABLED;
|
|
xpt_free_path(tptr->owner);
|
|
}
|
|
return;
|
|
}
|
|
isp->isp_osinfo.tmflags[bus] |= TM_TMODE_ENABLED;
|
|
isp_prt(isp, ISP_LOGINFO,
|
|
"Target Mode enabled on channel %d", bus);
|
|
} else if (cel->enable == 0 &&
|
|
(isp->isp_osinfo.tmflags[bus] & TM_TMODE_ENABLED) && wildcard) {
|
|
if (are_any_luns_enabled(isp, bus)) {
|
|
ccb->ccb_h.status = CAM_SCSI_BUSY;
|
|
return;
|
|
}
|
|
av = isp_control(isp, ISPCTL_TOGGLE_TMODE, &av);
|
|
if (av) {
|
|
ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
|
|
return;
|
|
}
|
|
isp->isp_osinfo.tmflags[bus] &= ~TM_TMODE_ENABLED;
|
|
isp_prt(isp, ISP_LOGINFO,
|
|
"Target Mode disabled on channel %d", bus);
|
|
}
|
|
|
|
if (wildcard) {
|
|
ccb->ccb_h.status = CAM_REQ_CMP;
|
|
return;
|
|
}
|
|
|
|
if (cel->enable) {
|
|
ccb->ccb_h.status =
|
|
create_lun_state(isp, bus, ccb->ccb_h.path, &tptr);
|
|
if (ccb->ccb_h.status != CAM_REQ_CMP) {
|
|
return;
|
|
}
|
|
} else {
|
|
tptr = get_lun_statep(isp, bus, lun);
|
|
if (tptr == NULL) {
|
|
ccb->ccb_h.status = CAM_LUN_INVALID;
|
|
return;
|
|
}
|
|
}
|
|
|
|
if (isp_psema_sig_rqe(isp, bus)) {
|
|
rls_lun_statep(isp, tptr);
|
|
if (cel->enable)
|
|
destroy_lun_state(isp, tptr);
|
|
ccb->ccb_h.status = CAM_REQ_CMP_ERR;
|
|
return;
|
|
}
|
|
|
|
if (cel->enable) {
|
|
u_int32_t seq = isp->isp_osinfo.rollinfo++;
|
|
int c, n, ulun = lun;
|
|
|
|
cmd = RQSTYPE_ENABLE_LUN;
|
|
c = DFLT_CMND_CNT;
|
|
n = DFLT_INOT_CNT;
|
|
if (IS_FC(isp) && lun != 0) {
|
|
cmd = RQSTYPE_MODIFY_LUN;
|
|
n = 0;
|
|
/*
|
|
* For SCC firmware, we only deal with setting
|
|
* (enabling or modifying) lun 0.
|
|
*/
|
|
ulun = 0;
|
|
}
|
|
rstat = LUN_ERR;
|
|
if (isp_lun_cmd(isp, cmd, bus, tgt, ulun, c, n, seq)) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGWARN, "isp_lun_cmd failed");
|
|
goto out;
|
|
}
|
|
if (isp_cv_wait_timed_rqe(isp, bus, 30 * hz)) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"wait for ENABLE/MODIFY LUN timed out");
|
|
goto out;
|
|
}
|
|
rstat = isp->isp_osinfo.rstatus[bus];
|
|
if (rstat != LUN_OK) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"ENABLE/MODIFY LUN returned 0x%x", rstat);
|
|
goto out;
|
|
}
|
|
} else {
|
|
int c, n, ulun = lun;
|
|
u_int32_t seq;
|
|
|
|
rstat = LUN_ERR;
|
|
seq = isp->isp_osinfo.rollinfo++;
|
|
cmd = -RQSTYPE_MODIFY_LUN;
|
|
|
|
c = DFLT_CMND_CNT;
|
|
n = DFLT_INOT_CNT;
|
|
if (IS_FC(isp) && lun != 0) {
|
|
n = 0;
|
|
/*
|
|
* For SCC firmware, we only deal with setting
|
|
* (enabling or modifying) lun 0.
|
|
*/
|
|
ulun = 0;
|
|
}
|
|
if (isp_lun_cmd(isp, cmd, bus, tgt, ulun, c, n, seq)) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGERR, "isp_lun_cmd failed");
|
|
goto out;
|
|
}
|
|
if (isp_cv_wait_timed_rqe(isp, bus, 30 * hz)) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"wait for MODIFY LUN timed out");
|
|
goto out;
|
|
}
|
|
rstat = isp->isp_osinfo.rstatus[bus];
|
|
if (rstat != LUN_OK) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"MODIFY LUN returned 0x%x", rstat);
|
|
goto out;
|
|
}
|
|
if (IS_FC(isp) && lun) {
|
|
goto out;
|
|
}
|
|
|
|
seq = isp->isp_osinfo.rollinfo++;
|
|
|
|
rstat = LUN_ERR;
|
|
cmd = -RQSTYPE_ENABLE_LUN;
|
|
if (isp_lun_cmd(isp, cmd, bus, tgt, lun, 0, 0, seq)) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGERR, "isp_lun_cmd failed");
|
|
goto out;
|
|
}
|
|
if (isp_cv_wait_timed_rqe(isp, bus, 30 * hz)) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"wait for DISABLE LUN timed out");
|
|
goto out;
|
|
}
|
|
rstat = isp->isp_osinfo.rstatus[bus];
|
|
if (rstat != LUN_OK) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"DISABLE LUN returned 0x%x", rstat);
|
|
goto out;
|
|
}
|
|
if (are_any_luns_enabled(isp, bus) == 0) {
|
|
av = isp_control(isp, ISPCTL_TOGGLE_TMODE, &av);
|
|
if (av) {
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"disable target mode on channel %d failed",
|
|
bus);
|
|
goto out;
|
|
}
|
|
isp->isp_osinfo.tmflags[bus] &= ~TM_TMODE_ENABLED;
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGINFO,
|
|
"Target Mode disabled on channel %d", bus);
|
|
}
|
|
}
|
|
|
|
out:
|
|
isp_vsema_rqe(isp, bus);
|
|
|
|
if (rstat != LUN_OK) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"lun %sable failed", (cel->enable) ? "en" : "dis");
|
|
ccb->ccb_h.status = CAM_REQ_CMP_ERR;
|
|
rls_lun_statep(isp, tptr);
|
|
if (cel->enable)
|
|
destroy_lun_state(isp, tptr);
|
|
} else {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGINFO, lfmt,
|
|
(cel->enable) ? "en" : "dis", bus);
|
|
rls_lun_statep(isp, tptr);
|
|
if (cel->enable == 0) {
|
|
destroy_lun_state(isp, tptr);
|
|
}
|
|
ccb->ccb_h.status = CAM_REQ_CMP;
|
|
}
|
|
}
|
|
|
|
static cam_status
|
|
isp_abort_tgt_ccb(struct ispsoftc *isp, union ccb *ccb)
|
|
{
|
|
tstate_t *tptr;
|
|
struct ccb_hdr_slist *lp;
|
|
struct ccb_hdr *curelm;
|
|
int found;
|
|
union ccb *accb = ccb->cab.abort_ccb;
|
|
|
|
if (accb->ccb_h.target_id != CAM_TARGET_WILDCARD) {
|
|
if (IS_FC(isp) && (accb->ccb_h.target_id !=
|
|
((fcparam *) isp->isp_param)->isp_loopid)) {
|
|
return (CAM_PATH_INVALID);
|
|
} else if (IS_SCSI(isp) && (accb->ccb_h.target_id !=
|
|
((sdparam *) isp->isp_param)->isp_initiator_id)) {
|
|
return (CAM_PATH_INVALID);
|
|
}
|
|
}
|
|
tptr = get_lun_statep(isp, XS_CHANNEL(ccb), accb->ccb_h.target_lun);
|
|
if (tptr == NULL) {
|
|
return (CAM_PATH_INVALID);
|
|
}
|
|
if (accb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) {
|
|
lp = &tptr->atios;
|
|
} else if (accb->ccb_h.func_code == XPT_IMMED_NOTIFY) {
|
|
lp = &tptr->inots;
|
|
} else {
|
|
rls_lun_statep(isp, tptr);
|
|
return (CAM_UA_ABORT);
|
|
}
|
|
curelm = SLIST_FIRST(lp);
|
|
found = 0;
|
|
if (curelm == &accb->ccb_h) {
|
|
found = 1;
|
|
SLIST_REMOVE_HEAD(lp, sim_links.sle);
|
|
} else {
|
|
while(curelm != NULL) {
|
|
struct ccb_hdr *nextelm;
|
|
|
|
nextelm = SLIST_NEXT(curelm, sim_links.sle);
|
|
if (nextelm == &accb->ccb_h) {
|
|
found = 1;
|
|
SLIST_NEXT(curelm, sim_links.sle) =
|
|
SLIST_NEXT(nextelm, sim_links.sle);
|
|
break;
|
|
}
|
|
curelm = nextelm;
|
|
}
|
|
}
|
|
rls_lun_statep(isp, tptr);
|
|
if (found) {
|
|
accb->ccb_h.status = CAM_REQ_ABORTED;
|
|
return (CAM_REQ_CMP);
|
|
}
|
|
return(CAM_PATH_INVALID);
|
|
}
|
|
|
|
static cam_status
|
|
isp_target_start_ctio(struct ispsoftc *isp, union ccb *ccb)
|
|
{
|
|
void *qe;
|
|
struct ccb_scsiio *cso = &ccb->csio;
|
|
u_int16_t *hp, save_handle;
|
|
u_int16_t nxti, optr;
|
|
u_int8_t local[QENTRY_LEN];
|
|
|
|
|
|
if (isp_getrqentry(isp, &nxti, &optr, &qe)) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
printf("Request Queue Overflow in isp_target_start_ctio\n");
|
|
return (CAM_RESRC_UNAVAIL);
|
|
}
|
|
bzero(local, QENTRY_LEN);
|
|
|
|
/*
|
|
* We're either moving data or completing a command here.
|
|
*/
|
|
|
|
if (IS_FC(isp)) {
|
|
atio_private_data_t *atp;
|
|
ct2_entry_t *cto = (ct2_entry_t *) local;
|
|
|
|
cto->ct_header.rqs_entry_type = RQSTYPE_CTIO2;
|
|
cto->ct_header.rqs_entry_count = 1;
|
|
cto->ct_iid = cso->init_id;
|
|
if ((FCPARAM(isp)->isp_fwattr & ISP_FW_ATTR_SCCLUN) == 0) {
|
|
cto->ct_lun = ccb->ccb_h.target_lun;
|
|
}
|
|
|
|
atp = isp_get_atpd(isp, cso->tag_id);
|
|
if (atp == NULL) {
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"cannot find private data adjunct for tag %x",
|
|
cso->tag_id);
|
|
return (-1);
|
|
}
|
|
|
|
cto->ct_rxid = cso->tag_id;
|
|
if (cso->dxfer_len == 0) {
|
|
cto->ct_flags |= CT2_FLAG_MODE1 | CT2_NO_DATA;
|
|
if (ccb->ccb_h.flags & CAM_SEND_STATUS) {
|
|
cto->ct_flags |= CT2_SENDSTATUS;
|
|
cto->rsp.m1.ct_scsi_status = cso->scsi_status;
|
|
cto->ct_resid =
|
|
atp->orig_datalen - atp->bytes_xfered;
|
|
if (cto->ct_resid < 0) {
|
|
cto->rsp.m1.ct_scsi_status |=
|
|
CT2_DATA_OVER;
|
|
} else if (cto->ct_resid > 0) {
|
|
cto->rsp.m1.ct_scsi_status |=
|
|
CT2_DATA_UNDER;
|
|
}
|
|
}
|
|
if ((ccb->ccb_h.flags & CAM_SEND_SENSE) != 0) {
|
|
int m = min(cso->sense_len, MAXRESPLEN);
|
|
bcopy(&cso->sense_data, cto->rsp.m1.ct_resp, m);
|
|
cto->rsp.m1.ct_senselen = m;
|
|
cto->rsp.m1.ct_scsi_status |= CT2_SNSLEN_VALID;
|
|
}
|
|
} else {
|
|
cto->ct_flags |= CT2_FLAG_MODE0;
|
|
if ((cso->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) {
|
|
cto->ct_flags |= CT2_DATA_IN;
|
|
} else {
|
|
cto->ct_flags |= CT2_DATA_OUT;
|
|
}
|
|
cto->ct_reloff = atp->bytes_xfered;
|
|
if ((ccb->ccb_h.flags & CAM_SEND_STATUS) != 0) {
|
|
cto->ct_flags |= CT2_SENDSTATUS;
|
|
cto->rsp.m0.ct_scsi_status = cso->scsi_status;
|
|
cto->ct_resid =
|
|
atp->orig_datalen -
|
|
(atp->bytes_xfered + cso->dxfer_len);
|
|
if (cto->ct_resid < 0) {
|
|
cto->rsp.m0.ct_scsi_status |=
|
|
CT2_DATA_OVER;
|
|
} else if (cto->ct_resid > 0) {
|
|
cto->rsp.m0.ct_scsi_status |=
|
|
CT2_DATA_UNDER;
|
|
}
|
|
} else {
|
|
atp->last_xframt = cso->dxfer_len;
|
|
}
|
|
/*
|
|
* If we're sending data and status back together,
|
|
* we can't also send back sense data as well.
|
|
*/
|
|
ccb->ccb_h.flags &= ~CAM_SEND_SENSE;
|
|
}
|
|
|
|
if (cto->ct_flags & CT2_SENDSTATUS) {
|
|
isp_prt(isp, ISP_LOGTDEBUG0,
|
|
"CTIO2[%x] STATUS %x origd %u curd %u resid %u",
|
|
cto->ct_rxid, cso->scsi_status, atp->orig_datalen,
|
|
cso->dxfer_len, cto->ct_resid);
|
|
cto->ct_flags |= CT2_CCINCR;
|
|
atp->state = ATPD_STATE_LAST_CTIO;
|
|
} else
|
|
atp->state = ATPD_STATE_CTIO;
|
|
cto->ct_timeout = 10;
|
|
hp = &cto->ct_syshandle;
|
|
} else {
|
|
ct_entry_t *cto = (ct_entry_t *) local;
|
|
|
|
cto->ct_header.rqs_entry_type = RQSTYPE_CTIO;
|
|
cto->ct_header.rqs_entry_count = 1;
|
|
cto->ct_iid = cso->init_id;
|
|
cto->ct_iid |= XS_CHANNEL(ccb) << 7;
|
|
cto->ct_tgt = ccb->ccb_h.target_id;
|
|
cto->ct_lun = ccb->ccb_h.target_lun;
|
|
cto->ct_fwhandle = AT_GET_HANDLE(cso->tag_id);
|
|
if (AT_HAS_TAG(cso->tag_id)) {
|
|
cto->ct_tag_val = (u_int8_t) AT_GET_TAG(cso->tag_id);
|
|
cto->ct_flags |= CT_TQAE;
|
|
}
|
|
if (ccb->ccb_h.flags & CAM_DIS_DISCONNECT) {
|
|
cto->ct_flags |= CT_NODISC;
|
|
}
|
|
if (cso->dxfer_len == 0) {
|
|
cto->ct_flags |= CT_NO_DATA;
|
|
} else if ((cso->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) {
|
|
cto->ct_flags |= CT_DATA_IN;
|
|
} else {
|
|
cto->ct_flags |= CT_DATA_OUT;
|
|
}
|
|
if (ccb->ccb_h.flags & CAM_SEND_STATUS) {
|
|
cto->ct_flags |= CT_SENDSTATUS|CT_CCINCR;
|
|
cto->ct_scsi_status = cso->scsi_status;
|
|
cto->ct_resid = cso->resid;
|
|
isp_prt(isp, ISP_LOGTDEBUG0,
|
|
"CTIO[%x] SCSI STATUS 0x%x resid %d tag_id %x",
|
|
cto->ct_fwhandle, cso->scsi_status, cso->resid,
|
|
cso->tag_id);
|
|
}
|
|
ccb->ccb_h.flags &= ~CAM_SEND_SENSE;
|
|
cto->ct_timeout = 10;
|
|
hp = &cto->ct_syshandle;
|
|
}
|
|
|
|
if (isp_save_xs(isp, (XS_T *)ccb, hp)) {
|
|
xpt_print_path(ccb->ccb_h.path);
|
|
printf("No XFLIST pointers for isp_target_start_ctio\n");
|
|
return (CAM_RESRC_UNAVAIL);
|
|
}
|
|
|
|
|
|
/*
|
|
* Call the dma setup routines for this entry (and any subsequent
|
|
* CTIOs) if there's data to move, and then tell the f/w it's got
|
|
* new things to play with. As with isp_start's usage of DMA setup,
|
|
* any swizzling is done in the machine dependent layer. Because
|
|
* of this, we put the request onto the queue area first in native
|
|
* format.
|
|
*/
|
|
|
|
save_handle = *hp;
|
|
|
|
switch (ISP_DMASETUP(isp, cso, (ispreq_t *) local, &nxti, optr)) {
|
|
case CMD_QUEUED:
|
|
ISP_ADD_REQUEST(isp, nxti);
|
|
return (CAM_REQ_INPROG);
|
|
|
|
case CMD_EAGAIN:
|
|
ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
|
|
isp_destroy_handle(isp, save_handle);
|
|
return (CAM_RESRC_UNAVAIL);
|
|
|
|
default:
|
|
isp_destroy_handle(isp, save_handle);
|
|
return (XS_ERR(ccb));
|
|
}
|
|
}
|
|
|
|
static void
|
|
isp_refire_putback_atio(void *arg)
|
|
{
|
|
int s = splcam();
|
|
isp_target_putback_atio(arg);
|
|
splx(s);
|
|
}
|
|
|
|
static void
|
|
isp_target_putback_atio(union ccb *ccb)
|
|
{
|
|
struct ispsoftc *isp;
|
|
struct ccb_scsiio *cso;
|
|
u_int16_t nxti, optr;
|
|
void *qe;
|
|
|
|
isp = XS_ISP(ccb);
|
|
|
|
if (isp_getrqentry(isp, &nxti, &optr, &qe)) {
|
|
(void) timeout(isp_refire_putback_atio, ccb, 10);
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"isp_target_putback_atio: Request Queue Overflow");
|
|
return;
|
|
}
|
|
bzero(qe, QENTRY_LEN);
|
|
cso = &ccb->csio;
|
|
if (IS_FC(isp)) {
|
|
at2_entry_t local, *at = &local;
|
|
MEMZERO(at, sizeof (at2_entry_t));
|
|
at->at_header.rqs_entry_type = RQSTYPE_ATIO2;
|
|
at->at_header.rqs_entry_count = 1;
|
|
if ((FCPARAM(isp)->isp_fwattr & ISP_FW_ATTR_SCCLUN) != 0) {
|
|
at->at_scclun = (uint16_t) ccb->ccb_h.target_lun;
|
|
} else {
|
|
at->at_lun = (uint8_t) ccb->ccb_h.target_lun;
|
|
}
|
|
at->at_status = CT_OK;
|
|
at->at_rxid = cso->tag_id;
|
|
at->at_iid = cso->ccb_h.target_id;
|
|
isp_put_atio2(isp, at, qe);
|
|
} else {
|
|
at_entry_t local, *at = &local;
|
|
MEMZERO(at, sizeof (at_entry_t));
|
|
at->at_header.rqs_entry_type = RQSTYPE_ATIO;
|
|
at->at_header.rqs_entry_count = 1;
|
|
at->at_iid = cso->init_id;
|
|
at->at_iid |= XS_CHANNEL(ccb) << 7;
|
|
at->at_tgt = cso->ccb_h.target_id;
|
|
at->at_lun = cso->ccb_h.target_lun;
|
|
at->at_status = CT_OK;
|
|
at->at_tag_val = AT_GET_TAG(cso->tag_id);
|
|
at->at_handle = AT_GET_HANDLE(cso->tag_id);
|
|
isp_put_atio(isp, at, qe);
|
|
}
|
|
ISP_TDQE(isp, "isp_target_putback_atio", (int) optr, qe);
|
|
ISP_ADD_REQUEST(isp, nxti);
|
|
isp_complete_ctio(ccb);
|
|
}
|
|
|
|
static void
|
|
isp_complete_ctio(union ccb *ccb)
|
|
{
|
|
struct ispsoftc *isp = XS_ISP(ccb);
|
|
if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INPROG) {
|
|
ccb->ccb_h.status |= CAM_REQ_CMP;
|
|
}
|
|
ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
|
|
xpt_done(ccb);
|
|
}
|
|
|
|
/*
|
|
* Handle ATIO stuff that the generic code can't.
|
|
* This means handling CDBs.
|
|
*/
|
|
|
|
static int
|
|
isp_handle_platform_atio(struct ispsoftc *isp, at_entry_t *aep)
|
|
{
|
|
tstate_t *tptr;
|
|
int status, bus, iswildcard;
|
|
struct ccb_accept_tio *atiop;
|
|
|
|
/*
|
|
* The firmware status (except for the QLTM_SVALID bit)
|
|
* indicates why this ATIO was sent to us.
|
|
*
|
|
* If QLTM_SVALID is set, the firware has recommended Sense Data.
|
|
*
|
|
* If the DISCONNECTS DISABLED bit is set in the flags field,
|
|
* we're still connected on the SCSI bus.
|
|
*/
|
|
status = aep->at_status;
|
|
if ((status & ~QLTM_SVALID) == AT_PHASE_ERROR) {
|
|
/*
|
|
* Bus Phase Sequence error. We should have sense data
|
|
* suggested by the f/w. I'm not sure quite yet what
|
|
* to do about this for CAM.
|
|
*/
|
|
isp_prt(isp, ISP_LOGWARN, "PHASE ERROR");
|
|
isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0);
|
|
return (0);
|
|
}
|
|
if ((status & ~QLTM_SVALID) != AT_CDB) {
|
|
isp_prt(isp, ISP_LOGWARN, "bad atio (0x%x) leaked to platform",
|
|
status);
|
|
isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0);
|
|
return (0);
|
|
}
|
|
|
|
bus = GET_BUS_VAL(aep->at_iid);
|
|
tptr = get_lun_statep(isp, bus, aep->at_lun);
|
|
if (tptr == NULL) {
|
|
tptr = get_lun_statep(isp, bus, CAM_LUN_WILDCARD);
|
|
iswildcard = 1;
|
|
} else {
|
|
iswildcard = 0;
|
|
}
|
|
|
|
if (tptr == NULL) {
|
|
/*
|
|
* Because we can't autofeed sense data back with
|
|
* a command for parallel SCSI, we can't give back
|
|
* a CHECK CONDITION. We'll give back a BUSY status
|
|
* instead. This works out okay because the only
|
|
* time we should, in fact, get this, is in the
|
|
* case that somebody configured us without the
|
|
* blackhole driver, so they get what they deserve.
|
|
*/
|
|
isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0);
|
|
return (0);
|
|
}
|
|
|
|
atiop = (struct ccb_accept_tio *) SLIST_FIRST(&tptr->atios);
|
|
if (atiop == NULL) {
|
|
/*
|
|
* Because we can't autofeed sense data back with
|
|
* a command for parallel SCSI, we can't give back
|
|
* a CHECK CONDITION. We'll give back a QUEUE FULL status
|
|
* instead. This works out okay because the only time we
|
|
* should, in fact, get this, is in the case that we've
|
|
* run out of ATIOS.
|
|
*/
|
|
xpt_print_path(tptr->owner);
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"no ATIOS for lun %d from initiator %d on channel %d",
|
|
aep->at_lun, GET_IID_VAL(aep->at_iid), bus);
|
|
if (aep->at_flags & AT_TQAE)
|
|
isp_endcmd(isp, aep, SCSI_STATUS_QUEUE_FULL, 0);
|
|
else
|
|
isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0);
|
|
rls_lun_statep(isp, tptr);
|
|
return (0);
|
|
}
|
|
SLIST_REMOVE_HEAD(&tptr->atios, sim_links.sle);
|
|
if (iswildcard) {
|
|
atiop->ccb_h.target_id = aep->at_tgt;
|
|
atiop->ccb_h.target_lun = aep->at_lun;
|
|
}
|
|
if (aep->at_flags & AT_NODISC) {
|
|
atiop->ccb_h.flags = CAM_DIS_DISCONNECT;
|
|
} else {
|
|
atiop->ccb_h.flags = 0;
|
|
}
|
|
|
|
if (status & QLTM_SVALID) {
|
|
size_t amt = imin(QLTM_SENSELEN, sizeof (atiop->sense_data));
|
|
atiop->sense_len = amt;
|
|
MEMCPY(&atiop->sense_data, aep->at_sense, amt);
|
|
} else {
|
|
atiop->sense_len = 0;
|
|
}
|
|
|
|
atiop->init_id = GET_IID_VAL(aep->at_iid);
|
|
atiop->cdb_len = aep->at_cdblen;
|
|
MEMCPY(atiop->cdb_io.cdb_bytes, aep->at_cdb, aep->at_cdblen);
|
|
atiop->ccb_h.status = CAM_CDB_RECVD;
|
|
/*
|
|
* Construct a tag 'id' based upon tag value (which may be 0..255)
|
|
* and the handle (which we have to preserve).
|
|
*/
|
|
AT_MAKE_TAGID(atiop->tag_id, aep);
|
|
if (aep->at_flags & AT_TQAE) {
|
|
atiop->tag_action = aep->at_tag_type;
|
|
atiop->ccb_h.status |= CAM_TAG_ACTION_VALID;
|
|
}
|
|
xpt_done((union ccb*)atiop);
|
|
isp_prt(isp, ISP_LOGTDEBUG0,
|
|
"ATIO[%x] CDB=0x%x bus %d iid%d->lun%d tag 0x%x ttype 0x%x %s",
|
|
aep->at_handle, aep->at_cdb[0] & 0xff, GET_BUS_VAL(aep->at_iid),
|
|
GET_IID_VAL(aep->at_iid), aep->at_lun, aep->at_tag_val & 0xff,
|
|
aep->at_tag_type, (aep->at_flags & AT_NODISC)?
|
|
"nondisc" : "disconnecting");
|
|
rls_lun_statep(isp, tptr);
|
|
return (0);
|
|
}
|
|
|
|
static int
|
|
isp_handle_platform_atio2(struct ispsoftc *isp, at2_entry_t *aep)
|
|
{
|
|
lun_id_t lun;
|
|
tstate_t *tptr;
|
|
struct ccb_accept_tio *atiop;
|
|
atio_private_data_t *atp;
|
|
|
|
/*
|
|
* The firmware status (except for the QLTM_SVALID bit)
|
|
* indicates why this ATIO was sent to us.
|
|
*
|
|
* If QLTM_SVALID is set, the firware has recommended Sense Data.
|
|
*/
|
|
if ((aep->at_status & ~QLTM_SVALID) != AT_CDB) {
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"bogus atio (0x%x) leaked to platform", aep->at_status);
|
|
isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0);
|
|
return (0);
|
|
}
|
|
|
|
if ((FCPARAM(isp)->isp_fwattr & ISP_FW_ATTR_SCCLUN) != 0) {
|
|
lun = aep->at_scclun;
|
|
} else {
|
|
lun = aep->at_lun;
|
|
}
|
|
tptr = get_lun_statep(isp, 0, lun);
|
|
if (tptr == NULL) {
|
|
isp_prt(isp, ISP_LOGWARN, "no state pointer for lun %d", lun);
|
|
tptr = get_lun_statep(isp, 0, CAM_LUN_WILDCARD);
|
|
}
|
|
|
|
if (tptr == NULL) {
|
|
/*
|
|
* What we'd like to know is whether or not we have a listener
|
|
* upstream that really hasn't configured yet. If we do, then
|
|
* we can give a more sensible reply here. If not, then we can
|
|
* reject this out of hand.
|
|
*
|
|
* Choices for what to send were
|
|
*
|
|
* Not Ready, Unit Not Self-Configured Yet
|
|
* (0x2,0x3e,0x00)
|
|
*
|
|
* for the former and
|
|
*
|
|
* Illegal Request, Logical Unit Not Supported
|
|
* (0x5,0x25,0x00)
|
|
*
|
|
* for the latter.
|
|
*
|
|
* We used to decide whether there was at least one listener
|
|
* based upon whether the black hole driver was configured.
|
|
* However, recent config(8) changes have made this hard to do
|
|
* at this time.
|
|
*
|
|
*/
|
|
isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0);
|
|
return (0);
|
|
}
|
|
|
|
atp = isp_get_atpd(isp, 0);
|
|
atiop = (struct ccb_accept_tio *) SLIST_FIRST(&tptr->atios);
|
|
if (atiop == NULL || atp == NULL) {
|
|
/*
|
|
* Because we can't autofeed sense data back with
|
|
* a command for parallel SCSI, we can't give back
|
|
* a CHECK CONDITION. We'll give back a QUEUE FULL status
|
|
* instead. This works out okay because the only time we
|
|
* should, in fact, get this, is in the case that we've
|
|
* run out of ATIOS.
|
|
*/
|
|
xpt_print_path(tptr->owner);
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"no %s for lun %d from initiator %d",
|
|
(atp == NULL && atiop == NULL)? "ATIO2s *or* ATPS" :
|
|
((atp == NULL)? "ATPs" : "ATIO2s"), lun, aep->at_iid);
|
|
rls_lun_statep(isp, tptr);
|
|
isp_endcmd(isp, aep, SCSI_STATUS_QUEUE_FULL, 0);
|
|
return (0);
|
|
}
|
|
atp->state = ATPD_STATE_ATIO;
|
|
SLIST_REMOVE_HEAD(&tptr->atios, sim_links.sle);
|
|
tptr->atio_count--;
|
|
isp_prt(isp, ISP_LOGTDEBUG0, "Take FREE ATIO2 lun %d, count now %d",
|
|
lun, tptr->atio_count);
|
|
|
|
if (tptr == &isp->isp_osinfo.tsdflt[0]) {
|
|
atiop->ccb_h.target_id =
|
|
((fcparam *)isp->isp_param)->isp_loopid;
|
|
atiop->ccb_h.target_lun = lun;
|
|
}
|
|
/*
|
|
* We don't get 'suggested' sense data as we do with SCSI cards.
|
|
*/
|
|
atiop->sense_len = 0;
|
|
|
|
atiop->init_id = aep->at_iid;
|
|
atiop->cdb_len = ATIO2_CDBLEN;
|
|
MEMCPY(atiop->cdb_io.cdb_bytes, aep->at_cdb, ATIO2_CDBLEN);
|
|
atiop->ccb_h.status = CAM_CDB_RECVD;
|
|
atiop->tag_id = aep->at_rxid;
|
|
switch (aep->at_taskflags & ATIO2_TC_ATTR_MASK) {
|
|
case ATIO2_TC_ATTR_SIMPLEQ:
|
|
atiop->tag_action = MSG_SIMPLE_Q_TAG;
|
|
break;
|
|
case ATIO2_TC_ATTR_HEADOFQ:
|
|
atiop->tag_action = MSG_HEAD_OF_Q_TAG;
|
|
break;
|
|
case ATIO2_TC_ATTR_ORDERED:
|
|
atiop->tag_action = MSG_ORDERED_Q_TAG;
|
|
break;
|
|
case ATIO2_TC_ATTR_ACAQ: /* ?? */
|
|
case ATIO2_TC_ATTR_UNTAGGED:
|
|
default:
|
|
atiop->tag_action = 0;
|
|
break;
|
|
}
|
|
atiop->ccb_h.flags = CAM_TAG_ACTION_VALID;
|
|
|
|
atp->tag = atiop->tag_id;
|
|
atp->lun = lun;
|
|
atp->orig_datalen = aep->at_datalen;
|
|
atp->last_xframt = 0;
|
|
atp->bytes_xfered = 0;
|
|
atp->state = ATPD_STATE_CAM;
|
|
xpt_done((union ccb*)atiop);
|
|
|
|
isp_prt(isp, ISP_LOGTDEBUG0,
|
|
"ATIO2[%x] CDB=0x%x iid%d->lun%d tattr 0x%x datalen %u",
|
|
aep->at_rxid, aep->at_cdb[0] & 0xff, aep->at_iid,
|
|
lun, aep->at_taskflags, aep->at_datalen);
|
|
rls_lun_statep(isp, tptr);
|
|
return (0);
|
|
}
|
|
|
|
static int
|
|
isp_handle_platform_ctio(struct ispsoftc *isp, void *arg)
|
|
{
|
|
union ccb *ccb;
|
|
int sentstatus, ok, notify_cam, resid = 0;
|
|
u_int16_t tval;
|
|
|
|
/*
|
|
* CTIO and CTIO2 are close enough....
|
|
*/
|
|
|
|
ccb = (union ccb *) isp_find_xs(isp, ((ct_entry_t *)arg)->ct_syshandle);
|
|
KASSERT((ccb != NULL), ("null ccb in isp_handle_platform_ctio"));
|
|
isp_destroy_handle(isp, ((ct_entry_t *)arg)->ct_syshandle);
|
|
|
|
if (IS_FC(isp)) {
|
|
ct2_entry_t *ct = arg;
|
|
atio_private_data_t *atp = isp_get_atpd(isp, ct->ct_rxid);
|
|
if (atp == NULL) {
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"cannot find adjunct for %x after I/O",
|
|
ct->ct_rxid);
|
|
return (0);
|
|
}
|
|
sentstatus = ct->ct_flags & CT2_SENDSTATUS;
|
|
ok = (ct->ct_status & ~QLTM_SVALID) == CT_OK;
|
|
if (ok && sentstatus && (ccb->ccb_h.flags & CAM_SEND_SENSE)) {
|
|
ccb->ccb_h.status |= CAM_SENT_SENSE;
|
|
}
|
|
notify_cam = ct->ct_header.rqs_seqno & 0x1;
|
|
if ((ct->ct_flags & CT2_DATAMASK) != CT2_NO_DATA) {
|
|
resid = ct->ct_resid;
|
|
atp->bytes_xfered += (atp->last_xframt - resid);
|
|
atp->last_xframt = 0;
|
|
}
|
|
if (sentstatus || !ok) {
|
|
atp->tag = 0;
|
|
}
|
|
isp_prt(isp, ok? ISP_LOGTDEBUG0 : ISP_LOGWARN,
|
|
"CTIO2[%x] sts 0x%x flg 0x%x sns %d resid %d %s",
|
|
ct->ct_rxid, ct->ct_status, ct->ct_flags,
|
|
(ccb->ccb_h.status & CAM_SENT_SENSE) != 0,
|
|
resid, sentstatus? "FIN" : "MID");
|
|
tval = ct->ct_rxid;
|
|
|
|
/* XXX: should really come after isp_complete_ctio */
|
|
atp->state = ATPD_STATE_PDON;
|
|
} else {
|
|
ct_entry_t *ct = arg;
|
|
sentstatus = ct->ct_flags & CT_SENDSTATUS;
|
|
ok = (ct->ct_status & ~QLTM_SVALID) == CT_OK;
|
|
/*
|
|
* We *ought* to be able to get back to the original ATIO
|
|
* here, but for some reason this gets lost. It's just as
|
|
* well because it's squirrelled away as part of periph
|
|
* private data.
|
|
*
|
|
* We can live without it as long as we continue to use
|
|
* the auto-replenish feature for CTIOs.
|
|
*/
|
|
notify_cam = ct->ct_header.rqs_seqno & 0x1;
|
|
if (ct->ct_status & QLTM_SVALID) {
|
|
char *sp = (char *)ct;
|
|
sp += CTIO_SENSE_OFFSET;
|
|
ccb->csio.sense_len =
|
|
min(sizeof (ccb->csio.sense_data), QLTM_SENSELEN);
|
|
MEMCPY(&ccb->csio.sense_data, sp, ccb->csio.sense_len);
|
|
ccb->ccb_h.status |= CAM_AUTOSNS_VALID;
|
|
}
|
|
if ((ct->ct_flags & CT_DATAMASK) != CT_NO_DATA) {
|
|
resid = ct->ct_resid;
|
|
}
|
|
isp_prt(isp, ISP_LOGTDEBUG0,
|
|
"CTIO[%x] tag %x iid %d lun %d sts %x flg %x resid %d %s",
|
|
ct->ct_fwhandle, ct->ct_tag_val, ct->ct_iid, ct->ct_lun,
|
|
ct->ct_status, ct->ct_flags, resid,
|
|
sentstatus? "FIN" : "MID");
|
|
tval = ct->ct_fwhandle;
|
|
}
|
|
ccb->csio.resid += resid;
|
|
|
|
/*
|
|
* We're here either because intermediate data transfers are done
|
|
* and/or the final status CTIO (which may have joined with a
|
|
* Data Transfer) is done.
|
|
*
|
|
* In any case, for this platform, the upper layers figure out
|
|
* what to do next, so all we do here is collect status and
|
|
* pass information along. Any DMA handles have already been
|
|
* freed.
|
|
*/
|
|
if (notify_cam == 0) {
|
|
isp_prt(isp, ISP_LOGTDEBUG0, " INTER CTIO[0x%x] done", tval);
|
|
return (0);
|
|
}
|
|
|
|
isp_prt(isp, ISP_LOGTDEBUG0, "%s CTIO[0x%x] done",
|
|
(sentstatus)? " FINAL " : "MIDTERM ", tval);
|
|
|
|
if (!ok) {
|
|
isp_target_putback_atio(ccb);
|
|
} else {
|
|
isp_complete_ctio(ccb);
|
|
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
static int
|
|
isp_handle_platform_notify_scsi(struct ispsoftc *isp, in_entry_t *inp)
|
|
{
|
|
return (0); /* XXXX */
|
|
}
|
|
|
|
static int
|
|
isp_handle_platform_notify_fc(struct ispsoftc *isp, in_fcentry_t *inp)
|
|
{
|
|
|
|
switch (inp->in_status) {
|
|
case IN_PORT_LOGOUT:
|
|
isp_prt(isp, ISP_LOGWARN, "port logout of iid %d",
|
|
inp->in_iid);
|
|
break;
|
|
case IN_PORT_CHANGED:
|
|
isp_prt(isp, ISP_LOGWARN, "port changed for iid %d",
|
|
inp->in_iid);
|
|
break;
|
|
case IN_GLOBAL_LOGO:
|
|
isp_prt(isp, ISP_LOGINFO, "all ports logged out");
|
|
break;
|
|
case IN_ABORT_TASK:
|
|
{
|
|
atio_private_data_t *atp = isp_get_atpd(isp, inp->in_seqid);
|
|
struct ccb_immed_notify *inot = NULL;
|
|
|
|
if (atp) {
|
|
tstate_t *tptr = get_lun_statep(isp, 0, atp->lun);
|
|
if (tptr) {
|
|
inot = (struct ccb_immed_notify *)
|
|
SLIST_FIRST(&tptr->inots);
|
|
if (inot) {
|
|
SLIST_REMOVE_HEAD(&tptr->inots,
|
|
sim_links.sle);
|
|
}
|
|
}
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"abort task RX_ID %x IID %d state %d",
|
|
inp->in_seqid, inp->in_iid, atp->state);
|
|
} else {
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"abort task RX_ID %x from iid %d, state unknown",
|
|
inp->in_seqid, inp->in_iid);
|
|
}
|
|
if (inot) {
|
|
inot->initiator_id = inp->in_iid;
|
|
inot->sense_len = 0;
|
|
inot->message_args[0] = MSG_ABORT_TAG;
|
|
inot->message_args[1] = inp->in_seqid & 0xff;
|
|
inot->message_args[2] = (inp->in_seqid >> 8) & 0xff;
|
|
inot->ccb_h.status = CAM_MESSAGE_RECV|CAM_DEV_QFRZN;
|
|
xpt_done((union ccb *)inot);
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
break;
|
|
}
|
|
return (0);
|
|
}
|
|
#endif
|
|
|
|
static void
|
|
isp_cam_async(void *cbarg, u_int32_t code, struct cam_path *path, void *arg)
|
|
{
|
|
struct cam_sim *sim;
|
|
struct ispsoftc *isp;
|
|
|
|
sim = (struct cam_sim *)cbarg;
|
|
isp = (struct ispsoftc *) cam_sim_softc(sim);
|
|
switch (code) {
|
|
case AC_LOST_DEVICE:
|
|
if (IS_SCSI(isp)) {
|
|
u_int16_t oflags, nflags;
|
|
sdparam *sdp = isp->isp_param;
|
|
int tgt;
|
|
|
|
tgt = xpt_path_target_id(path);
|
|
ISP_LOCK(isp);
|
|
sdp += cam_sim_bus(sim);
|
|
nflags = sdp->isp_devparam[tgt].nvrm_flags;
|
|
#ifndef ISP_TARGET_MODE
|
|
nflags &= DPARM_SAFE_DFLT;
|
|
if (isp->isp_loaded_fw) {
|
|
nflags |= DPARM_NARROW | DPARM_ASYNC;
|
|
}
|
|
#else
|
|
nflags = DPARM_DEFAULT;
|
|
#endif
|
|
oflags = sdp->isp_devparam[tgt].goal_flags;
|
|
sdp->isp_devparam[tgt].goal_flags = nflags;
|
|
sdp->isp_devparam[tgt].dev_update = 1;
|
|
isp->isp_update |= (1 << cam_sim_bus(sim));
|
|
(void) isp_control(isp, ISPCTL_UPDATE_PARAMS, NULL);
|
|
sdp->isp_devparam[tgt].goal_flags = oflags;
|
|
ISP_UNLOCK(isp);
|
|
}
|
|
break;
|
|
default:
|
|
isp_prt(isp, ISP_LOGWARN, "isp_cam_async: Code 0x%x", code);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
isp_poll(struct cam_sim *sim)
|
|
{
|
|
struct ispsoftc *isp = cam_sim_softc(sim);
|
|
u_int16_t isr, sema, mbox;
|
|
|
|
ISP_LOCK(isp);
|
|
if (ISP_READ_ISR(isp, &isr, &sema, &mbox)) {
|
|
isp_intr(isp, isr, sema, mbox);
|
|
}
|
|
ISP_UNLOCK(isp);
|
|
}
|
|
|
|
|
|
static void
|
|
isp_watchdog(void *arg)
|
|
{
|
|
XS_T *xs = arg;
|
|
struct ispsoftc *isp = XS_ISP(xs);
|
|
u_int32_t handle;
|
|
int iok;
|
|
|
|
/*
|
|
* We've decided this command is dead. Make sure we're not trying
|
|
* to kill a command that's already dead by getting it's handle and
|
|
* and seeing whether it's still alive.
|
|
*/
|
|
ISP_LOCK(isp);
|
|
iok = isp->isp_osinfo.intsok;
|
|
isp->isp_osinfo.intsok = 0;
|
|
handle = isp_find_handle(isp, xs);
|
|
if (handle) {
|
|
u_int16_t isr, sema, mbox;
|
|
|
|
if (XS_CMD_DONE_P(xs)) {
|
|
isp_prt(isp, ISP_LOGDEBUG1,
|
|
"watchdog found done cmd (handle 0x%x)", handle);
|
|
ISP_UNLOCK(isp);
|
|
return;
|
|
}
|
|
|
|
if (XS_CMD_WDOG_P(xs)) {
|
|
isp_prt(isp, ISP_LOGDEBUG2,
|
|
"recursive watchdog (handle 0x%x)", handle);
|
|
ISP_UNLOCK(isp);
|
|
return;
|
|
}
|
|
|
|
XS_CMD_S_WDOG(xs);
|
|
if (ISP_READ_ISR(isp, &isr, &sema, &mbox)) {
|
|
isp_intr(isp, isr, sema, mbox);
|
|
}
|
|
if (XS_CMD_DONE_P(xs)) {
|
|
isp_prt(isp, ISP_LOGDEBUG2,
|
|
"watchdog cleanup for handle 0x%x", handle);
|
|
xpt_done((union ccb *) xs);
|
|
} else if (XS_CMD_GRACE_P(xs)) {
|
|
/*
|
|
* Make sure the command is *really* dead before we
|
|
* release the handle (and DMA resources) for reuse.
|
|
*/
|
|
(void) isp_control(isp, ISPCTL_ABORT_CMD, arg);
|
|
|
|
/*
|
|
* After this point, the comamnd is really dead.
|
|
*/
|
|
if (XS_XFRLEN(xs)) {
|
|
ISP_DMAFREE(isp, xs, handle);
|
|
}
|
|
isp_destroy_handle(isp, handle);
|
|
xpt_print_path(xs->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"watchdog timeout for handle 0x%x", handle);
|
|
XS_SETERR(xs, CAM_CMD_TIMEOUT);
|
|
XS_CMD_C_WDOG(xs);
|
|
isp_done(xs);
|
|
} else {
|
|
u_int16_t nxti, optr;
|
|
ispreq_t local, *mp= &local, *qe;
|
|
|
|
XS_CMD_C_WDOG(xs);
|
|
xs->ccb_h.timeout_ch = timeout(isp_watchdog, xs, hz);
|
|
if (isp_getrqentry(isp, &nxti, &optr, (void **) &qe)) {
|
|
ISP_UNLOCK(isp);
|
|
return;
|
|
}
|
|
XS_CMD_S_GRACE(xs);
|
|
MEMZERO((void *) mp, sizeof (*mp));
|
|
mp->req_header.rqs_entry_count = 1;
|
|
mp->req_header.rqs_entry_type = RQSTYPE_MARKER;
|
|
mp->req_modifier = SYNC_ALL;
|
|
mp->req_target = XS_CHANNEL(xs) << 7;
|
|
isp_put_request(isp, mp, qe);
|
|
ISP_ADD_REQUEST(isp, nxti);
|
|
}
|
|
} else {
|
|
isp_prt(isp, ISP_LOGDEBUG2, "watchdog with no command");
|
|
}
|
|
isp->isp_osinfo.intsok = iok;
|
|
ISP_UNLOCK(isp);
|
|
}
|
|
|
|
static void
|
|
isp_kthread(void *arg)
|
|
{
|
|
struct ispsoftc *isp = arg;
|
|
|
|
mtx_lock(&isp->isp_lock);
|
|
/*
|
|
* The first loop is for our usage where we have yet to have
|
|
* gotten good fibre channel state.
|
|
*/
|
|
for (;;) {
|
|
int wasfrozen;
|
|
|
|
isp_prt(isp, ISP_LOGDEBUG0, "kthread: checking FC state");
|
|
while (isp_fc_runstate(isp, 2 * 1000000) != 0) {
|
|
isp_prt(isp, ISP_LOGDEBUG0, "kthread: FC state ungood");
|
|
if (FCPARAM(isp)->isp_fwstate != FW_READY ||
|
|
FCPARAM(isp)->isp_loopstate < LOOP_PDB_RCVD) {
|
|
if (FCPARAM(isp)->loop_seen_once == 0 ||
|
|
isp->isp_osinfo.ktmature == 0) {
|
|
break;
|
|
}
|
|
}
|
|
msleep(isp_kthread, &isp->isp_lock,
|
|
PRIBIO, "isp_fcthrd", hz);
|
|
}
|
|
|
|
/*
|
|
* Even if we didn't get good loop state we may be
|
|
* unfreezing the SIMQ so that we can kill off
|
|
* commands (if we've never seen loop before, for example).
|
|
*/
|
|
isp->isp_osinfo.ktmature = 1;
|
|
wasfrozen = isp->isp_osinfo.simqfrozen & SIMQFRZ_LOOPDOWN;
|
|
isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_LOOPDOWN;
|
|
if (wasfrozen && isp->isp_osinfo.simqfrozen == 0) {
|
|
isp_prt(isp, ISP_LOGDEBUG0, "kthread: releasing simq");
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_release_simq(isp->isp_sim, 1);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
}
|
|
isp_prt(isp, ISP_LOGDEBUG0, "kthread: waiting until called");
|
|
cv_wait(&isp->isp_osinfo.kthread_cv, &isp->isp_lock);
|
|
}
|
|
}
|
|
|
|
static void
|
|
isp_action(struct cam_sim *sim, union ccb *ccb)
|
|
{
|
|
int bus, tgt, error;
|
|
struct ispsoftc *isp;
|
|
struct ccb_trans_settings *cts;
|
|
|
|
CAM_DEBUG(ccb->ccb_h.path, CAM_DEBUG_TRACE, ("isp_action\n"));
|
|
|
|
isp = (struct ispsoftc *)cam_sim_softc(sim);
|
|
ccb->ccb_h.sim_priv.entries[0].field = 0;
|
|
ccb->ccb_h.sim_priv.entries[1].ptr = isp;
|
|
if (isp->isp_state != ISP_RUNSTATE &&
|
|
ccb->ccb_h.func_code == XPT_SCSI_IO) {
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
isp_init(isp);
|
|
if (isp->isp_state != ISP_INITSTATE) {
|
|
ISP_UNLOCK(isp);
|
|
/*
|
|
* Lie. Say it was a selection timeout.
|
|
*/
|
|
ccb->ccb_h.status = CAM_SEL_TIMEOUT | CAM_DEV_QFRZN;
|
|
xpt_freeze_devq(ccb->ccb_h.path, 1);
|
|
xpt_done(ccb);
|
|
return;
|
|
}
|
|
isp->isp_state = ISP_RUNSTATE;
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
}
|
|
isp_prt(isp, ISP_LOGDEBUG2, "isp_action code %x", ccb->ccb_h.func_code);
|
|
|
|
|
|
switch (ccb->ccb_h.func_code) {
|
|
case XPT_SCSI_IO: /* Execute the requested I/O operation */
|
|
/*
|
|
* Do a couple of preliminary checks...
|
|
*/
|
|
if ((ccb->ccb_h.flags & CAM_CDB_POINTER) != 0) {
|
|
if ((ccb->ccb_h.flags & CAM_CDB_PHYS) != 0) {
|
|
ccb->ccb_h.status = CAM_REQ_INVALID;
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
}
|
|
#ifdef DIAGNOSTIC
|
|
if (ccb->ccb_h.target_id > (ISP_MAX_TARGETS(isp) - 1)) {
|
|
ccb->ccb_h.status = CAM_PATH_INVALID;
|
|
} else if (ccb->ccb_h.target_lun > (ISP_MAX_LUNS(isp) - 1)) {
|
|
ccb->ccb_h.status = CAM_PATH_INVALID;
|
|
}
|
|
if (ccb->ccb_h.status == CAM_PATH_INVALID) {
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"invalid tgt/lun (%d.%d) in XPT_SCSI_IO",
|
|
ccb->ccb_h.target_id, ccb->ccb_h.target_lun);
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
#endif
|
|
((struct ccb_scsiio *) ccb)->scsi_status = SCSI_STATUS_OK;
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
error = isp_start((XS_T *) ccb);
|
|
switch (error) {
|
|
case CMD_QUEUED:
|
|
ccb->ccb_h.status |= CAM_SIM_QUEUED;
|
|
if (ccb->ccb_h.timeout != CAM_TIME_INFINITY) {
|
|
u_int64_t ticks = (u_int64_t) hz;
|
|
if (ccb->ccb_h.timeout == CAM_TIME_DEFAULT)
|
|
ticks = 60 * 1000 * ticks;
|
|
else
|
|
ticks = ccb->ccb_h.timeout * hz;
|
|
ticks = ((ticks + 999) / 1000) + hz + hz;
|
|
if (ticks >= 0x80000000) {
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"timeout overflow");
|
|
ticks = 0x7fffffff;
|
|
}
|
|
ccb->ccb_h.timeout_ch = timeout(isp_watchdog,
|
|
(caddr_t)ccb, (int)ticks);
|
|
} else {
|
|
callout_handle_init(&ccb->ccb_h.timeout_ch);
|
|
}
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
break;
|
|
case CMD_RQLATER:
|
|
/*
|
|
* This can only happen for Fibre Channel
|
|
*/
|
|
KASSERT((IS_FC(isp)), ("CMD_RQLATER for FC only"));
|
|
if (FCPARAM(isp)->loop_seen_once == 0 &&
|
|
isp->isp_osinfo.ktmature) {
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
XS_SETERR(ccb, CAM_SEL_TIMEOUT);
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
cv_signal(&isp->isp_osinfo.kthread_cv);
|
|
isp_freeze_loopdown(isp, "isp_action(RQLATER)");
|
|
isp->isp_osinfo.simqfrozen |= SIMQFRZ_LOOPDOWN;
|
|
XS_SETERR(ccb, CAM_REQUEUE_REQ);
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_done(ccb);
|
|
break;
|
|
case CMD_EAGAIN:
|
|
XS_SETERR(ccb, CAM_REQUEUE_REQ);
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_done(ccb);
|
|
break;
|
|
case CMD_COMPLETE:
|
|
isp_done((struct ccb_scsiio *) ccb);
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
break;
|
|
default:
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"What's this? 0x%x at %d in file %s",
|
|
error, __LINE__, __FILE__);
|
|
XS_SETERR(ccb, CAM_REQ_CMP_ERR);
|
|
xpt_done(ccb);
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
}
|
|
break;
|
|
|
|
#ifdef ISP_TARGET_MODE
|
|
case XPT_EN_LUN: /* Enable LUN as a target */
|
|
{
|
|
int iok;
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
iok = isp->isp_osinfo.intsok;
|
|
isp->isp_osinfo.intsok = 0;
|
|
isp_en_lun(isp, ccb);
|
|
isp->isp_osinfo.intsok = iok;
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
case XPT_NOTIFY_ACK: /* recycle notify ack */
|
|
case XPT_IMMED_NOTIFY: /* Add Immediate Notify Resource */
|
|
case XPT_ACCEPT_TARGET_IO: /* Add Accept Target IO Resource */
|
|
{
|
|
tstate_t *tptr =
|
|
get_lun_statep(isp, XS_CHANNEL(ccb), ccb->ccb_h.target_lun);
|
|
if (tptr == NULL) {
|
|
ccb->ccb_h.status = CAM_LUN_INVALID;
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
ccb->ccb_h.sim_priv.entries[0].field = 0;
|
|
ccb->ccb_h.sim_priv.entries[1].ptr = isp;
|
|
ccb->ccb_h.flags = 0;
|
|
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
if (ccb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) {
|
|
/*
|
|
* Note that the command itself may not be done-
|
|
* it may not even have had the first CTIO sent.
|
|
*/
|
|
tptr->atio_count++;
|
|
isp_prt(isp, ISP_LOGTDEBUG0,
|
|
"Put FREE ATIO2, lun %d, count now %d",
|
|
ccb->ccb_h.target_lun, tptr->atio_count);
|
|
SLIST_INSERT_HEAD(&tptr->atios, &ccb->ccb_h,
|
|
sim_links.sle);
|
|
} else if (ccb->ccb_h.func_code == XPT_IMMED_NOTIFY) {
|
|
SLIST_INSERT_HEAD(&tptr->inots, &ccb->ccb_h,
|
|
sim_links.sle);
|
|
} else {
|
|
;
|
|
}
|
|
rls_lun_statep(isp, tptr);
|
|
ccb->ccb_h.status = CAM_REQ_INPROG;
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
break;
|
|
}
|
|
case XPT_CONT_TARGET_IO:
|
|
{
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
ccb->ccb_h.status = isp_target_start_ctio(isp, ccb);
|
|
if (ccb->ccb_h.status != CAM_REQ_INPROG) {
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"XPT_CONT_TARGET_IO: status 0x%x",
|
|
ccb->ccb_h.status);
|
|
XS_SETERR(ccb, CAM_REQUEUE_REQ);
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_done(ccb);
|
|
} else {
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
ccb->ccb_h.status |= CAM_SIM_QUEUED;
|
|
}
|
|
break;
|
|
}
|
|
#endif
|
|
case XPT_RESET_DEV: /* BDR the specified SCSI device */
|
|
|
|
bus = cam_sim_bus(xpt_path_sim(ccb->ccb_h.path));
|
|
tgt = ccb->ccb_h.target_id;
|
|
tgt |= (bus << 16);
|
|
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
error = isp_control(isp, ISPCTL_RESET_DEV, &tgt);
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
if (error) {
|
|
ccb->ccb_h.status = CAM_REQ_CMP_ERR;
|
|
} else {
|
|
ccb->ccb_h.status = CAM_REQ_CMP;
|
|
}
|
|
xpt_done(ccb);
|
|
break;
|
|
case XPT_ABORT: /* Abort the specified CCB */
|
|
{
|
|
union ccb *accb = ccb->cab.abort_ccb;
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
switch (accb->ccb_h.func_code) {
|
|
#ifdef ISP_TARGET_MODE
|
|
case XPT_ACCEPT_TARGET_IO:
|
|
case XPT_IMMED_NOTIFY:
|
|
ccb->ccb_h.status = isp_abort_tgt_ccb(isp, ccb);
|
|
break;
|
|
case XPT_CONT_TARGET_IO:
|
|
isp_prt(isp, ISP_LOGERR, "cannot abort CTIOs yet");
|
|
ccb->ccb_h.status = CAM_UA_ABORT;
|
|
break;
|
|
#endif
|
|
case XPT_SCSI_IO:
|
|
error = isp_control(isp, ISPCTL_ABORT_CMD, ccb);
|
|
if (error) {
|
|
ccb->ccb_h.status = CAM_UA_ABORT;
|
|
} else {
|
|
ccb->ccb_h.status = CAM_REQ_CMP;
|
|
}
|
|
break;
|
|
default:
|
|
ccb->ccb_h.status = CAM_REQ_INVALID;
|
|
break;
|
|
}
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
#ifdef CAM_NEW_TRAN_CODE
|
|
#define IS_CURRENT_SETTINGS(c) (c->type == CTS_TYPE_CURRENT_SETTINGS)
|
|
#else
|
|
#define IS_CURRENT_SETTINGS(c) (c->flags & CCB_TRANS_CURRENT_SETTINGS)
|
|
#endif
|
|
case XPT_SET_TRAN_SETTINGS: /* Nexus Settings */
|
|
cts = &ccb->cts;
|
|
if (!IS_CURRENT_SETTINGS(cts)) {
|
|
ccb->ccb_h.status = CAM_REQ_INVALID;
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
tgt = cts->ccb_h.target_id;
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
if (IS_SCSI(isp)) {
|
|
#ifndef CAM_NEW_TRAN_CODE
|
|
sdparam *sdp = isp->isp_param;
|
|
u_int16_t *dptr;
|
|
|
|
bus = cam_sim_bus(xpt_path_sim(cts->ccb_h.path));
|
|
|
|
sdp += bus;
|
|
/*
|
|
* We always update (internally) from goal_flags
|
|
* so any request to change settings just gets
|
|
* vectored to that location.
|
|
*/
|
|
dptr = &sdp->isp_devparam[tgt].goal_flags;
|
|
|
|
/*
|
|
* Note that these operations affect the
|
|
* the goal flags (goal_flags)- not
|
|
* the current state flags. Then we mark
|
|
* things so that the next operation to
|
|
* this HBA will cause the update to occur.
|
|
*/
|
|
if (cts->valid & CCB_TRANS_DISC_VALID) {
|
|
if ((cts->flags & CCB_TRANS_DISC_ENB) != 0) {
|
|
*dptr |= DPARM_DISC;
|
|
} else {
|
|
*dptr &= ~DPARM_DISC;
|
|
}
|
|
}
|
|
if (cts->valid & CCB_TRANS_TQ_VALID) {
|
|
if ((cts->flags & CCB_TRANS_TAG_ENB) != 0) {
|
|
*dptr |= DPARM_TQING;
|
|
} else {
|
|
*dptr &= ~DPARM_TQING;
|
|
}
|
|
}
|
|
if (cts->valid & CCB_TRANS_BUS_WIDTH_VALID) {
|
|
switch (cts->bus_width) {
|
|
case MSG_EXT_WDTR_BUS_16_BIT:
|
|
*dptr |= DPARM_WIDE;
|
|
break;
|
|
default:
|
|
*dptr &= ~DPARM_WIDE;
|
|
}
|
|
}
|
|
/*
|
|
* Any SYNC RATE of nonzero and SYNC_OFFSET
|
|
* of nonzero will cause us to go to the
|
|
* selected (from NVRAM) maximum value for
|
|
* this device. At a later point, we'll
|
|
* allow finer control.
|
|
*/
|
|
if ((cts->valid & CCB_TRANS_SYNC_RATE_VALID) &&
|
|
(cts->valid & CCB_TRANS_SYNC_OFFSET_VALID) &&
|
|
(cts->sync_offset > 0)) {
|
|
*dptr |= DPARM_SYNC;
|
|
} else {
|
|
*dptr &= ~DPARM_SYNC;
|
|
}
|
|
*dptr |= DPARM_SAFE_DFLT;
|
|
#else
|
|
struct ccb_trans_settings_scsi *scsi =
|
|
&cts->proto_specific.scsi;
|
|
struct ccb_trans_settings_spi *spi =
|
|
&cts->xport_specific.spi;
|
|
sdparam *sdp = isp->isp_param;
|
|
u_int16_t *dptr;
|
|
|
|
bus = cam_sim_bus(xpt_path_sim(cts->ccb_h.path));
|
|
sdp += bus;
|
|
/*
|
|
* We always update (internally) from goal_flags
|
|
* so any request to change settings just gets
|
|
* vectored to that location.
|
|
*/
|
|
dptr = &sdp->isp_devparam[tgt].goal_flags;
|
|
|
|
if ((spi->valid & CTS_SPI_VALID_DISC) != 0) {
|
|
if ((spi->flags & CTS_SPI_FLAGS_DISC_ENB) != 0)
|
|
*dptr |= DPARM_DISC;
|
|
else
|
|
*dptr &= ~DPARM_DISC;
|
|
}
|
|
|
|
if ((scsi->valid & CTS_SCSI_VALID_TQ) != 0) {
|
|
if ((scsi->flags & CTS_SCSI_FLAGS_TAG_ENB) != 0)
|
|
*dptr |= DPARM_TQING;
|
|
else
|
|
*dptr &= ~DPARM_TQING;
|
|
}
|
|
|
|
if ((spi->valid & CTS_SPI_VALID_BUS_WIDTH) != 0) {
|
|
if (spi->bus_width == MSG_EXT_WDTR_BUS_16_BIT)
|
|
*dptr |= DPARM_WIDE;
|
|
else
|
|
*dptr &= ~DPARM_WIDE;
|
|
}
|
|
|
|
/*
|
|
* XXX: FIX ME
|
|
*/
|
|
if ((spi->valid & CTS_SPI_VALID_SYNC_OFFSET) &&
|
|
(spi->valid & CTS_SPI_VALID_SYNC_RATE) &&
|
|
(spi->sync_period && spi->sync_offset)) {
|
|
*dptr |= DPARM_SYNC;
|
|
/*
|
|
* XXX: CHECK FOR LEGALITY
|
|
*/
|
|
sdp->isp_devparam[tgt].goal_period =
|
|
spi->sync_period;
|
|
sdp->isp_devparam[tgt].goal_offset =
|
|
spi->sync_offset;
|
|
} else {
|
|
*dptr &= ~DPARM_SYNC;
|
|
}
|
|
#endif
|
|
isp_prt(isp, ISP_LOGDEBUG0,
|
|
"SET bus %d targ %d to flags %x off %x per %x",
|
|
bus, tgt, sdp->isp_devparam[tgt].goal_flags,
|
|
sdp->isp_devparam[tgt].goal_offset,
|
|
sdp->isp_devparam[tgt].goal_period);
|
|
sdp->isp_devparam[tgt].dev_update = 1;
|
|
isp->isp_update |= (1 << bus);
|
|
}
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
ccb->ccb_h.status = CAM_REQ_CMP;
|
|
xpt_done(ccb);
|
|
break;
|
|
case XPT_GET_TRAN_SETTINGS:
|
|
cts = &ccb->cts;
|
|
tgt = cts->ccb_h.target_id;
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
if (IS_FC(isp)) {
|
|
#ifndef CAM_NEW_TRAN_CODE
|
|
/*
|
|
* a lot of normal SCSI things don't make sense.
|
|
*/
|
|
cts->flags = CCB_TRANS_TAG_ENB | CCB_TRANS_DISC_ENB;
|
|
cts->valid = CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID;
|
|
/*
|
|
* How do you measure the width of a high
|
|
* speed serial bus? Well, in bytes.
|
|
*
|
|
* Offset and period make no sense, though, so we set
|
|
* (above) a 'base' transfer speed to be gigabit.
|
|
*/
|
|
cts->bus_width = MSG_EXT_WDTR_BUS_8_BIT;
|
|
#else
|
|
fcparam *fcp = isp->isp_param;
|
|
struct ccb_trans_settings_fc *fc =
|
|
&cts->xport_specific.fc;
|
|
|
|
cts->protocol = PROTO_SCSI;
|
|
cts->protocol_version = SCSI_REV_2;
|
|
cts->transport = XPORT_FC;
|
|
cts->transport_version = 0;
|
|
|
|
fc->valid = CTS_FC_VALID_SPEED;
|
|
if (fcp->isp_gbspeed == 2)
|
|
fc->bitrate = 200000;
|
|
else
|
|
fc->bitrate = 100000;
|
|
if (tgt > 0 && tgt < MAX_FC_TARG) {
|
|
struct lportdb *lp = &fcp->portdb[tgt];
|
|
fc->wwnn = lp->node_wwn;
|
|
fc->wwpn = lp->port_wwn;
|
|
fc->port = lp->portid;
|
|
fc->valid |= CTS_FC_VALID_WWNN |
|
|
CTS_FC_VALID_WWPN | CTS_FC_VALID_PORT;
|
|
}
|
|
#endif
|
|
} else {
|
|
#ifdef CAM_NEW_TRAN_CODE
|
|
struct ccb_trans_settings_scsi *scsi =
|
|
&cts->proto_specific.scsi;
|
|
struct ccb_trans_settings_spi *spi =
|
|
&cts->xport_specific.spi;
|
|
#endif
|
|
sdparam *sdp = isp->isp_param;
|
|
int bus = cam_sim_bus(xpt_path_sim(cts->ccb_h.path));
|
|
u_int16_t dval, pval, oval;
|
|
|
|
sdp += bus;
|
|
|
|
if (IS_CURRENT_SETTINGS(cts)) {
|
|
sdp->isp_devparam[tgt].dev_refresh = 1;
|
|
isp->isp_update |= (1 << bus);
|
|
(void) isp_control(isp, ISPCTL_UPDATE_PARAMS,
|
|
NULL);
|
|
dval = sdp->isp_devparam[tgt].actv_flags;
|
|
oval = sdp->isp_devparam[tgt].actv_offset;
|
|
pval = sdp->isp_devparam[tgt].actv_period;
|
|
} else {
|
|
dval = sdp->isp_devparam[tgt].nvrm_flags;
|
|
oval = sdp->isp_devparam[tgt].nvrm_offset;
|
|
pval = sdp->isp_devparam[tgt].nvrm_period;
|
|
}
|
|
|
|
#ifndef CAM_NEW_TRAN_CODE
|
|
cts->flags &= ~(CCB_TRANS_DISC_ENB|CCB_TRANS_TAG_ENB);
|
|
|
|
if (dval & DPARM_DISC) {
|
|
cts->flags |= CCB_TRANS_DISC_ENB;
|
|
}
|
|
if (dval & DPARM_TQING) {
|
|
cts->flags |= CCB_TRANS_TAG_ENB;
|
|
}
|
|
if (dval & DPARM_WIDE) {
|
|
cts->bus_width = MSG_EXT_WDTR_BUS_16_BIT;
|
|
} else {
|
|
cts->bus_width = MSG_EXT_WDTR_BUS_8_BIT;
|
|
}
|
|
cts->valid = CCB_TRANS_BUS_WIDTH_VALID |
|
|
CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID;
|
|
|
|
if ((dval & DPARM_SYNC) && oval != 0) {
|
|
cts->sync_period = pval;
|
|
cts->sync_offset = oval;
|
|
cts->valid |=
|
|
CCB_TRANS_SYNC_RATE_VALID |
|
|
CCB_TRANS_SYNC_OFFSET_VALID;
|
|
}
|
|
#else
|
|
cts->protocol = PROTO_SCSI;
|
|
cts->protocol_version = SCSI_REV_2;
|
|
cts->transport = XPORT_SPI;
|
|
cts->transport_version = 2;
|
|
|
|
scsi->flags &= ~CTS_SCSI_FLAGS_TAG_ENB;
|
|
spi->flags &= ~CTS_SPI_FLAGS_DISC_ENB;
|
|
if (dval & DPARM_DISC) {
|
|
spi->flags |= CTS_SPI_FLAGS_DISC_ENB;
|
|
}
|
|
if (dval & DPARM_TQING) {
|
|
scsi->flags |= CTS_SCSI_FLAGS_TAG_ENB;
|
|
}
|
|
if ((dval & DPARM_SYNC) && oval && pval) {
|
|
spi->sync_offset = oval;
|
|
spi->sync_period = pval;
|
|
spi->valid |= CTS_SPI_VALID_SYNC_OFFSET;
|
|
spi->valid |= CTS_SPI_VALID_SYNC_RATE;
|
|
}
|
|
spi->valid |= CTS_SPI_VALID_BUS_WIDTH;
|
|
if (dval & DPARM_WIDE) {
|
|
spi->bus_width = MSG_EXT_WDTR_BUS_16_BIT;
|
|
} else {
|
|
spi->bus_width = MSG_EXT_WDTR_BUS_8_BIT;
|
|
}
|
|
if (cts->ccb_h.target_lun != CAM_LUN_WILDCARD) {
|
|
scsi->valid = CTS_SCSI_VALID_TQ;
|
|
spi->valid |= CTS_SPI_VALID_DISC;
|
|
} else {
|
|
scsi->valid = 0;
|
|
}
|
|
#endif
|
|
isp_prt(isp, ISP_LOGDEBUG0,
|
|
"GET %s bus %d targ %d to flags %x off %x per %x",
|
|
IS_CURRENT_SETTINGS(cts)? "ACTIVE" : "NVRAM",
|
|
bus, tgt, dval, oval, pval);
|
|
}
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
ccb->ccb_h.status = CAM_REQ_CMP;
|
|
xpt_done(ccb);
|
|
break;
|
|
|
|
case XPT_CALC_GEOMETRY:
|
|
{
|
|
struct ccb_calc_geometry *ccg;
|
|
u_int32_t secs_per_cylinder;
|
|
u_int32_t size_mb;
|
|
|
|
ccg = &ccb->ccg;
|
|
if (ccg->block_size == 0) {
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"%d.%d XPT_CALC_GEOMETRY block size 0?",
|
|
ccg->ccb_h.target_id, ccg->ccb_h.target_lun);
|
|
ccb->ccb_h.status = CAM_REQ_INVALID;
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
size_mb = ccg->volume_size /((1024L * 1024L) / ccg->block_size);
|
|
if (size_mb > 1024) {
|
|
ccg->heads = 255;
|
|
ccg->secs_per_track = 63;
|
|
} else {
|
|
ccg->heads = 64;
|
|
ccg->secs_per_track = 32;
|
|
}
|
|
secs_per_cylinder = ccg->heads * ccg->secs_per_track;
|
|
ccg->cylinders = ccg->volume_size / secs_per_cylinder;
|
|
ccb->ccb_h.status = CAM_REQ_CMP;
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
case XPT_RESET_BUS: /* Reset the specified bus */
|
|
bus = cam_sim_bus(sim);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
error = isp_control(isp, ISPCTL_RESET_BUS, &bus);
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
if (error)
|
|
ccb->ccb_h.status = CAM_REQ_CMP_ERR;
|
|
else {
|
|
if (cam_sim_bus(sim) && isp->isp_path2 != NULL)
|
|
xpt_async(AC_BUS_RESET, isp->isp_path2, NULL);
|
|
else if (isp->isp_path != NULL)
|
|
xpt_async(AC_BUS_RESET, isp->isp_path, NULL);
|
|
ccb->ccb_h.status = CAM_REQ_CMP;
|
|
}
|
|
xpt_done(ccb);
|
|
break;
|
|
|
|
case XPT_TERM_IO: /* Terminate the I/O process */
|
|
ccb->ccb_h.status = CAM_REQ_INVALID;
|
|
xpt_done(ccb);
|
|
break;
|
|
|
|
case XPT_PATH_INQ: /* Path routing inquiry */
|
|
{
|
|
struct ccb_pathinq *cpi = &ccb->cpi;
|
|
|
|
cpi->version_num = 1;
|
|
#ifdef ISP_TARGET_MODE
|
|
cpi->target_sprt = PIT_PROCESSOR | PIT_DISCONNECT | PIT_TERM_IO;
|
|
#else
|
|
cpi->target_sprt = 0;
|
|
#endif
|
|
cpi->hba_eng_cnt = 0;
|
|
cpi->max_target = ISP_MAX_TARGETS(isp) - 1;
|
|
cpi->max_lun = ISP_MAX_LUNS(isp) - 1;
|
|
cpi->bus_id = cam_sim_bus(sim);
|
|
if (IS_FC(isp)) {
|
|
cpi->hba_misc = PIM_NOBUSRESET;
|
|
/*
|
|
* Because our loop ID can shift from time to time,
|
|
* make our initiator ID out of range of our bus.
|
|
*/
|
|
cpi->initiator_id = cpi->max_target + 1;
|
|
|
|
/*
|
|
* Set base transfer capabilities for Fibre Channel.
|
|
* Technically not correct because we don't know
|
|
* what media we're running on top of- but we'll
|
|
* look good if we always say 100MB/s.
|
|
*/
|
|
if (FCPARAM(isp)->isp_gbspeed == 2)
|
|
cpi->base_transfer_speed = 200000;
|
|
else
|
|
cpi->base_transfer_speed = 100000;
|
|
cpi->hba_inquiry = PI_TAG_ABLE;
|
|
#ifdef CAM_NEW_TRAN_CODE
|
|
cpi->transport = XPORT_FC;
|
|
cpi->transport_version = 0; /* WHAT'S THIS FOR? */
|
|
#endif
|
|
} else {
|
|
sdparam *sdp = isp->isp_param;
|
|
sdp += cam_sim_bus(xpt_path_sim(cpi->ccb_h.path));
|
|
cpi->hba_inquiry = PI_SDTR_ABLE|PI_TAG_ABLE|PI_WIDE_16;
|
|
cpi->hba_misc = 0;
|
|
cpi->initiator_id = sdp->isp_initiator_id;
|
|
cpi->base_transfer_speed = 3300;
|
|
#ifdef CAM_NEW_TRAN_CODE
|
|
cpi->transport = XPORT_SPI;
|
|
cpi->transport_version = 2; /* WHAT'S THIS FOR? */
|
|
#endif
|
|
}
|
|
#ifdef CAM_NEW_TRAN_CODE
|
|
cpi->protocol = PROTO_SCSI;
|
|
cpi->protocol_version = SCSI_REV_2;
|
|
#endif
|
|
strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
|
|
strncpy(cpi->hba_vid, "Qlogic", HBA_IDLEN);
|
|
strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
|
|
cpi->unit_number = cam_sim_unit(sim);
|
|
cpi->ccb_h.status = CAM_REQ_CMP;
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
default:
|
|
ccb->ccb_h.status = CAM_REQ_INVALID;
|
|
xpt_done(ccb);
|
|
break;
|
|
}
|
|
}
|
|
|
|
#define ISPDDB (CAM_DEBUG_INFO|CAM_DEBUG_TRACE|CAM_DEBUG_CDB)
|
|
void
|
|
isp_done(struct ccb_scsiio *sccb)
|
|
{
|
|
struct ispsoftc *isp = XS_ISP(sccb);
|
|
|
|
if (XS_NOERR(sccb))
|
|
XS_SETERR(sccb, CAM_REQ_CMP);
|
|
|
|
if ((sccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP &&
|
|
(sccb->scsi_status != SCSI_STATUS_OK)) {
|
|
sccb->ccb_h.status &= ~CAM_STATUS_MASK;
|
|
if ((sccb->scsi_status == SCSI_STATUS_CHECK_COND) &&
|
|
(sccb->ccb_h.status & CAM_AUTOSNS_VALID) == 0) {
|
|
sccb->ccb_h.status |= CAM_AUTOSENSE_FAIL;
|
|
} else {
|
|
sccb->ccb_h.status |= CAM_SCSI_STATUS_ERROR;
|
|
}
|
|
}
|
|
|
|
sccb->ccb_h.status &= ~CAM_SIM_QUEUED;
|
|
if ((sccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) {
|
|
if ((sccb->ccb_h.status & CAM_DEV_QFRZN) == 0) {
|
|
sccb->ccb_h.status |= CAM_DEV_QFRZN;
|
|
xpt_freeze_devq(sccb->ccb_h.path, 1);
|
|
isp_prt(isp, ISP_LOGDEBUG0,
|
|
"freeze devq %d.%d cam sts %x scsi sts %x",
|
|
sccb->ccb_h.target_id, sccb->ccb_h.target_lun,
|
|
sccb->ccb_h.status, sccb->scsi_status);
|
|
}
|
|
}
|
|
|
|
if ((CAM_DEBUGGED(sccb->ccb_h.path, ISPDDB)) &&
|
|
(sccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) {
|
|
xpt_print_path(sccb->ccb_h.path);
|
|
isp_prt(isp, ISP_LOGINFO,
|
|
"cam completion status 0x%x", sccb->ccb_h.status);
|
|
}
|
|
|
|
XS_CMD_S_DONE(sccb);
|
|
if (XS_CMD_WDOG_P(sccb) == 0) {
|
|
untimeout(isp_watchdog, (caddr_t)sccb, sccb->ccb_h.timeout_ch);
|
|
if (XS_CMD_GRACE_P(sccb)) {
|
|
isp_prt(isp, ISP_LOGDEBUG2,
|
|
"finished command on borrowed time");
|
|
}
|
|
XS_CMD_S_CLEAR(sccb);
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_done((union ccb *) sccb);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
}
|
|
}
|
|
|
|
int
|
|
isp_async(struct ispsoftc *isp, ispasync_t cmd, void *arg)
|
|
{
|
|
int bus, rv = 0;
|
|
switch (cmd) {
|
|
case ISPASYNC_NEW_TGT_PARAMS:
|
|
{
|
|
#ifdef CAM_NEW_TRAN_CODE
|
|
struct ccb_trans_settings_scsi *scsi;
|
|
struct ccb_trans_settings_spi *spi;
|
|
#endif
|
|
int flags, tgt;
|
|
sdparam *sdp = isp->isp_param;
|
|
struct ccb_trans_settings cts;
|
|
struct cam_path *tmppath;
|
|
|
|
bzero(&cts, sizeof (struct ccb_trans_settings));
|
|
|
|
tgt = *((int *)arg);
|
|
bus = (tgt >> 16) & 0xffff;
|
|
tgt &= 0xffff;
|
|
sdp += bus;
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
if (xpt_create_path(&tmppath, NULL,
|
|
cam_sim_path(bus? isp->isp_sim2 : isp->isp_sim),
|
|
tgt, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"isp_async cannot make temp path for %d.%d",
|
|
tgt, bus);
|
|
rv = -1;
|
|
break;
|
|
}
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
flags = sdp->isp_devparam[tgt].actv_flags;
|
|
#ifdef CAM_NEW_TRAN_CODE
|
|
cts.type = CTS_TYPE_CURRENT_SETTINGS;
|
|
cts.protocol = PROTO_SCSI;
|
|
cts.transport = XPORT_SPI;
|
|
|
|
scsi = &cts.proto_specific.scsi;
|
|
spi = &cts.xport_specific.spi;
|
|
|
|
if (flags & DPARM_TQING) {
|
|
scsi->valid |= CTS_SCSI_VALID_TQ;
|
|
scsi->flags |= CTS_SCSI_FLAGS_TAG_ENB;
|
|
spi->flags |= CTS_SPI_FLAGS_TAG_ENB;
|
|
}
|
|
|
|
if (flags & DPARM_DISC) {
|
|
spi->valid |= CTS_SPI_VALID_DISC;
|
|
spi->flags |= CTS_SPI_FLAGS_DISC_ENB;
|
|
}
|
|
spi->flags |= CTS_SPI_VALID_BUS_WIDTH;
|
|
if (flags & DPARM_WIDE) {
|
|
spi->bus_width = MSG_EXT_WDTR_BUS_16_BIT;
|
|
} else {
|
|
spi->bus_width = MSG_EXT_WDTR_BUS_8_BIT;
|
|
}
|
|
if (flags & DPARM_SYNC) {
|
|
spi->valid |= CTS_SPI_VALID_SYNC_RATE;
|
|
spi->valid |= CTS_SPI_VALID_SYNC_OFFSET;
|
|
spi->sync_period = sdp->isp_devparam[tgt].actv_period;
|
|
spi->sync_offset = sdp->isp_devparam[tgt].actv_offset;
|
|
}
|
|
#else
|
|
cts.flags = CCB_TRANS_CURRENT_SETTINGS;
|
|
cts.valid = CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID;
|
|
if (flags & DPARM_DISC) {
|
|
cts.flags |= CCB_TRANS_DISC_ENB;
|
|
}
|
|
if (flags & DPARM_TQING) {
|
|
cts.flags |= CCB_TRANS_TAG_ENB;
|
|
}
|
|
cts.valid |= CCB_TRANS_BUS_WIDTH_VALID;
|
|
cts.bus_width = (flags & DPARM_WIDE)?
|
|
MSG_EXT_WDTR_BUS_8_BIT : MSG_EXT_WDTR_BUS_16_BIT;
|
|
cts.sync_period = sdp->isp_devparam[tgt].actv_period;
|
|
cts.sync_offset = sdp->isp_devparam[tgt].actv_offset;
|
|
if (flags & DPARM_SYNC) {
|
|
cts.valid |=
|
|
CCB_TRANS_SYNC_RATE_VALID |
|
|
CCB_TRANS_SYNC_OFFSET_VALID;
|
|
}
|
|
#endif
|
|
isp_prt(isp, ISP_LOGDEBUG2,
|
|
"NEW_TGT_PARAMS bus %d tgt %d period %x offset %x flags %x",
|
|
bus, tgt, sdp->isp_devparam[tgt].actv_period,
|
|
sdp->isp_devparam[tgt].actv_offset, flags);
|
|
xpt_setup_ccb(&cts.ccb_h, tmppath, 1);
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_async(AC_TRANSFER_NEG, tmppath, &cts);
|
|
xpt_free_path(tmppath);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
break;
|
|
}
|
|
case ISPASYNC_BUS_RESET:
|
|
bus = *((int *)arg);
|
|
isp_prt(isp, ISP_LOGINFO, "SCSI bus reset on bus %d detected",
|
|
bus);
|
|
if (bus > 0 && isp->isp_path2) {
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_async(AC_BUS_RESET, isp->isp_path2, NULL);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
} else if (isp->isp_path) {
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
xpt_async(AC_BUS_RESET, isp->isp_path, NULL);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
}
|
|
break;
|
|
case ISPASYNC_LIP:
|
|
if (isp->isp_path) {
|
|
isp_freeze_loopdown(isp, "ISPASYNC_LIP");
|
|
}
|
|
isp_prt(isp, ISP_LOGINFO, "LIP Received");
|
|
break;
|
|
case ISPASYNC_LOOP_RESET:
|
|
if (isp->isp_path) {
|
|
isp_freeze_loopdown(isp, "ISPASYNC_LOOP_RESET");
|
|
}
|
|
isp_prt(isp, ISP_LOGINFO, "Loop Reset Received");
|
|
break;
|
|
case ISPASYNC_LOOP_DOWN:
|
|
if (isp->isp_path) {
|
|
isp_freeze_loopdown(isp, "ISPASYNC_LOOP_DOWN");
|
|
}
|
|
isp_prt(isp, ISP_LOGINFO, "Loop DOWN");
|
|
break;
|
|
case ISPASYNC_LOOP_UP:
|
|
/*
|
|
* Now we just note that Loop has come up. We don't
|
|
* actually do anything because we're waiting for a
|
|
* Change Notify before activating the FC cleanup
|
|
* thread to look at the state of the loop again.
|
|
*/
|
|
isp_prt(isp, ISP_LOGINFO, "Loop UP");
|
|
break;
|
|
case ISPASYNC_PROMENADE:
|
|
{
|
|
struct cam_path *tmppath;
|
|
const char *fmt = "Target %d (Loop 0x%x) Port ID 0x%x "
|
|
"(role %s) %s\n Port WWN 0x%08x%08x\n Node WWN 0x%08x%08x";
|
|
static const char *roles[4] = {
|
|
"(none)", "Target", "Initiator", "Target/Initiator"
|
|
};
|
|
fcparam *fcp = isp->isp_param;
|
|
int tgt = *((int *) arg);
|
|
int is_tgt_mask = (SVC3_TGT_ROLE >> SVC3_ROLE_SHIFT);
|
|
struct lportdb *lp = &fcp->portdb[tgt];
|
|
|
|
isp_prt(isp, ISP_LOGINFO, fmt, tgt, lp->loopid, lp->portid,
|
|
roles[lp->roles & 0x3],
|
|
(lp->valid)? "Arrived" : "Departed",
|
|
(u_int32_t) (lp->port_wwn >> 32),
|
|
(u_int32_t) (lp->port_wwn & 0xffffffffLL),
|
|
(u_int32_t) (lp->node_wwn >> 32),
|
|
(u_int32_t) (lp->node_wwn & 0xffffffffLL));
|
|
|
|
ISPLOCK_2_CAMLOCK(isp);
|
|
if (xpt_create_path(&tmppath, NULL, cam_sim_path(isp->isp_sim),
|
|
(target_id_t)tgt, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
break;
|
|
}
|
|
/*
|
|
* Policy: only announce targets.
|
|
*/
|
|
if (lp->roles & is_tgt_mask) {
|
|
if (lp->valid) {
|
|
xpt_async(AC_FOUND_DEVICE, tmppath, NULL);
|
|
} else {
|
|
xpt_async(AC_LOST_DEVICE, tmppath, NULL);
|
|
}
|
|
}
|
|
xpt_free_path(tmppath);
|
|
CAMLOCK_2_ISPLOCK(isp);
|
|
break;
|
|
}
|
|
case ISPASYNC_CHANGE_NOTIFY:
|
|
if (arg == ISPASYNC_CHANGE_PDB) {
|
|
isp_prt(isp, ISP_LOGINFO,
|
|
"Port Database Changed");
|
|
} else if (arg == ISPASYNC_CHANGE_SNS) {
|
|
isp_prt(isp, ISP_LOGINFO,
|
|
"Name Server Database Changed");
|
|
}
|
|
cv_signal(&isp->isp_osinfo.kthread_cv);
|
|
break;
|
|
case ISPASYNC_FABRIC_DEV:
|
|
{
|
|
int target, base, lim;
|
|
fcparam *fcp = isp->isp_param;
|
|
struct lportdb *lp = NULL;
|
|
struct lportdb *clp = (struct lportdb *) arg;
|
|
char *pt;
|
|
|
|
switch (clp->port_type) {
|
|
case 1:
|
|
pt = " N_Port";
|
|
break;
|
|
case 2:
|
|
pt = " NL_Port";
|
|
break;
|
|
case 3:
|
|
pt = "F/NL_Port";
|
|
break;
|
|
case 0x7f:
|
|
pt = " Nx_Port";
|
|
break;
|
|
case 0x81:
|
|
pt = " F_port";
|
|
break;
|
|
case 0x82:
|
|
pt = " FL_Port";
|
|
break;
|
|
case 0x84:
|
|
pt = " E_port";
|
|
break;
|
|
default:
|
|
pt = " ";
|
|
break;
|
|
}
|
|
|
|
isp_prt(isp, ISP_LOGINFO,
|
|
"%s Fabric Device @ PortID 0x%x", pt, clp->portid);
|
|
|
|
/*
|
|
* If we don't have an initiator role we bail.
|
|
*
|
|
* We just use ISPASYNC_FABRIC_DEV for announcement purposes.
|
|
*/
|
|
|
|
if ((isp->isp_role & ISP_ROLE_INITIATOR) == 0) {
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* Is this entry for us? If so, we bail.
|
|
*/
|
|
|
|
if (fcp->isp_portid == clp->portid) {
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* Else, the default policy is to find room for it in
|
|
* our local port database. Later, when we execute
|
|
* the call to isp_pdb_sync either this newly arrived
|
|
* or already logged in device will be (re)announced.
|
|
*/
|
|
|
|
if (fcp->isp_topo == TOPO_FL_PORT)
|
|
base = FC_SNS_ID+1;
|
|
else
|
|
base = 0;
|
|
|
|
if (fcp->isp_topo == TOPO_N_PORT)
|
|
lim = 1;
|
|
else
|
|
lim = MAX_FC_TARG;
|
|
|
|
/*
|
|
* Is it already in our list?
|
|
*/
|
|
for (target = base; target < lim; target++) {
|
|
if (target >= FL_PORT_ID && target <= FC_SNS_ID) {
|
|
continue;
|
|
}
|
|
lp = &fcp->portdb[target];
|
|
if (lp->port_wwn == clp->port_wwn &&
|
|
lp->node_wwn == clp->node_wwn) {
|
|
lp->fabric_dev = 1;
|
|
break;
|
|
}
|
|
}
|
|
if (target < lim) {
|
|
break;
|
|
}
|
|
for (target = base; target < lim; target++) {
|
|
if (target >= FL_PORT_ID && target <= FC_SNS_ID) {
|
|
continue;
|
|
}
|
|
lp = &fcp->portdb[target];
|
|
if (lp->port_wwn == 0) {
|
|
break;
|
|
}
|
|
}
|
|
if (target == lim) {
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"out of space for fabric devices");
|
|
break;
|
|
}
|
|
lp->port_type = clp->port_type;
|
|
lp->fc4_type = clp->fc4_type;
|
|
lp->node_wwn = clp->node_wwn;
|
|
lp->port_wwn = clp->port_wwn;
|
|
lp->portid = clp->portid;
|
|
lp->fabric_dev = 1;
|
|
break;
|
|
}
|
|
#ifdef ISP_TARGET_MODE
|
|
case ISPASYNC_TARGET_MESSAGE:
|
|
{
|
|
tmd_msg_t *mp = arg;
|
|
isp_prt(isp, ISP_LOGALL,
|
|
"bus %d iid %d tgt %d lun %d ttype %x tval %x msg[0]=%x",
|
|
mp->nt_bus, (int) mp->nt_iid, (int) mp->nt_tgt,
|
|
(int) mp->nt_lun, mp->nt_tagtype, mp->nt_tagval,
|
|
mp->nt_msg[0]);
|
|
break;
|
|
}
|
|
case ISPASYNC_TARGET_EVENT:
|
|
{
|
|
tmd_event_t *ep = arg;
|
|
isp_prt(isp, ISP_LOGALL,
|
|
"bus %d event code 0x%x", ep->ev_bus, ep->ev_event);
|
|
break;
|
|
}
|
|
case ISPASYNC_TARGET_ACTION:
|
|
switch (((isphdr_t *)arg)->rqs_entry_type) {
|
|
default:
|
|
isp_prt(isp, ISP_LOGWARN,
|
|
"event 0x%x for unhandled target action",
|
|
((isphdr_t *)arg)->rqs_entry_type);
|
|
break;
|
|
case RQSTYPE_NOTIFY:
|
|
if (IS_SCSI(isp)) {
|
|
rv = isp_handle_platform_notify_scsi(isp,
|
|
(in_entry_t *) arg);
|
|
} else {
|
|
rv = isp_handle_platform_notify_fc(isp,
|
|
(in_fcentry_t *) arg);
|
|
}
|
|
break;
|
|
case RQSTYPE_ATIO:
|
|
rv = isp_handle_platform_atio(isp, (at_entry_t *) arg);
|
|
break;
|
|
case RQSTYPE_ATIO2:
|
|
rv = isp_handle_platform_atio2(isp, (at2_entry_t *)arg);
|
|
break;
|
|
case RQSTYPE_CTIO2:
|
|
case RQSTYPE_CTIO:
|
|
rv = isp_handle_platform_ctio(isp, arg);
|
|
break;
|
|
case RQSTYPE_ENABLE_LUN:
|
|
case RQSTYPE_MODIFY_LUN:
|
|
if (IS_DUALBUS(isp)) {
|
|
bus =
|
|
GET_BUS_VAL(((lun_entry_t *)arg)->le_rsvd);
|
|
} else {
|
|
bus = 0;
|
|
}
|
|
isp_cv_signal_rqe(isp, bus,
|
|
((lun_entry_t *)arg)->le_status);
|
|
break;
|
|
}
|
|
break;
|
|
#endif
|
|
case ISPASYNC_FW_CRASH:
|
|
{
|
|
u_int16_t mbox1, mbox6;
|
|
mbox1 = ISP_READ(isp, OUTMAILBOX1);
|
|
if (IS_DUALBUS(isp)) {
|
|
mbox6 = ISP_READ(isp, OUTMAILBOX6);
|
|
} else {
|
|
mbox6 = 0;
|
|
}
|
|
isp_prt(isp, ISP_LOGERR,
|
|
"Internal Firmware Error on bus %d @ RISC Address 0x%x",
|
|
mbox6, mbox1);
|
|
isp_reinit(isp);
|
|
break;
|
|
}
|
|
case ISPASYNC_UNHANDLED_RESPONSE:
|
|
break;
|
|
default:
|
|
isp_prt(isp, ISP_LOGERR, "unknown isp_async event %d", cmd);
|
|
break;
|
|
}
|
|
return (rv);
|
|
}
|
|
|
|
|
|
/*
|
|
* Locks are held before coming here.
|
|
*/
|
|
void
|
|
isp_uninit(struct ispsoftc *isp)
|
|
{
|
|
ISP_WRITE(isp, HCCR, HCCR_CMD_RESET);
|
|
DISABLE_INTS(isp);
|
|
}
|
|
|
|
void
|
|
isp_prt(struct ispsoftc *isp, int level, const char *fmt, ...)
|
|
{
|
|
va_list ap;
|
|
if (level != ISP_LOGALL && (level & isp->isp_dblev) == 0) {
|
|
return;
|
|
}
|
|
printf("%s: ", device_get_nameunit(isp->isp_dev));
|
|
va_start(ap, fmt);
|
|
vprintf(fmt, ap);
|
|
va_end(ap);
|
|
printf("\n");
|
|
}
|