1537078d8f
Mainly focus on files that use BSD 2-Clause license, however the tool I was using misidentified many licenses so this was mostly a manual - error prone - task. The Software Package Data Exchange (SPDX) group provides a specification to make it easier for automated tools to detect and summarize well known opensource licenses. We are gradually adopting the specification, noting that the tags are considered only advisory and do not, in any way, superceed or replace the license texts.
842 lines
20 KiB
C
842 lines
20 KiB
C
/*-
|
|
* SPDX-License-Identifier: BSD-2-Clause-FreeBSD
|
|
*
|
|
* Copyright (c) 1999,2000 Jonathan Lemon
|
|
* All rights reserved.
|
|
*
|
|
# Derived from the original IDA Compaq RAID driver, which is
|
|
* Copyright (c) 1996, 1997, 1998, 1999
|
|
* Mark Dawson and David James. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
/*
|
|
* Generic driver for Compaq SMART RAID adapters.
|
|
*/
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/lock.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/mutex.h>
|
|
#include <sys/stat.h>
|
|
|
|
#include <sys/bio.h>
|
|
#include <sys/bus.h>
|
|
#include <sys/conf.h>
|
|
#include <sys/endian.h>
|
|
|
|
#include <machine/bus.h>
|
|
#include <sys/rman.h>
|
|
|
|
#include <geom/geom_disk.h>
|
|
|
|
#include <dev/ida/idareg.h>
|
|
#include <dev/ida/idavar.h>
|
|
#include <dev/ida/idaio.h>
|
|
|
|
/* prototypes */
|
|
static int ida_alloc_qcbs(struct ida_softc *ida);
|
|
static void ida_done(struct ida_softc *ida, struct ida_qcb *qcb);
|
|
static void ida_start(struct ida_softc *ida);
|
|
static void ida_startio(struct ida_softc *ida);
|
|
static void ida_startup(void *arg);
|
|
static void ida_timeout(void *arg);
|
|
static int ida_wait(struct ida_softc *ida, struct ida_qcb *qcb);
|
|
|
|
static d_ioctl_t ida_ioctl;
|
|
static struct cdevsw ida_cdevsw = {
|
|
.d_version = D_VERSION,
|
|
.d_ioctl = ida_ioctl,
|
|
.d_name = "ida",
|
|
};
|
|
|
|
void
|
|
ida_free(struct ida_softc *ida)
|
|
{
|
|
int i;
|
|
|
|
if (ida->ih != NULL)
|
|
bus_teardown_intr(ida->dev, ida->irq, ida->ih);
|
|
|
|
mtx_lock(&ida->lock);
|
|
callout_stop(&ida->ch);
|
|
mtx_unlock(&ida->lock);
|
|
callout_drain(&ida->ch);
|
|
|
|
if (ida->buffer_dmat) {
|
|
for (i = 0; i < IDA_QCB_MAX; i++)
|
|
bus_dmamap_destroy(ida->buffer_dmat, ida->qcbs[i].dmamap);
|
|
bus_dma_tag_destroy(ida->buffer_dmat);
|
|
}
|
|
|
|
if (ida->hwqcb_dmat) {
|
|
if (ida->hwqcb_busaddr)
|
|
bus_dmamap_unload(ida->hwqcb_dmat, ida->hwqcb_dmamap);
|
|
if (ida->hwqcbs)
|
|
bus_dmamem_free(ida->hwqcb_dmat, ida->hwqcbs,
|
|
ida->hwqcb_dmamap);
|
|
bus_dma_tag_destroy(ida->hwqcb_dmat);
|
|
}
|
|
|
|
if (ida->qcbs != NULL)
|
|
free(ida->qcbs, M_DEVBUF);
|
|
|
|
if (ida->irq != NULL)
|
|
bus_release_resource(ida->dev, ida->irq_res_type,
|
|
0, ida->irq);
|
|
|
|
if (ida->parent_dmat != NULL)
|
|
bus_dma_tag_destroy(ida->parent_dmat);
|
|
|
|
if (ida->regs != NULL)
|
|
bus_release_resource(ida->dev, ida->regs_res_type,
|
|
ida->regs_res_id, ida->regs);
|
|
|
|
mtx_destroy(&ida->lock);
|
|
}
|
|
|
|
/*
|
|
* record bus address from bus_dmamap_load
|
|
*/
|
|
static void
|
|
ida_dma_map_cb(void *arg, bus_dma_segment_t *segs, int nseg, int error)
|
|
{
|
|
bus_addr_t *baddr;
|
|
|
|
baddr = (bus_addr_t *)arg;
|
|
*baddr = segs->ds_addr;
|
|
}
|
|
|
|
static __inline struct ida_qcb *
|
|
ida_get_qcb(struct ida_softc *ida)
|
|
{
|
|
struct ida_qcb *qcb;
|
|
|
|
if ((qcb = SLIST_FIRST(&ida->free_qcbs)) != NULL) {
|
|
SLIST_REMOVE_HEAD(&ida->free_qcbs, link.sle);
|
|
bzero(qcb->hwqcb, sizeof(struct ida_hdr) + sizeof(struct ida_req));
|
|
}
|
|
return (qcb);
|
|
}
|
|
|
|
static __inline void
|
|
ida_free_qcb(struct ida_softc *ida, struct ida_qcb *qcb)
|
|
{
|
|
|
|
qcb->state = QCB_FREE;
|
|
qcb->buf = NULL;
|
|
qcb->error = 0;
|
|
SLIST_INSERT_HEAD(&ida->free_qcbs, qcb, link.sle);
|
|
}
|
|
|
|
static __inline bus_addr_t
|
|
idahwqcbvtop(struct ida_softc *ida, struct ida_hardware_qcb *hwqcb)
|
|
{
|
|
return (ida->hwqcb_busaddr +
|
|
((bus_addr_t)hwqcb - (bus_addr_t)ida->hwqcbs));
|
|
}
|
|
|
|
static __inline struct ida_qcb *
|
|
idahwqcbptov(struct ida_softc *ida, bus_addr_t hwqcb_addr)
|
|
{
|
|
struct ida_hardware_qcb *hwqcb;
|
|
|
|
hwqcb = (struct ida_hardware_qcb *)
|
|
((bus_addr_t)ida->hwqcbs + (hwqcb_addr - ida->hwqcb_busaddr));
|
|
return (hwqcb->qcb);
|
|
}
|
|
|
|
static int
|
|
ida_alloc_qcbs(struct ida_softc *ida)
|
|
{
|
|
struct ida_qcb *qcb;
|
|
int error, i;
|
|
|
|
for (i = 0; i < IDA_QCB_MAX; i++) {
|
|
qcb = &ida->qcbs[i];
|
|
|
|
error = bus_dmamap_create(ida->buffer_dmat, /*flags*/0, &qcb->dmamap);
|
|
if (error != 0)
|
|
return (error);
|
|
|
|
qcb->ida = ida;
|
|
qcb->flags = QCB_FREE;
|
|
qcb->hwqcb = &ida->hwqcbs[i];
|
|
qcb->hwqcb->qcb = qcb;
|
|
qcb->hwqcb_busaddr = idahwqcbvtop(ida, qcb->hwqcb);
|
|
SLIST_INSERT_HEAD(&ida->free_qcbs, qcb, link.sle);
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
ida_setup(struct ida_softc *ida)
|
|
{
|
|
struct ida_controller_info cinfo;
|
|
device_t child;
|
|
int error, i, unit;
|
|
|
|
SLIST_INIT(&ida->free_qcbs);
|
|
STAILQ_INIT(&ida->qcb_queue);
|
|
bioq_init(&ida->bio_queue);
|
|
|
|
ida->qcbs = (struct ida_qcb *)
|
|
malloc(IDA_QCB_MAX * sizeof(struct ida_qcb), M_DEVBUF,
|
|
M_NOWAIT | M_ZERO);
|
|
if (ida->qcbs == NULL)
|
|
return (ENOMEM);
|
|
|
|
/*
|
|
* Create our DMA tags
|
|
*/
|
|
|
|
/* DMA tag for our hardware QCB structures */
|
|
error = bus_dma_tag_create(
|
|
/* parent */ ida->parent_dmat,
|
|
/* alignment */ 1,
|
|
/* boundary */ 0,
|
|
/* lowaddr */ BUS_SPACE_MAXADDR,
|
|
/* highaddr */ BUS_SPACE_MAXADDR,
|
|
/* filter */ NULL,
|
|
/* filterarg */ NULL,
|
|
/* maxsize */ IDA_QCB_MAX * sizeof(struct ida_hardware_qcb),
|
|
/* nsegments */ 1,
|
|
/* maxsegsz */ BUS_SPACE_MAXSIZE_32BIT,
|
|
/* flags */ 0,
|
|
/* lockfunc */ NULL,
|
|
/* lockarg */ NULL,
|
|
&ida->hwqcb_dmat);
|
|
if (error)
|
|
return (ENOMEM);
|
|
|
|
/* DMA tag for mapping buffers into device space */
|
|
error = bus_dma_tag_create(
|
|
/* parent */ ida->parent_dmat,
|
|
/* alignment */ 1,
|
|
/* boundary */ 0,
|
|
/* lowaddr */ BUS_SPACE_MAXADDR,
|
|
/* highaddr */ BUS_SPACE_MAXADDR,
|
|
/* filter */ NULL,
|
|
/* filterarg */ NULL,
|
|
/* maxsize */ DFLTPHYS,
|
|
/* nsegments */ IDA_NSEG,
|
|
/* maxsegsz */ BUS_SPACE_MAXSIZE_32BIT,
|
|
/* flags */ 0,
|
|
/* lockfunc */ busdma_lock_mutex,
|
|
/* lockarg */ &Giant,
|
|
&ida->buffer_dmat);
|
|
if (error)
|
|
return (ENOMEM);
|
|
|
|
/* Allocation of hardware QCBs */
|
|
/* XXX allocation is rounded to hardware page size */
|
|
error = bus_dmamem_alloc(ida->hwqcb_dmat,
|
|
(void **)&ida->hwqcbs, BUS_DMA_NOWAIT, &ida->hwqcb_dmamap);
|
|
if (error)
|
|
return (ENOMEM);
|
|
|
|
/* And permanently map them in */
|
|
bus_dmamap_load(ida->hwqcb_dmat, ida->hwqcb_dmamap,
|
|
ida->hwqcbs, IDA_QCB_MAX * sizeof(struct ida_hardware_qcb),
|
|
ida_dma_map_cb, &ida->hwqcb_busaddr, /*flags*/0);
|
|
|
|
bzero(ida->hwqcbs, IDA_QCB_MAX * sizeof(struct ida_hardware_qcb));
|
|
|
|
error = ida_alloc_qcbs(ida);
|
|
if (error)
|
|
return (error);
|
|
|
|
mtx_lock(&ida->lock);
|
|
ida->cmd.int_enable(ida, 0);
|
|
|
|
error = ida_command(ida, CMD_GET_CTRL_INFO, &cinfo, sizeof(cinfo),
|
|
IDA_CONTROLLER, 0, DMA_DATA_IN);
|
|
if (error) {
|
|
mtx_unlock(&ida->lock);
|
|
device_printf(ida->dev, "CMD_GET_CTRL_INFO failed.\n");
|
|
return (error);
|
|
}
|
|
|
|
device_printf(ida->dev, "drives=%d firm_rev=%c%c%c%c\n",
|
|
cinfo.num_drvs, cinfo.firm_rev[0], cinfo.firm_rev[1],
|
|
cinfo.firm_rev[2], cinfo.firm_rev[3]);
|
|
|
|
if (ida->flags & IDA_FIRMWARE) {
|
|
int data;
|
|
|
|
error = ida_command(ida, CMD_START_FIRMWARE,
|
|
&data, sizeof(data), IDA_CONTROLLER, 0, DMA_DATA_IN);
|
|
if (error) {
|
|
mtx_unlock(&ida->lock);
|
|
device_printf(ida->dev, "CMD_START_FIRMWARE failed.\n");
|
|
return (error);
|
|
}
|
|
}
|
|
|
|
ida->cmd.int_enable(ida, 1);
|
|
ida->flags |= IDA_ATTACHED;
|
|
mtx_unlock(&ida->lock);
|
|
|
|
for (i = 0; i < cinfo.num_drvs; i++) {
|
|
child = device_add_child(ida->dev, /*"idad"*/NULL, -1);
|
|
if (child != NULL)
|
|
device_set_ivars(child, (void *)(intptr_t)i);
|
|
}
|
|
|
|
ida->ich.ich_func = ida_startup;
|
|
ida->ich.ich_arg = ida;
|
|
if (config_intrhook_establish(&ida->ich) != 0) {
|
|
device_delete_children(ida->dev);
|
|
device_printf(ida->dev, "Cannot establish configuration hook\n");
|
|
return (error);
|
|
}
|
|
|
|
unit = device_get_unit(ida->dev);
|
|
ida->ida_dev_t = make_dev(&ida_cdevsw, unit,
|
|
UID_ROOT, GID_OPERATOR, S_IRUSR | S_IWUSR,
|
|
"ida%d", unit);
|
|
ida->ida_dev_t->si_drv1 = ida;
|
|
|
|
return (0);
|
|
}
|
|
|
|
static void
|
|
ida_startup(void *arg)
|
|
{
|
|
struct ida_softc *ida;
|
|
|
|
ida = arg;
|
|
|
|
config_intrhook_disestablish(&ida->ich);
|
|
|
|
mtx_lock(&Giant);
|
|
bus_generic_attach(ida->dev);
|
|
mtx_unlock(&Giant);
|
|
}
|
|
|
|
int
|
|
ida_detach(device_t dev)
|
|
{
|
|
struct ida_softc *ida;
|
|
int error;
|
|
|
|
ida = (struct ida_softc *)device_get_softc(dev);
|
|
|
|
error = bus_generic_detach(dev);
|
|
if (error)
|
|
return (error);
|
|
error = device_delete_children(dev);
|
|
if (error)
|
|
return (error);
|
|
|
|
/*
|
|
* XXX
|
|
* before detaching, we must make sure that the system is
|
|
* quiescent; nothing mounted, no pending activity.
|
|
*/
|
|
|
|
/*
|
|
* XXX
|
|
* now, how are we supposed to maintain a list of our drives?
|
|
* iterate over our "child devices"?
|
|
*/
|
|
|
|
destroy_dev(ida->ida_dev_t);
|
|
ida_free(ida);
|
|
return (error);
|
|
}
|
|
|
|
static void
|
|
ida_data_cb(void *arg, bus_dma_segment_t *segs, int nsegments, int error)
|
|
{
|
|
struct ida_hardware_qcb *hwqcb;
|
|
struct ida_softc *ida;
|
|
struct ida_qcb *qcb;
|
|
bus_dmasync_op_t op;
|
|
int i;
|
|
|
|
qcb = arg;
|
|
ida = qcb->ida;
|
|
if (!dumping)
|
|
mtx_assert(&ida->lock, MA_OWNED);
|
|
if (error) {
|
|
qcb->error = error;
|
|
ida_done(ida, qcb);
|
|
return;
|
|
}
|
|
|
|
hwqcb = qcb->hwqcb;
|
|
hwqcb->hdr.size = htole16((sizeof(struct ida_req) +
|
|
sizeof(struct ida_sgb) * IDA_NSEG) >> 2);
|
|
|
|
for (i = 0; i < nsegments; i++) {
|
|
hwqcb->seg[i].addr = htole32(segs[i].ds_addr);
|
|
hwqcb->seg[i].length = htole32(segs[i].ds_len);
|
|
}
|
|
hwqcb->req.sgcount = nsegments;
|
|
if (qcb->flags & DMA_DATA_TRANSFER) {
|
|
switch (qcb->flags & DMA_DATA_TRANSFER) {
|
|
case DMA_DATA_TRANSFER:
|
|
op = BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE;
|
|
break;
|
|
case DMA_DATA_IN:
|
|
op = BUS_DMASYNC_PREREAD;
|
|
break;
|
|
default:
|
|
KASSERT((qcb->flags & DMA_DATA_TRANSFER) ==
|
|
DMA_DATA_OUT, ("bad DMA data flags"));
|
|
op = BUS_DMASYNC_PREWRITE;
|
|
break;
|
|
}
|
|
bus_dmamap_sync(ida->buffer_dmat, qcb->dmamap, op);
|
|
}
|
|
bus_dmamap_sync(ida->hwqcb_dmat, ida->hwqcb_dmamap,
|
|
BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD);
|
|
|
|
STAILQ_INSERT_TAIL(&ida->qcb_queue, qcb, link.stqe);
|
|
ida_start(ida);
|
|
ida->flags &= ~IDA_QFROZEN;
|
|
}
|
|
|
|
static int
|
|
ida_map_qcb(struct ida_softc *ida, struct ida_qcb *qcb, void *data,
|
|
bus_size_t datasize)
|
|
{
|
|
int error, flags;
|
|
|
|
if (ida->flags & IDA_INTERRUPTS)
|
|
flags = BUS_DMA_WAITOK;
|
|
else
|
|
flags = BUS_DMA_NOWAIT;
|
|
error = bus_dmamap_load(ida->buffer_dmat, qcb->dmamap, data, datasize,
|
|
ida_data_cb, qcb, flags);
|
|
if (error == EINPROGRESS) {
|
|
ida->flags |= IDA_QFROZEN;
|
|
error = 0;
|
|
}
|
|
return (error);
|
|
}
|
|
|
|
int
|
|
ida_command(struct ida_softc *ida, int command, void *data, int datasize,
|
|
int drive, u_int32_t pblkno, int flags)
|
|
{
|
|
struct ida_hardware_qcb *hwqcb;
|
|
struct ida_qcb *qcb;
|
|
int error;
|
|
|
|
if (!dumping)
|
|
mtx_assert(&ida->lock, MA_OWNED);
|
|
qcb = ida_get_qcb(ida);
|
|
|
|
if (qcb == NULL) {
|
|
device_printf(ida->dev, "out of QCBs\n");
|
|
return (EAGAIN);
|
|
}
|
|
|
|
qcb->flags = flags | IDA_COMMAND;
|
|
hwqcb = qcb->hwqcb;
|
|
hwqcb->hdr.drive = drive;
|
|
hwqcb->req.blkno = htole32(pblkno);
|
|
hwqcb->req.bcount = htole16(howmany(datasize, DEV_BSIZE));
|
|
hwqcb->req.command = command;
|
|
|
|
error = ida_map_qcb(ida, qcb, data, datasize);
|
|
if (error == 0) {
|
|
error = ida_wait(ida, qcb);
|
|
/* Don't free QCB on a timeout in case it later completes. */
|
|
if (error)
|
|
return (error);
|
|
error = qcb->error;
|
|
}
|
|
|
|
/* XXX should have status returned here? */
|
|
/* XXX have "status pointer" area in QCB? */
|
|
|
|
ida_free_qcb(ida, qcb);
|
|
return (error);
|
|
}
|
|
|
|
void
|
|
ida_submit_buf(struct ida_softc *ida, struct bio *bp)
|
|
{
|
|
mtx_lock(&ida->lock);
|
|
bioq_insert_tail(&ida->bio_queue, bp);
|
|
ida_startio(ida);
|
|
mtx_unlock(&ida->lock);
|
|
}
|
|
|
|
static void
|
|
ida_startio(struct ida_softc *ida)
|
|
{
|
|
struct ida_hardware_qcb *hwqcb;
|
|
struct ida_qcb *qcb;
|
|
struct idad_softc *drv;
|
|
struct bio *bp;
|
|
int error;
|
|
|
|
mtx_assert(&ida->lock, MA_OWNED);
|
|
for (;;) {
|
|
if (ida->flags & IDA_QFROZEN)
|
|
return;
|
|
bp = bioq_first(&ida->bio_queue);
|
|
if (bp == NULL)
|
|
return; /* no more buffers */
|
|
|
|
qcb = ida_get_qcb(ida);
|
|
if (qcb == NULL)
|
|
return; /* out of resources */
|
|
|
|
bioq_remove(&ida->bio_queue, bp);
|
|
qcb->buf = bp;
|
|
qcb->flags = bp->bio_cmd == BIO_READ ? DMA_DATA_IN : DMA_DATA_OUT;
|
|
|
|
hwqcb = qcb->hwqcb;
|
|
drv = bp->bio_driver1;
|
|
hwqcb->hdr.drive = drv->drive;
|
|
hwqcb->req.blkno = bp->bio_pblkno;
|
|
hwqcb->req.bcount = howmany(bp->bio_bcount, DEV_BSIZE);
|
|
hwqcb->req.command = bp->bio_cmd == BIO_READ ? CMD_READ : CMD_WRITE;
|
|
|
|
error = ida_map_qcb(ida, qcb, bp->bio_data, bp->bio_bcount);
|
|
if (error) {
|
|
qcb->error = error;
|
|
ida_done(ida, qcb);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
ida_start(struct ida_softc *ida)
|
|
{
|
|
struct ida_qcb *qcb;
|
|
|
|
if (!dumping)
|
|
mtx_assert(&ida->lock, MA_OWNED);
|
|
while ((qcb = STAILQ_FIRST(&ida->qcb_queue)) != NULL) {
|
|
if (ida->cmd.fifo_full(ida))
|
|
break;
|
|
STAILQ_REMOVE_HEAD(&ida->qcb_queue, link.stqe);
|
|
/*
|
|
* XXX
|
|
* place the qcb on an active list?
|
|
*/
|
|
|
|
/* Set a timeout. */
|
|
if (!ida->qactive && !dumping)
|
|
callout_reset(&ida->ch, hz * 5, ida_timeout, ida);
|
|
ida->qactive++;
|
|
|
|
qcb->state = QCB_ACTIVE;
|
|
ida->cmd.submit(ida, qcb);
|
|
}
|
|
}
|
|
|
|
static int
|
|
ida_wait(struct ida_softc *ida, struct ida_qcb *qcb)
|
|
{
|
|
struct ida_qcb *qcb_done = NULL;
|
|
bus_addr_t completed;
|
|
int delay;
|
|
|
|
if (!dumping)
|
|
mtx_assert(&ida->lock, MA_OWNED);
|
|
if (ida->flags & IDA_INTERRUPTS) {
|
|
if (mtx_sleep(qcb, &ida->lock, PRIBIO, "idacmd", 5 * hz)) {
|
|
qcb->state = QCB_TIMEDOUT;
|
|
return (ETIMEDOUT);
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
again:
|
|
delay = 5 * 1000 * 100; /* 5 sec delay */
|
|
while ((completed = ida->cmd.done(ida)) == 0) {
|
|
if (delay-- == 0) {
|
|
qcb->state = QCB_TIMEDOUT;
|
|
return (ETIMEDOUT);
|
|
}
|
|
DELAY(10);
|
|
}
|
|
|
|
qcb_done = idahwqcbptov(ida, completed & ~3);
|
|
if (qcb_done != qcb)
|
|
goto again;
|
|
ida_done(ida, qcb);
|
|
return (0);
|
|
}
|
|
|
|
void
|
|
ida_intr(void *data)
|
|
{
|
|
struct ida_softc *ida;
|
|
struct ida_qcb *qcb;
|
|
bus_addr_t completed;
|
|
|
|
ida = (struct ida_softc *)data;
|
|
|
|
mtx_lock(&ida->lock);
|
|
if (ida->cmd.int_pending(ida) == 0) {
|
|
mtx_unlock(&ida->lock);
|
|
return; /* not our interrupt */
|
|
}
|
|
|
|
while ((completed = ida->cmd.done(ida)) != 0) {
|
|
qcb = idahwqcbptov(ida, completed & ~3);
|
|
|
|
if (qcb == NULL || qcb->state != QCB_ACTIVE) {
|
|
device_printf(ida->dev,
|
|
"ignoring completion %jx\n", (intmax_t)completed);
|
|
continue;
|
|
}
|
|
/* Handle "Bad Command List" errors. */
|
|
if ((completed & 3) && (qcb->hwqcb->req.error == 0))
|
|
qcb->hwqcb->req.error = CMD_REJECTED;
|
|
ida_done(ida, qcb);
|
|
}
|
|
ida_startio(ida);
|
|
mtx_unlock(&ida->lock);
|
|
}
|
|
|
|
/*
|
|
* should switch out command type; may be status, not just I/O.
|
|
*/
|
|
static void
|
|
ida_done(struct ida_softc *ida, struct ida_qcb *qcb)
|
|
{
|
|
bus_dmasync_op_t op;
|
|
int active, error = 0;
|
|
|
|
/*
|
|
* finish up command
|
|
*/
|
|
if (!dumping)
|
|
mtx_assert(&ida->lock, MA_OWNED);
|
|
active = (qcb->state != QCB_FREE);
|
|
if (qcb->flags & DMA_DATA_TRANSFER && active) {
|
|
switch (qcb->flags & DMA_DATA_TRANSFER) {
|
|
case DMA_DATA_TRANSFER:
|
|
op = BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE;
|
|
break;
|
|
case DMA_DATA_IN:
|
|
op = BUS_DMASYNC_POSTREAD;
|
|
break;
|
|
default:
|
|
KASSERT((qcb->flags & DMA_DATA_TRANSFER) ==
|
|
DMA_DATA_OUT, ("bad DMA data flags"));
|
|
op = BUS_DMASYNC_POSTWRITE;
|
|
break;
|
|
}
|
|
bus_dmamap_sync(ida->buffer_dmat, qcb->dmamap, op);
|
|
bus_dmamap_unload(ida->buffer_dmat, qcb->dmamap);
|
|
}
|
|
if (active)
|
|
bus_dmamap_sync(ida->hwqcb_dmat, ida->hwqcb_dmamap,
|
|
BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
|
|
|
|
if (qcb->hwqcb->req.error & SOFT_ERROR) {
|
|
if (qcb->buf)
|
|
device_printf(ida->dev, "soft %s error\n",
|
|
qcb->buf->bio_cmd == BIO_READ ?
|
|
"read" : "write");
|
|
else
|
|
device_printf(ida->dev, "soft error\n");
|
|
}
|
|
if (qcb->hwqcb->req.error & HARD_ERROR) {
|
|
error = 1;
|
|
if (qcb->buf)
|
|
device_printf(ida->dev, "hard %s error\n",
|
|
qcb->buf->bio_cmd == BIO_READ ?
|
|
"read" : "write");
|
|
else
|
|
device_printf(ida->dev, "hard error\n");
|
|
}
|
|
if (qcb->hwqcb->req.error & CMD_REJECTED) {
|
|
error = 1;
|
|
device_printf(ida->dev, "invalid request\n");
|
|
}
|
|
if (qcb->error) {
|
|
error = 1;
|
|
device_printf(ida->dev, "request failed to map: %d\n", qcb->error);
|
|
}
|
|
|
|
if (qcb->flags & IDA_COMMAND) {
|
|
if (ida->flags & IDA_INTERRUPTS)
|
|
wakeup(qcb);
|
|
if (qcb->state == QCB_TIMEDOUT)
|
|
ida_free_qcb(ida, qcb);
|
|
} else {
|
|
KASSERT(qcb->buf != NULL, ("ida_done(): qcb->buf is NULL!"));
|
|
if (error)
|
|
qcb->buf->bio_flags |= BIO_ERROR;
|
|
idad_intr(qcb->buf);
|
|
ida_free_qcb(ida, qcb);
|
|
}
|
|
|
|
if (!active)
|
|
return;
|
|
|
|
ida->qactive--;
|
|
/* Reschedule or cancel timeout */
|
|
if (ida->qactive)
|
|
callout_reset(&ida->ch, hz * 5, ida_timeout, ida);
|
|
else
|
|
callout_stop(&ida->ch);
|
|
}
|
|
|
|
static void
|
|
ida_timeout(void *arg)
|
|
{
|
|
struct ida_softc *ida;
|
|
|
|
ida = (struct ida_softc *)arg;
|
|
device_printf(ida->dev, "%s() qactive %d\n", __func__, ida->qactive);
|
|
|
|
if (ida->flags & IDA_INTERRUPTS)
|
|
device_printf(ida->dev, "IDA_INTERRUPTS\n");
|
|
|
|
device_printf(ida->dev, "\t R_CMD_FIFO: %08x\n"
|
|
"\t R_DONE_FIFO: %08x\n"
|
|
"\t R_INT_MASK: %08x\n"
|
|
"\t R_STATUS: %08x\n"
|
|
"\tR_INT_PENDING: %08x\n",
|
|
ida_inl(ida, R_CMD_FIFO),
|
|
ida_inl(ida, R_DONE_FIFO),
|
|
ida_inl(ida, R_INT_MASK),
|
|
ida_inl(ida, R_STATUS),
|
|
ida_inl(ida, R_INT_PENDING));
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* IOCTL stuff follows.
|
|
*/
|
|
struct cmd_info {
|
|
int cmd;
|
|
int len;
|
|
int flags;
|
|
};
|
|
static struct cmd_info *ida_cmd_lookup(int);
|
|
|
|
static int
|
|
ida_ioctl (struct cdev *dev, u_long cmd, caddr_t addr, int32_t flag, struct thread *td)
|
|
{
|
|
struct ida_softc *sc;
|
|
struct ida_user_command *uc;
|
|
struct cmd_info *ci;
|
|
int len;
|
|
int flags;
|
|
int error;
|
|
int data;
|
|
void *daddr;
|
|
|
|
sc = (struct ida_softc *)dev->si_drv1;
|
|
uc = (struct ida_user_command *)addr;
|
|
error = 0;
|
|
|
|
switch (cmd) {
|
|
case IDAIO_COMMAND:
|
|
ci = ida_cmd_lookup(uc->command);
|
|
if (ci == NULL) {
|
|
error = EINVAL;
|
|
break;
|
|
}
|
|
len = ci->len;
|
|
flags = ci->flags;
|
|
if (len)
|
|
daddr = &uc->d.buf;
|
|
else {
|
|
daddr = &data;
|
|
len = sizeof(data);
|
|
}
|
|
mtx_lock(&sc->lock);
|
|
error = ida_command(sc, uc->command, daddr, len,
|
|
uc->drive, uc->blkno, flags);
|
|
mtx_unlock(&sc->lock);
|
|
break;
|
|
default:
|
|
error = ENOIOCTL;
|
|
break;
|
|
}
|
|
return (error);
|
|
}
|
|
|
|
static struct cmd_info ci_list[] = {
|
|
{ CMD_GET_LOG_DRV_INFO,
|
|
sizeof(struct ida_drive_info), DMA_DATA_IN },
|
|
{ CMD_GET_CTRL_INFO,
|
|
sizeof(struct ida_controller_info), DMA_DATA_IN },
|
|
{ CMD_SENSE_DRV_STATUS,
|
|
sizeof(struct ida_drive_status), DMA_DATA_IN },
|
|
{ CMD_START_RECOVERY, 0, 0 },
|
|
{ CMD_GET_PHYS_DRV_INFO,
|
|
sizeof(struct ida_phys_drv_info), DMA_DATA_TRANSFER },
|
|
{ CMD_BLINK_DRV_LEDS,
|
|
sizeof(struct ida_blink_drv_leds), DMA_DATA_OUT },
|
|
{ CMD_SENSE_DRV_LEDS,
|
|
sizeof(struct ida_blink_drv_leds), DMA_DATA_IN },
|
|
{ CMD_GET_LOG_DRV_EXT,
|
|
sizeof(struct ida_drive_info_ext), DMA_DATA_IN },
|
|
{ CMD_RESET_CTRL, 0, 0 },
|
|
{ CMD_GET_CONFIG, 0, 0 },
|
|
{ CMD_SET_CONFIG, 0, 0 },
|
|
{ CMD_LABEL_LOG_DRV,
|
|
sizeof(struct ida_label_logical), DMA_DATA_OUT },
|
|
{ CMD_SET_SURFACE_DELAY, 0, 0 },
|
|
{ CMD_SENSE_BUS_PARAMS, 0, 0 },
|
|
{ CMD_SENSE_SUBSYS_INFO, 0, 0 },
|
|
{ CMD_SENSE_SURFACE_ATS, 0, 0 },
|
|
{ CMD_PASSTHROUGH, 0, 0 },
|
|
{ CMD_RESET_SCSI_DEV, 0, 0 },
|
|
{ CMD_PAUSE_BG_ACT, 0, 0 },
|
|
{ CMD_RESUME_BG_ACT, 0, 0 },
|
|
{ CMD_START_FIRMWARE, 0, 0 },
|
|
{ CMD_SENSE_DRV_ERR_LOG, 0, 0 },
|
|
{ CMD_START_CPM, 0, 0 },
|
|
{ CMD_SENSE_CP, 0, 0 },
|
|
{ CMD_STOP_CPM, 0, 0 },
|
|
{ CMD_FLUSH_CACHE, 0, 0 },
|
|
{ CMD_ACCEPT_MEDIA_EXCH, 0, 0 },
|
|
{ 0, 0, 0 }
|
|
};
|
|
|
|
static struct cmd_info *
|
|
ida_cmd_lookup (int command)
|
|
{
|
|
struct cmd_info *ci;
|
|
|
|
ci = ci_list;
|
|
while (ci->cmd) {
|
|
if (ci->cmd == command)
|
|
return (ci);
|
|
ci++;
|
|
}
|
|
return (NULL);
|
|
}
|