2019-08-21 22:17:55 +00:00
|
|
|
/*-
|
|
|
|
* Copyright (C) 2012-2016 Intel Corporation
|
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
|
|
* SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <sys/cdefs.h>
|
|
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
|
|
|
|
#include <sys/param.h>
|
|
|
|
#include <sys/systm.h>
|
|
|
|
#include <sys/buf.h>
|
|
|
|
#include <sys/bus.h>
|
|
|
|
#include <sys/conf.h>
|
|
|
|
#include <sys/proc.h>
|
|
|
|
#include <sys/smp.h>
|
2019-09-23 17:53:47 +00:00
|
|
|
#include <vm/vm.h>
|
2019-08-21 22:17:55 +00:00
|
|
|
|
|
|
|
#include <dev/pci/pcireg.h>
|
|
|
|
#include <dev/pci/pcivar.h>
|
|
|
|
|
|
|
|
#include "nvme_private.h"
|
|
|
|
|
|
|
|
static int nvme_pci_probe(device_t);
|
|
|
|
static int nvme_pci_attach(device_t);
|
|
|
|
static int nvme_pci_detach(device_t);
|
Implement nvme suspend / resume for pci attachment
When we suspend, we need to properly shutdown the NVME controller. The
controller may go into D3 state (or may have the power removed), and
to properly flush the metadata to non-volatile RAM, we must complete a
normal shutdown. This consists of deleting the I/O queues and setting
the shutodown bit. We have to do some extra stuff to make sure we
reset the software state of the queues as well.
On resume, we have to reset the card twice, for reasons described in
the attach funcion. Once we've done that, we can restart the card. If
any of this fails, we'll fail the NVMe card, just like we do when a
reset fails.
Set is_resetting for the duration of the suspend / resume. This keeps
the reset taskqueue from running a concurrent reset, and also is
needed to prevent any hw completions from queueing more I/O to the
card. Pass resetting flag to nvme_ctrlr_start. It doesn't need to get
that from the global state of the ctrlr. Wait for any pending reset to
finish. All queued I/O will get sent to the hardware as part of
nvme_ctrlr_start(), though the upper layers shouldn't send any
down. Disabling the qpairs is the other failsafe to ensure all I/O is
queued.
Rename nvme_ctrlr_destory_qpairs to nvme_ctrlr_delete_qpairs to avoid
confusion with all the other destroy functions. It just removes the
queues in hardware, while the other _destroy_ functions tear down
driver data structures.
Split parts of the hardware reset function up so that I can
do part of the reset in suspsend. Split out the software disabling
of the qpairs into nvme_ctrlr_disable_qpairs.
Finally, fix a couple of spelling errors in comments related to
this.
Relnotes: Yes
MFC After: 1 week
Reviewed by: scottl@ (prior version)
Differential Revision: https://reviews.freebsd.org/D21493
2019-09-03 15:26:11 +00:00
|
|
|
static int nvme_pci_suspend(device_t);
|
|
|
|
static int nvme_pci_resume(device_t);
|
2019-08-21 22:17:55 +00:00
|
|
|
|
|
|
|
static void nvme_ctrlr_setup_interrupts(struct nvme_controller *ctrlr);
|
|
|
|
|
|
|
|
static device_method_t nvme_pci_methods[] = {
|
|
|
|
/* Device interface */
|
|
|
|
DEVMETHOD(device_probe, nvme_pci_probe),
|
|
|
|
DEVMETHOD(device_attach, nvme_pci_attach),
|
|
|
|
DEVMETHOD(device_detach, nvme_pci_detach),
|
Implement nvme suspend / resume for pci attachment
When we suspend, we need to properly shutdown the NVME controller. The
controller may go into D3 state (or may have the power removed), and
to properly flush the metadata to non-volatile RAM, we must complete a
normal shutdown. This consists of deleting the I/O queues and setting
the shutodown bit. We have to do some extra stuff to make sure we
reset the software state of the queues as well.
On resume, we have to reset the card twice, for reasons described in
the attach funcion. Once we've done that, we can restart the card. If
any of this fails, we'll fail the NVMe card, just like we do when a
reset fails.
Set is_resetting for the duration of the suspend / resume. This keeps
the reset taskqueue from running a concurrent reset, and also is
needed to prevent any hw completions from queueing more I/O to the
card. Pass resetting flag to nvme_ctrlr_start. It doesn't need to get
that from the global state of the ctrlr. Wait for any pending reset to
finish. All queued I/O will get sent to the hardware as part of
nvme_ctrlr_start(), though the upper layers shouldn't send any
down. Disabling the qpairs is the other failsafe to ensure all I/O is
queued.
Rename nvme_ctrlr_destory_qpairs to nvme_ctrlr_delete_qpairs to avoid
confusion with all the other destroy functions. It just removes the
queues in hardware, while the other _destroy_ functions tear down
driver data structures.
Split parts of the hardware reset function up so that I can
do part of the reset in suspsend. Split out the software disabling
of the qpairs into nvme_ctrlr_disable_qpairs.
Finally, fix a couple of spelling errors in comments related to
this.
Relnotes: Yes
MFC After: 1 week
Reviewed by: scottl@ (prior version)
Differential Revision: https://reviews.freebsd.org/D21493
2019-09-03 15:26:11 +00:00
|
|
|
DEVMETHOD(device_suspend, nvme_pci_suspend),
|
|
|
|
DEVMETHOD(device_resume, nvme_pci_resume),
|
2019-08-21 22:17:55 +00:00
|
|
|
DEVMETHOD(device_shutdown, nvme_shutdown),
|
|
|
|
{ 0, 0 }
|
|
|
|
};
|
|
|
|
|
|
|
|
static driver_t nvme_pci_driver = {
|
|
|
|
"nvme",
|
|
|
|
nvme_pci_methods,
|
|
|
|
sizeof(struct nvme_controller),
|
|
|
|
};
|
|
|
|
|
|
|
|
DRIVER_MODULE(nvme, pci, nvme_pci_driver, nvme_devclass, NULL, 0);
|
|
|
|
|
|
|
|
static struct _pcsid
|
|
|
|
{
|
|
|
|
uint32_t devid;
|
|
|
|
int match_subdevice;
|
|
|
|
uint16_t subdevice;
|
|
|
|
const char *desc;
|
|
|
|
uint32_t quirks;
|
|
|
|
} pci_ids[] = {
|
|
|
|
{ 0x01118086, 0, 0, "NVMe Controller" },
|
|
|
|
{ IDT32_PCI_ID, 0, 0, "IDT NVMe Controller (32 channel)" },
|
|
|
|
{ IDT8_PCI_ID, 0, 0, "IDT NVMe Controller (8 channel)" },
|
|
|
|
{ 0x09538086, 1, 0x3702, "DC P3700 SSD" },
|
|
|
|
{ 0x09538086, 1, 0x3703, "DC P3700 SSD [2.5\" SFF]" },
|
|
|
|
{ 0x09538086, 1, 0x3704, "DC P3500 SSD [Add-in Card]" },
|
|
|
|
{ 0x09538086, 1, 0x3705, "DC P3500 SSD [2.5\" SFF]" },
|
|
|
|
{ 0x09538086, 1, 0x3709, "DC P3600 SSD [Add-in Card]" },
|
|
|
|
{ 0x09538086, 1, 0x370a, "DC P3600 SSD [2.5\" SFF]" },
|
|
|
|
{ 0x00031c58, 0, 0, "HGST SN100", QUIRK_DELAY_B4_CHK_RDY },
|
|
|
|
{ 0x00231c58, 0, 0, "WDC SN200", QUIRK_DELAY_B4_CHK_RDY },
|
|
|
|
{ 0x05401c5f, 0, 0, "Memblaze Pblaze4", QUIRK_DELAY_B4_CHK_RDY },
|
|
|
|
{ 0xa821144d, 0, 0, "Samsung PM1725", QUIRK_DELAY_B4_CHK_RDY },
|
|
|
|
{ 0xa822144d, 0, 0, "Samsung PM1725a", QUIRK_DELAY_B4_CHK_RDY },
|
|
|
|
{ 0x00000000, 0, 0, NULL }
|
|
|
|
};
|
|
|
|
|
|
|
|
static int
|
|
|
|
nvme_match(uint32_t devid, uint16_t subdevice, struct _pcsid *ep)
|
|
|
|
{
|
|
|
|
if (devid != ep->devid)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (!ep->match_subdevice)
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
if (subdevice == ep->subdevice)
|
|
|
|
return 1;
|
|
|
|
else
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
nvme_pci_probe (device_t device)
|
|
|
|
{
|
|
|
|
struct nvme_controller *ctrlr = DEVICE2SOFTC(device);
|
|
|
|
struct _pcsid *ep;
|
|
|
|
uint32_t devid;
|
|
|
|
uint16_t subdevice;
|
|
|
|
|
|
|
|
devid = pci_get_devid(device);
|
|
|
|
subdevice = pci_get_subdevice(device);
|
|
|
|
ep = pci_ids;
|
|
|
|
|
|
|
|
while (ep->devid) {
|
|
|
|
if (nvme_match(devid, subdevice, ep))
|
|
|
|
break;
|
|
|
|
++ep;
|
|
|
|
}
|
|
|
|
if (ep->devid)
|
|
|
|
ctrlr->quirks = ep->quirks;
|
|
|
|
|
|
|
|
if (ep->desc) {
|
|
|
|
device_set_desc(device, ep->desc);
|
|
|
|
return (BUS_PROBE_DEFAULT);
|
|
|
|
}
|
|
|
|
|
|
|
|
#if defined(PCIS_STORAGE_NVM)
|
|
|
|
if (pci_get_class(device) == PCIC_STORAGE &&
|
|
|
|
pci_get_subclass(device) == PCIS_STORAGE_NVM &&
|
|
|
|
pci_get_progif(device) == PCIP_STORAGE_NVM_ENTERPRISE_NVMHCI_1_0) {
|
|
|
|
device_set_desc(device, "Generic NVMe Device");
|
|
|
|
return (BUS_PROBE_GENERIC);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
return (ENXIO);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
nvme_ctrlr_allocate_bar(struct nvme_controller *ctrlr)
|
|
|
|
{
|
|
|
|
|
|
|
|
ctrlr->resource_id = PCIR_BAR(0);
|
|
|
|
|
|
|
|
ctrlr->resource = bus_alloc_resource_any(ctrlr->dev, SYS_RES_MEMORY,
|
|
|
|
&ctrlr->resource_id, RF_ACTIVE);
|
|
|
|
|
|
|
|
if(ctrlr->resource == NULL) {
|
|
|
|
nvme_printf(ctrlr, "unable to allocate pci resource\n");
|
|
|
|
return (ENOMEM);
|
|
|
|
}
|
|
|
|
|
|
|
|
ctrlr->bus_tag = rman_get_bustag(ctrlr->resource);
|
|
|
|
ctrlr->bus_handle = rman_get_bushandle(ctrlr->resource);
|
|
|
|
ctrlr->regs = (struct nvme_registers *)ctrlr->bus_handle;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The NVMe spec allows for the MSI-X table to be placed behind
|
|
|
|
* BAR 4/5, separate from the control/doorbell registers. Always
|
|
|
|
* try to map this bar, because it must be mapped prior to calling
|
|
|
|
* pci_alloc_msix(). If the table isn't behind BAR 4/5,
|
|
|
|
* bus_alloc_resource() will just return NULL which is OK.
|
|
|
|
*/
|
|
|
|
ctrlr->bar4_resource_id = PCIR_BAR(4);
|
|
|
|
ctrlr->bar4_resource = bus_alloc_resource_any(ctrlr->dev, SYS_RES_MEMORY,
|
|
|
|
&ctrlr->bar4_resource_id, RF_ACTIVE);
|
|
|
|
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
nvme_pci_attach(device_t dev)
|
|
|
|
{
|
|
|
|
struct nvme_controller*ctrlr = DEVICE2SOFTC(dev);
|
|
|
|
int status;
|
|
|
|
|
|
|
|
ctrlr->dev = dev;
|
|
|
|
status = nvme_ctrlr_allocate_bar(ctrlr);
|
|
|
|
if (status != 0)
|
|
|
|
goto bad;
|
|
|
|
pci_enable_busmaster(dev);
|
|
|
|
nvme_ctrlr_setup_interrupts(ctrlr);
|
|
|
|
return nvme_attach(dev);
|
|
|
|
bad:
|
|
|
|
if (ctrlr->resource != NULL) {
|
|
|
|
bus_release_resource(dev, SYS_RES_MEMORY,
|
|
|
|
ctrlr->resource_id, ctrlr->resource);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ctrlr->bar4_resource != NULL) {
|
|
|
|
bus_release_resource(dev, SYS_RES_MEMORY,
|
|
|
|
ctrlr->bar4_resource_id, ctrlr->bar4_resource);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ctrlr->tag)
|
|
|
|
bus_teardown_intr(dev, ctrlr->res, ctrlr->tag);
|
|
|
|
|
|
|
|
if (ctrlr->res)
|
|
|
|
bus_release_resource(dev, SYS_RES_IRQ,
|
|
|
|
rman_get_rid(ctrlr->res), ctrlr->res);
|
|
|
|
|
|
|
|
if (ctrlr->msix_enabled)
|
|
|
|
pci_release_msi(dev);
|
|
|
|
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
nvme_pci_detach(device_t dev)
|
|
|
|
{
|
|
|
|
struct nvme_controller*ctrlr = DEVICE2SOFTC(dev);
|
2019-08-22 20:09:32 +00:00
|
|
|
int rv;
|
2019-08-21 22:17:55 +00:00
|
|
|
|
2019-08-22 20:09:32 +00:00
|
|
|
rv = nvme_detach(dev);
|
2019-08-21 22:17:55 +00:00
|
|
|
if (ctrlr->msix_enabled)
|
|
|
|
pci_release_msi(dev);
|
|
|
|
pci_disable_busmaster(dev);
|
2019-08-22 20:09:32 +00:00
|
|
|
return (rv);
|
2019-08-21 22:17:55 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
nvme_ctrlr_configure_intx(struct nvme_controller *ctrlr)
|
|
|
|
{
|
|
|
|
|
|
|
|
ctrlr->msix_enabled = 0;
|
|
|
|
ctrlr->num_io_queues = 1;
|
|
|
|
ctrlr->rid = 0;
|
|
|
|
ctrlr->res = bus_alloc_resource_any(ctrlr->dev, SYS_RES_IRQ,
|
|
|
|
&ctrlr->rid, RF_SHAREABLE | RF_ACTIVE);
|
|
|
|
|
|
|
|
if (ctrlr->res == NULL) {
|
|
|
|
nvme_printf(ctrlr, "unable to allocate shared IRQ\n");
|
|
|
|
return (ENOMEM);
|
|
|
|
}
|
|
|
|
|
2020-05-02 20:47:58 +00:00
|
|
|
if (bus_setup_intr(ctrlr->dev, ctrlr->res,
|
2019-08-21 22:17:55 +00:00
|
|
|
INTR_TYPE_MISC | INTR_MPSAFE, NULL, nvme_ctrlr_intx_handler,
|
2020-05-02 20:47:58 +00:00
|
|
|
ctrlr, &ctrlr->tag) != 0) {
|
2019-08-21 22:17:55 +00:00
|
|
|
nvme_printf(ctrlr, "unable to setup intx handler\n");
|
|
|
|
return (ENOMEM);
|
|
|
|
}
|
|
|
|
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
nvme_ctrlr_setup_interrupts(struct nvme_controller *ctrlr)
|
|
|
|
{
|
|
|
|
device_t dev;
|
2019-09-23 17:53:47 +00:00
|
|
|
int force_intx, num_io_queues, per_cpu_io_queues;
|
2019-08-21 22:17:55 +00:00
|
|
|
int min_cpus_per_ioq;
|
|
|
|
int num_vectors_requested, num_vectors_allocated;
|
|
|
|
|
|
|
|
dev = ctrlr->dev;
|
|
|
|
|
2019-09-23 17:53:47 +00:00
|
|
|
force_intx = 0;
|
|
|
|
TUNABLE_INT_FETCH("hw.nvme.force_intx", &force_intx);
|
|
|
|
if (force_intx || pci_msix_count(dev) < 2) {
|
|
|
|
nvme_ctrlr_configure_intx(ctrlr);
|
|
|
|
return;
|
2019-08-21 22:17:55 +00:00
|
|
|
}
|
|
|
|
|
2019-09-23 17:53:47 +00:00
|
|
|
num_io_queues = mp_ncpus;
|
|
|
|
TUNABLE_INT_FETCH("hw.nvme.num_io_queues", &num_io_queues);
|
|
|
|
if (num_io_queues < 1 || num_io_queues > mp_ncpus)
|
|
|
|
num_io_queues = mp_ncpus;
|
|
|
|
|
2019-08-21 22:17:55 +00:00
|
|
|
per_cpu_io_queues = 1;
|
|
|
|
TUNABLE_INT_FETCH("hw.nvme.per_cpu_io_queues", &per_cpu_io_queues);
|
2019-09-23 17:53:47 +00:00
|
|
|
if (per_cpu_io_queues == 0)
|
|
|
|
num_io_queues = 1;
|
2019-08-21 22:17:55 +00:00
|
|
|
|
2019-09-23 17:53:47 +00:00
|
|
|
min_cpus_per_ioq = smp_threads_per_core;
|
|
|
|
TUNABLE_INT_FETCH("hw.nvme.min_cpus_per_ioq", &min_cpus_per_ioq);
|
|
|
|
if (min_cpus_per_ioq > 1) {
|
|
|
|
num_io_queues = min(num_io_queues,
|
|
|
|
max(1, mp_ncpus / min_cpus_per_ioq));
|
2019-08-21 22:17:55 +00:00
|
|
|
}
|
|
|
|
|
2019-09-23 17:53:47 +00:00
|
|
|
num_io_queues = min(num_io_queues, pci_msix_count(dev) - 1);
|
2019-08-21 22:17:55 +00:00
|
|
|
|
2019-09-23 17:53:47 +00:00
|
|
|
again:
|
|
|
|
if (num_io_queues > vm_ndomains)
|
|
|
|
num_io_queues -= num_io_queues % vm_ndomains;
|
2019-08-21 22:17:55 +00:00
|
|
|
/* One vector for per core I/O queue, plus one vector for admin queue. */
|
2019-09-23 17:53:47 +00:00
|
|
|
num_vectors_requested = num_io_queues + 1;
|
2019-08-21 22:17:55 +00:00
|
|
|
num_vectors_allocated = num_vectors_requested;
|
|
|
|
if (pci_alloc_msix(dev, &num_vectors_allocated) != 0) {
|
|
|
|
nvme_ctrlr_configure_intx(ctrlr);
|
|
|
|
return;
|
|
|
|
}
|
2019-09-23 17:53:47 +00:00
|
|
|
if (num_vectors_allocated < 2) {
|
2019-08-21 22:17:55 +00:00
|
|
|
pci_release_msi(dev);
|
|
|
|
nvme_ctrlr_configure_intx(ctrlr);
|
|
|
|
return;
|
|
|
|
}
|
2019-09-23 17:53:47 +00:00
|
|
|
if (num_vectors_allocated != num_vectors_requested) {
|
|
|
|
pci_release_msi(dev);
|
|
|
|
num_io_queues = num_vectors_allocated - 1;
|
|
|
|
goto again;
|
|
|
|
}
|
2019-08-21 22:17:55 +00:00
|
|
|
|
|
|
|
ctrlr->msix_enabled = 1;
|
2019-09-23 17:53:47 +00:00
|
|
|
ctrlr->num_io_queues = num_io_queues;
|
2019-08-21 22:17:55 +00:00
|
|
|
}
|
Implement nvme suspend / resume for pci attachment
When we suspend, we need to properly shutdown the NVME controller. The
controller may go into D3 state (or may have the power removed), and
to properly flush the metadata to non-volatile RAM, we must complete a
normal shutdown. This consists of deleting the I/O queues and setting
the shutodown bit. We have to do some extra stuff to make sure we
reset the software state of the queues as well.
On resume, we have to reset the card twice, for reasons described in
the attach funcion. Once we've done that, we can restart the card. If
any of this fails, we'll fail the NVMe card, just like we do when a
reset fails.
Set is_resetting for the duration of the suspend / resume. This keeps
the reset taskqueue from running a concurrent reset, and also is
needed to prevent any hw completions from queueing more I/O to the
card. Pass resetting flag to nvme_ctrlr_start. It doesn't need to get
that from the global state of the ctrlr. Wait for any pending reset to
finish. All queued I/O will get sent to the hardware as part of
nvme_ctrlr_start(), though the upper layers shouldn't send any
down. Disabling the qpairs is the other failsafe to ensure all I/O is
queued.
Rename nvme_ctrlr_destory_qpairs to nvme_ctrlr_delete_qpairs to avoid
confusion with all the other destroy functions. It just removes the
queues in hardware, while the other _destroy_ functions tear down
driver data structures.
Split parts of the hardware reset function up so that I can
do part of the reset in suspsend. Split out the software disabling
of the qpairs into nvme_ctrlr_disable_qpairs.
Finally, fix a couple of spelling errors in comments related to
this.
Relnotes: Yes
MFC After: 1 week
Reviewed by: scottl@ (prior version)
Differential Revision: https://reviews.freebsd.org/D21493
2019-09-03 15:26:11 +00:00
|
|
|
|
|
|
|
static int
|
|
|
|
nvme_pci_suspend(device_t dev)
|
|
|
|
{
|
|
|
|
struct nvme_controller *ctrlr;
|
|
|
|
|
|
|
|
ctrlr = DEVICE2SOFTC(dev);
|
|
|
|
return (nvme_ctrlr_suspend(ctrlr));
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
nvme_pci_resume(device_t dev)
|
|
|
|
{
|
|
|
|
struct nvme_controller *ctrlr;
|
|
|
|
|
|
|
|
ctrlr = DEVICE2SOFTC(dev);
|
|
|
|
return (nvme_ctrlr_resume(ctrlr));
|
|
|
|
}
|