45f8cc9a6c
- In sys/ofed/drivers/infiniband/core/cma.c, an enum struct member is interpreted as an int, so cast it to an int. - In sys/ofed/drivers/infiniband/core/ud_header.c, initialize the packet_length variable in ib_ud_header_init(), to prevent undefined behaviour. - In sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c, call rdma_notify() with the correct enum type and value. - In sys/ofed/include/linux/pci.h, change the PCI_DEVICE and PCI_VDEVICE macros to use C99 struct initializers, so additional members can be overridden. Reviewed by: delphij, Garrett Cooper <yanegomi@gmail.com> MFC after: 1 week
583 lines
14 KiB
C
583 lines
14 KiB
C
/*-
|
|
* Copyright (c) 2010 Isilon Systems, Inc.
|
|
* Copyright (c) 2010 iX Systems, Inc.
|
|
* Copyright (c) 2010 Panasas, Inc.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice unmodified, this list of conditions, and the following
|
|
* disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
|
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
|
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
|
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
|
|
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
|
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
|
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#ifndef _LINUX_PCI_H_
|
|
#define _LINUX_PCI_H_
|
|
|
|
#define CONFIG_PCI_MSI
|
|
|
|
#include <linux/types.h>
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/bus.h>
|
|
#include <sys/pciio.h>
|
|
#include <sys/rman.h>
|
|
#include <dev/pci/pcivar.h>
|
|
#include <dev/pci/pcireg.h>
|
|
#include <dev/pci/pci_private.h>
|
|
|
|
#include <machine/resource.h>
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/list.h>
|
|
#include <linux/dmapool.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/compiler.h>
|
|
#include <linux/errno.h>
|
|
#include <asm/atomic.h>
|
|
#include <linux/device.h>
|
|
|
|
struct pci_device_id {
|
|
uint32_t vendor;
|
|
uint32_t device;
|
|
uint32_t subvendor;
|
|
uint32_t subdevice;
|
|
uint32_t class_mask;
|
|
uintptr_t driver_data;
|
|
};
|
|
|
|
#define MODULE_DEVICE_TABLE(bus, table)
|
|
#define PCI_ANY_ID (-1)
|
|
#define PCI_VENDOR_ID_MELLANOX 0x15b3
|
|
#define PCI_VENDOR_ID_TOPSPIN 0x1867
|
|
#define PCI_DEVICE_ID_MELLANOX_TAVOR 0x5a44
|
|
#define PCI_DEVICE_ID_MELLANOX_TAVOR_BRIDGE 0x5a46
|
|
#define PCI_DEVICE_ID_MELLANOX_ARBEL_COMPAT 0x6278
|
|
#define PCI_DEVICE_ID_MELLANOX_ARBEL 0x6282
|
|
#define PCI_DEVICE_ID_MELLANOX_SINAI_OLD 0x5e8c
|
|
#define PCI_DEVICE_ID_MELLANOX_SINAI 0x6274
|
|
|
|
|
|
#define PCI_VDEVICE(_vendor, _device) \
|
|
.vendor = PCI_VENDOR_ID_##_vendor, .device = (_device), \
|
|
.subvendor = PCI_ANY_ID, .subdevice = PCI_ANY_ID
|
|
#define PCI_DEVICE(_vendor, _device) \
|
|
.vendor = (_vendor), .device = (_device), \
|
|
.subvendor = PCI_ANY_ID, .subdevice = PCI_ANY_ID
|
|
|
|
#define to_pci_dev(n) container_of(n, struct pci_dev, dev)
|
|
|
|
#define PCI_VENDOR_ID PCIR_DEVVENDOR
|
|
#define PCI_COMMAND PCIR_COMMAND
|
|
#define PCI_EXP_DEVCTL PCIER_DEVICE_CTL
|
|
#define PCI_EXP_LNKCTL PCIER_LINK_CTL
|
|
|
|
#define IORESOURCE_MEM SYS_RES_MEMORY
|
|
#define IORESOURCE_IO SYS_RES_IOPORT
|
|
#define IORESOURCE_IRQ SYS_RES_IRQ
|
|
|
|
struct pci_dev;
|
|
|
|
struct pci_driver {
|
|
struct list_head links;
|
|
char *name;
|
|
struct pci_device_id *id_table;
|
|
int (*probe)(struct pci_dev *dev, const struct pci_device_id *id);
|
|
void (*remove)(struct pci_dev *dev);
|
|
driver_t driver;
|
|
devclass_t bsdclass;
|
|
};
|
|
|
|
extern struct list_head pci_drivers;
|
|
extern struct list_head pci_devices;
|
|
extern spinlock_t pci_lock;
|
|
|
|
#define __devexit_p(x) x
|
|
|
|
struct pci_dev {
|
|
struct device dev;
|
|
struct list_head links;
|
|
struct pci_driver *pdrv;
|
|
uint64_t dma_mask;
|
|
uint16_t device;
|
|
uint16_t vendor;
|
|
unsigned int irq;
|
|
};
|
|
|
|
static inline struct resource_list_entry *
|
|
_pci_get_rle(struct pci_dev *pdev, int type, int rid)
|
|
{
|
|
struct pci_devinfo *dinfo;
|
|
struct resource_list *rl;
|
|
|
|
dinfo = device_get_ivars(pdev->dev.bsddev);
|
|
rl = &dinfo->resources;
|
|
return resource_list_find(rl, type, rid);
|
|
}
|
|
|
|
static inline struct resource_list_entry *
|
|
_pci_get_bar(struct pci_dev *pdev, int bar)
|
|
{
|
|
struct resource_list_entry *rle;
|
|
|
|
bar = PCIR_BAR(bar);
|
|
if ((rle = _pci_get_rle(pdev, SYS_RES_MEMORY, bar)) == NULL)
|
|
rle = _pci_get_rle(pdev, SYS_RES_IOPORT, bar);
|
|
return (rle);
|
|
}
|
|
|
|
static inline struct device *
|
|
_pci_find_irq_dev(unsigned int irq)
|
|
{
|
|
struct pci_dev *pdev;
|
|
|
|
spin_lock(&pci_lock);
|
|
list_for_each_entry(pdev, &pci_devices, links) {
|
|
if (irq == pdev->dev.irq)
|
|
break;
|
|
if (irq >= pdev->dev.msix && irq < pdev->dev.msix_max)
|
|
break;
|
|
}
|
|
spin_unlock(&pci_lock);
|
|
if (pdev)
|
|
return &pdev->dev;
|
|
return (NULL);
|
|
}
|
|
|
|
static inline unsigned long
|
|
pci_resource_start(struct pci_dev *pdev, int bar)
|
|
{
|
|
struct resource_list_entry *rle;
|
|
|
|
if ((rle = _pci_get_bar(pdev, bar)) == NULL)
|
|
return (0);
|
|
return rle->start;
|
|
}
|
|
|
|
static inline unsigned long
|
|
pci_resource_len(struct pci_dev *pdev, int bar)
|
|
{
|
|
struct resource_list_entry *rle;
|
|
|
|
if ((rle = _pci_get_bar(pdev, bar)) == NULL)
|
|
return (0);
|
|
return rle->count;
|
|
}
|
|
|
|
/*
|
|
* All drivers just seem to want to inspect the type not flags.
|
|
*/
|
|
static inline int
|
|
pci_resource_flags(struct pci_dev *pdev, int bar)
|
|
{
|
|
struct resource_list_entry *rle;
|
|
|
|
if ((rle = _pci_get_bar(pdev, bar)) == NULL)
|
|
return (0);
|
|
return rle->type;
|
|
}
|
|
|
|
static inline const char *
|
|
pci_name(struct pci_dev *d)
|
|
{
|
|
|
|
return device_get_desc(d->dev.bsddev);
|
|
}
|
|
|
|
static inline void *
|
|
pci_get_drvdata(struct pci_dev *pdev)
|
|
{
|
|
|
|
return dev_get_drvdata(&pdev->dev);
|
|
}
|
|
|
|
static inline void
|
|
pci_set_drvdata(struct pci_dev *pdev, void *data)
|
|
{
|
|
|
|
dev_set_drvdata(&pdev->dev, data);
|
|
}
|
|
|
|
static inline int
|
|
pci_enable_device(struct pci_dev *pdev)
|
|
{
|
|
|
|
pci_enable_io(pdev->dev.bsddev, SYS_RES_IOPORT);
|
|
pci_enable_io(pdev->dev.bsddev, SYS_RES_MEMORY);
|
|
return (0);
|
|
}
|
|
|
|
static inline void
|
|
pci_disable_device(struct pci_dev *pdev)
|
|
{
|
|
}
|
|
|
|
static inline int
|
|
pci_set_master(struct pci_dev *pdev)
|
|
{
|
|
|
|
pci_enable_busmaster(pdev->dev.bsddev);
|
|
return (0);
|
|
}
|
|
|
|
static inline int
|
|
pci_request_region(struct pci_dev *pdev, int bar, const char *res_name)
|
|
{
|
|
int rid;
|
|
int type;
|
|
|
|
type = pci_resource_flags(pdev, bar);
|
|
if (type == 0)
|
|
return (-ENODEV);
|
|
rid = PCIR_BAR(bar);
|
|
if (bus_alloc_resource_any(pdev->dev.bsddev, type, &rid,
|
|
RF_ACTIVE) == NULL)
|
|
return (-EINVAL);
|
|
return (0);
|
|
}
|
|
|
|
static inline void
|
|
pci_release_region(struct pci_dev *pdev, int bar)
|
|
{
|
|
struct resource_list_entry *rle;
|
|
|
|
if ((rle = _pci_get_bar(pdev, bar)) == NULL)
|
|
return;
|
|
bus_release_resource(pdev->dev.bsddev, rle->type, rle->rid, rle->res);
|
|
}
|
|
|
|
static inline void
|
|
pci_release_regions(struct pci_dev *pdev)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i <= PCIR_MAX_BAR_0; i++)
|
|
pci_release_region(pdev, i);
|
|
}
|
|
|
|
static inline int
|
|
pci_request_regions(struct pci_dev *pdev, const char *res_name)
|
|
{
|
|
int error;
|
|
int i;
|
|
|
|
for (i = 0; i <= PCIR_MAX_BAR_0; i++) {
|
|
error = pci_request_region(pdev, i, res_name);
|
|
if (error && error != -ENODEV) {
|
|
pci_release_regions(pdev);
|
|
return (error);
|
|
}
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
static inline void
|
|
pci_disable_msix(struct pci_dev *pdev)
|
|
{
|
|
|
|
pci_release_msi(pdev->dev.bsddev);
|
|
}
|
|
|
|
#define PCI_CAP_ID_EXP PCIY_EXPRESS
|
|
#define PCI_CAP_ID_PCIX PCIY_PCIX
|
|
|
|
static inline int
|
|
pci_find_capability(struct pci_dev *pdev, int capid)
|
|
{
|
|
int reg;
|
|
|
|
if (pci_find_cap(pdev->dev.bsddev, capid, ®))
|
|
return (0);
|
|
return (reg);
|
|
}
|
|
|
|
static inline int
|
|
pci_read_config_byte(struct pci_dev *pdev, int where, u8 *val)
|
|
{
|
|
|
|
*val = (u8)pci_read_config(pdev->dev.bsddev, where, 1);
|
|
return (0);
|
|
}
|
|
|
|
static inline int
|
|
pci_read_config_word(struct pci_dev *pdev, int where, u16 *val)
|
|
{
|
|
|
|
*val = (u16)pci_read_config(pdev->dev.bsddev, where, 2);
|
|
return (0);
|
|
}
|
|
|
|
static inline int
|
|
pci_read_config_dword(struct pci_dev *pdev, int where, u32 *val)
|
|
{
|
|
|
|
*val = (u32)pci_read_config(pdev->dev.bsddev, where, 4);
|
|
return (0);
|
|
}
|
|
|
|
static inline int
|
|
pci_write_config_byte(struct pci_dev *pdev, int where, u8 val)
|
|
{
|
|
|
|
pci_write_config(pdev->dev.bsddev, where, val, 1);
|
|
return (0);
|
|
}
|
|
|
|
static inline int
|
|
pci_write_config_word(struct pci_dev *pdev, int where, u16 val)
|
|
{
|
|
|
|
pci_write_config(pdev->dev.bsddev, where, val, 2);
|
|
return (0);
|
|
}
|
|
|
|
static inline int
|
|
pci_write_config_dword(struct pci_dev *pdev, int where, u32 val)
|
|
{
|
|
|
|
pci_write_config(pdev->dev.bsddev, where, val, 4);
|
|
return (0);
|
|
}
|
|
|
|
static struct pci_driver *
|
|
linux_pci_find(device_t dev, struct pci_device_id **idp)
|
|
{
|
|
struct pci_device_id *id;
|
|
struct pci_driver *pdrv;
|
|
uint16_t vendor;
|
|
uint16_t device;
|
|
|
|
vendor = pci_get_vendor(dev);
|
|
device = pci_get_device(dev);
|
|
|
|
spin_lock(&pci_lock);
|
|
list_for_each_entry(pdrv, &pci_drivers, links) {
|
|
for (id = pdrv->id_table; id->vendor != 0; id++) {
|
|
if (vendor == id->vendor && device == id->device) {
|
|
*idp = id;
|
|
spin_unlock(&pci_lock);
|
|
return (pdrv);
|
|
}
|
|
}
|
|
}
|
|
spin_unlock(&pci_lock);
|
|
return (NULL);
|
|
}
|
|
|
|
static inline int
|
|
linux_pci_probe(device_t dev)
|
|
{
|
|
struct pci_device_id *id;
|
|
struct pci_driver *pdrv;
|
|
|
|
if ((pdrv = linux_pci_find(dev, &id)) == NULL)
|
|
return (ENXIO);
|
|
if (device_get_driver(dev) != &pdrv->driver)
|
|
return (ENXIO);
|
|
device_set_desc(dev, pdrv->name);
|
|
return (0);
|
|
}
|
|
|
|
static inline int
|
|
linux_pci_attach(device_t dev)
|
|
{
|
|
struct resource_list_entry *rle;
|
|
struct pci_dev *pdev;
|
|
struct pci_driver *pdrv;
|
|
struct pci_device_id *id;
|
|
int error;
|
|
|
|
pdrv = linux_pci_find(dev, &id);
|
|
pdev = device_get_softc(dev);
|
|
pdev->dev.parent = &linux_rootdev;
|
|
pdev->dev.bsddev = dev;
|
|
INIT_LIST_HEAD(&pdev->dev.irqents);
|
|
pdev->device = id->device;
|
|
pdev->vendor = id->vendor;
|
|
pdev->dev.dma_mask = &pdev->dma_mask;
|
|
pdev->pdrv = pdrv;
|
|
kobject_init(&pdev->dev.kobj, &dev_ktype);
|
|
kobject_set_name(&pdev->dev.kobj, device_get_nameunit(dev));
|
|
kobject_add(&pdev->dev.kobj, &linux_rootdev.kobj,
|
|
kobject_name(&pdev->dev.kobj));
|
|
rle = _pci_get_rle(pdev, SYS_RES_IRQ, 0);
|
|
if (rle)
|
|
pdev->dev.irq = rle->start;
|
|
else
|
|
pdev->dev.irq = 0;
|
|
pdev->irq = pdev->dev.irq;
|
|
mtx_unlock(&Giant);
|
|
spin_lock(&pci_lock);
|
|
list_add(&pdev->links, &pci_devices);
|
|
spin_unlock(&pci_lock);
|
|
error = pdrv->probe(pdev, id);
|
|
mtx_lock(&Giant);
|
|
if (error) {
|
|
spin_lock(&pci_lock);
|
|
list_del(&pdev->links);
|
|
spin_unlock(&pci_lock);
|
|
put_device(&pdev->dev);
|
|
return (-error);
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
static inline int
|
|
linux_pci_detach(device_t dev)
|
|
{
|
|
struct pci_dev *pdev;
|
|
|
|
pdev = device_get_softc(dev);
|
|
mtx_unlock(&Giant);
|
|
pdev->pdrv->remove(pdev);
|
|
mtx_lock(&Giant);
|
|
spin_lock(&pci_lock);
|
|
list_del(&pdev->links);
|
|
spin_unlock(&pci_lock);
|
|
put_device(&pdev->dev);
|
|
|
|
return (0);
|
|
}
|
|
|
|
static device_method_t pci_methods[] = {
|
|
DEVMETHOD(device_probe, linux_pci_probe),
|
|
DEVMETHOD(device_attach, linux_pci_attach),
|
|
DEVMETHOD(device_detach, linux_pci_detach),
|
|
{0, 0}
|
|
};
|
|
|
|
static inline int
|
|
pci_register_driver(struct pci_driver *pdrv)
|
|
{
|
|
devclass_t bus;
|
|
int error;
|
|
|
|
spin_lock(&pci_lock);
|
|
list_add(&pdrv->links, &pci_drivers);
|
|
spin_unlock(&pci_lock);
|
|
bus = devclass_find("pci");
|
|
pdrv->driver.name = pdrv->name;
|
|
pdrv->driver.methods = pci_methods;
|
|
pdrv->driver.size = sizeof(struct pci_dev);
|
|
mtx_lock(&Giant);
|
|
error = devclass_add_driver(bus, &pdrv->driver, BUS_PASS_DEFAULT,
|
|
&pdrv->bsdclass);
|
|
mtx_unlock(&Giant);
|
|
if (error)
|
|
return (-error);
|
|
return (0);
|
|
}
|
|
|
|
static inline void
|
|
pci_unregister_driver(struct pci_driver *pdrv)
|
|
{
|
|
devclass_t bus;
|
|
|
|
list_del(&pdrv->links);
|
|
bus = devclass_find("pci");
|
|
mtx_lock(&Giant);
|
|
devclass_delete_driver(bus, &pdrv->driver);
|
|
mtx_unlock(&Giant);
|
|
}
|
|
|
|
struct msix_entry {
|
|
int entry;
|
|
int vector;
|
|
};
|
|
|
|
/*
|
|
* Enable msix, positive errors indicate actual number of available
|
|
* vectors. Negative errors are failures.
|
|
*/
|
|
static inline int
|
|
pci_enable_msix(struct pci_dev *pdev, struct msix_entry *entries, int nreq)
|
|
{
|
|
struct resource_list_entry *rle;
|
|
int error;
|
|
int avail;
|
|
int i;
|
|
|
|
avail = pci_msix_count(pdev->dev.bsddev);
|
|
if (avail < nreq) {
|
|
if (avail == 0)
|
|
return -EINVAL;
|
|
return avail;
|
|
}
|
|
avail = nreq;
|
|
if ((error = -pci_alloc_msix(pdev->dev.bsddev, &avail)) != 0)
|
|
return error;
|
|
rle = _pci_get_rle(pdev, SYS_RES_IRQ, 1);
|
|
pdev->dev.msix = rle->start;
|
|
pdev->dev.msix_max = rle->start + avail;
|
|
for (i = 0; i < nreq; i++)
|
|
entries[i].vector = pdev->dev.msix + i;
|
|
return (0);
|
|
}
|
|
|
|
/* XXX This should not be necessary. */
|
|
#define pcix_set_mmrbc(d, v) 0
|
|
#define pcix_get_max_mmrbc(d) 0
|
|
#define pcie_set_readrq(d, v) 0
|
|
|
|
#define PCI_DMA_BIDIRECTIONAL 0
|
|
#define PCI_DMA_TODEVICE 1
|
|
#define PCI_DMA_FROMDEVICE 2
|
|
#define PCI_DMA_NONE 3
|
|
|
|
#define pci_pool dma_pool
|
|
#define pci_pool_destroy dma_pool_destroy
|
|
#define pci_pool_alloc dma_pool_alloc
|
|
#define pci_pool_free dma_pool_free
|
|
#define pci_pool_create(_name, _pdev, _size, _align, _alloc) \
|
|
dma_pool_create(_name, &(_pdev)->dev, _size, _align, _alloc)
|
|
#define pci_free_consistent(_hwdev, _size, _vaddr, _dma_handle) \
|
|
dma_free_coherent((_hwdev) == NULL ? NULL : &(_hwdev)->dev, \
|
|
_size, _vaddr, _dma_handle)
|
|
#define pci_map_sg(_hwdev, _sg, _nents, _dir) \
|
|
dma_map_sg((_hwdev) == NULL ? NULL : &(_hwdev->dev), \
|
|
_sg, _nents, (enum dma_data_direction)_dir)
|
|
#define pci_map_single(_hwdev, _ptr, _size, _dir) \
|
|
dma_map_single((_hwdev) == NULL ? NULL : &(_hwdev->dev), \
|
|
(_ptr), (_size), (enum dma_data_direction)_dir)
|
|
#define pci_unmap_single(_hwdev, _addr, _size, _dir) \
|
|
dma_unmap_single((_hwdev) == NULL ? NULL : &(_hwdev)->dev, \
|
|
_addr, _size, (enum dma_data_direction)_dir)
|
|
#define pci_unmap_sg(_hwdev, _sg, _nents, _dir) \
|
|
dma_unmap_sg((_hwdev) == NULL ? NULL : &(_hwdev)->dev, \
|
|
_sg, _nents, (enum dma_data_direction)_dir)
|
|
#define pci_map_page(_hwdev, _page, _offset, _size, _dir) \
|
|
dma_map_page((_hwdev) == NULL ? NULL : &(_hwdev)->dev, _page,\
|
|
_offset, _size, (enum dma_data_direction)_dir)
|
|
#define pci_unmap_page(_hwdev, _dma_address, _size, _dir) \
|
|
dma_unmap_page((_hwdev) == NULL ? NULL : &(_hwdev)->dev, \
|
|
_dma_address, _size, (enum dma_data_direction)_dir)
|
|
#define pci_set_dma_mask(_pdev, mask) dma_set_mask(&(_pdev)->dev, (mask))
|
|
#define pci_dma_mapping_error(_pdev, _dma_addr) \
|
|
dma_mapping_error(&(_pdev)->dev, _dma_addr)
|
|
#define pci_set_consistent_dma_mask(_pdev, _mask) \
|
|
dma_set_coherent_mask(&(_pdev)->dev, (_mask))
|
|
#define DECLARE_PCI_UNMAP_ADDR(x) DEFINE_DMA_UNMAP_ADDR(x);
|
|
#define DECLARE_PCI_UNMAP_LEN(x) DEFINE_DMA_UNMAP_LEN(x);
|
|
#define pci_unmap_addr dma_unmap_addr
|
|
#define pci_unmap_addr_set dma_unmap_addr_set
|
|
#define pci_unmap_len dma_unmap_len
|
|
#define pci_unmap_len_set dma_unmap_len_set
|
|
|
|
|
|
#endif /* _LINUX_PCI_H_ */
|