fd90e2ed54
years for head. However, it is continuously misused as the mpsafe argument for callout_init(9). Deprecate the flag and clean up callout_init() calls to make them more consistent. Differential Revision: https://reviews.freebsd.org/D2613 Reviewed by: jhb MFC after: 2 weeks
3575 lines
88 KiB
C
3575 lines
88 KiB
C
/**************************************************************************
|
|
|
|
Copyright (c) 2007-2009, Chelsio Inc.
|
|
All rights reserved.
|
|
|
|
Redistribution and use in source and binary forms, with or without
|
|
modification, are permitted provided that the following conditions are met:
|
|
|
|
1. Redistributions of source code must retain the above copyright notice,
|
|
this list of conditions and the following disclaimer.
|
|
|
|
2. Neither the name of the Chelsio Corporation nor the names of its
|
|
contributors may be used to endorse or promote products derived from
|
|
this software without specific prior written permission.
|
|
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
***************************************************************************/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include "opt_inet.h"
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/bus.h>
|
|
#include <sys/module.h>
|
|
#include <sys/pciio.h>
|
|
#include <sys/conf.h>
|
|
#include <machine/bus.h>
|
|
#include <machine/resource.h>
|
|
#include <sys/bus_dma.h>
|
|
#include <sys/ktr.h>
|
|
#include <sys/rman.h>
|
|
#include <sys/ioccom.h>
|
|
#include <sys/mbuf.h>
|
|
#include <sys/linker.h>
|
|
#include <sys/firmware.h>
|
|
#include <sys/socket.h>
|
|
#include <sys/sockio.h>
|
|
#include <sys/smp.h>
|
|
#include <sys/sysctl.h>
|
|
#include <sys/syslog.h>
|
|
#include <sys/queue.h>
|
|
#include <sys/taskqueue.h>
|
|
#include <sys/proc.h>
|
|
|
|
#include <net/bpf.h>
|
|
#include <net/ethernet.h>
|
|
#include <net/if.h>
|
|
#include <net/if_var.h>
|
|
#include <net/if_arp.h>
|
|
#include <net/if_dl.h>
|
|
#include <net/if_media.h>
|
|
#include <net/if_types.h>
|
|
#include <net/if_vlan_var.h>
|
|
|
|
#include <netinet/in_systm.h>
|
|
#include <netinet/in.h>
|
|
#include <netinet/if_ether.h>
|
|
#include <netinet/ip.h>
|
|
#include <netinet/ip.h>
|
|
#include <netinet/tcp.h>
|
|
#include <netinet/udp.h>
|
|
|
|
#include <dev/pci/pcireg.h>
|
|
#include <dev/pci/pcivar.h>
|
|
#include <dev/pci/pci_private.h>
|
|
|
|
#include <cxgb_include.h>
|
|
|
|
#ifdef PRIV_SUPPORTED
|
|
#include <sys/priv.h>
|
|
#endif
|
|
|
|
static int cxgb_setup_interrupts(adapter_t *);
|
|
static void cxgb_teardown_interrupts(adapter_t *);
|
|
static void cxgb_init(void *);
|
|
static int cxgb_init_locked(struct port_info *);
|
|
static int cxgb_uninit_locked(struct port_info *);
|
|
static int cxgb_uninit_synchronized(struct port_info *);
|
|
static int cxgb_ioctl(struct ifnet *, unsigned long, caddr_t);
|
|
static int cxgb_media_change(struct ifnet *);
|
|
static int cxgb_ifm_type(int);
|
|
static void cxgb_build_medialist(struct port_info *);
|
|
static void cxgb_media_status(struct ifnet *, struct ifmediareq *);
|
|
static uint64_t cxgb_get_counter(struct ifnet *, ift_counter);
|
|
static int setup_sge_qsets(adapter_t *);
|
|
static void cxgb_async_intr(void *);
|
|
static void cxgb_tick_handler(void *, int);
|
|
static void cxgb_tick(void *);
|
|
static void link_check_callout(void *);
|
|
static void check_link_status(void *, int);
|
|
static void setup_rss(adapter_t *sc);
|
|
static int alloc_filters(struct adapter *);
|
|
static int setup_hw_filters(struct adapter *);
|
|
static int set_filter(struct adapter *, int, const struct filter_info *);
|
|
static inline void mk_set_tcb_field(struct cpl_set_tcb_field *, unsigned int,
|
|
unsigned int, u64, u64);
|
|
static inline void set_tcb_field_ulp(struct cpl_set_tcb_field *, unsigned int,
|
|
unsigned int, u64, u64);
|
|
#ifdef TCP_OFFLOAD
|
|
static int cpl_not_handled(struct sge_qset *, struct rsp_desc *, struct mbuf *);
|
|
#endif
|
|
|
|
/* Attachment glue for the PCI controller end of the device. Each port of
|
|
* the device is attached separately, as defined later.
|
|
*/
|
|
static int cxgb_controller_probe(device_t);
|
|
static int cxgb_controller_attach(device_t);
|
|
static int cxgb_controller_detach(device_t);
|
|
static void cxgb_free(struct adapter *);
|
|
static __inline void reg_block_dump(struct adapter *ap, uint8_t *buf, unsigned int start,
|
|
unsigned int end);
|
|
static void cxgb_get_regs(adapter_t *sc, struct ch_ifconf_regs *regs, uint8_t *buf);
|
|
static int cxgb_get_regs_len(void);
|
|
static void touch_bars(device_t dev);
|
|
static void cxgb_update_mac_settings(struct port_info *p);
|
|
#ifdef TCP_OFFLOAD
|
|
static int toe_capability(struct port_info *, int);
|
|
#endif
|
|
|
|
static device_method_t cxgb_controller_methods[] = {
|
|
DEVMETHOD(device_probe, cxgb_controller_probe),
|
|
DEVMETHOD(device_attach, cxgb_controller_attach),
|
|
DEVMETHOD(device_detach, cxgb_controller_detach),
|
|
|
|
DEVMETHOD_END
|
|
};
|
|
|
|
static driver_t cxgb_controller_driver = {
|
|
"cxgbc",
|
|
cxgb_controller_methods,
|
|
sizeof(struct adapter)
|
|
};
|
|
|
|
static int cxgbc_mod_event(module_t, int, void *);
|
|
static devclass_t cxgb_controller_devclass;
|
|
DRIVER_MODULE(cxgbc, pci, cxgb_controller_driver, cxgb_controller_devclass,
|
|
cxgbc_mod_event, 0);
|
|
MODULE_VERSION(cxgbc, 1);
|
|
MODULE_DEPEND(cxgbc, firmware, 1, 1, 1);
|
|
|
|
/*
|
|
* Attachment glue for the ports. Attachment is done directly to the
|
|
* controller device.
|
|
*/
|
|
static int cxgb_port_probe(device_t);
|
|
static int cxgb_port_attach(device_t);
|
|
static int cxgb_port_detach(device_t);
|
|
|
|
static device_method_t cxgb_port_methods[] = {
|
|
DEVMETHOD(device_probe, cxgb_port_probe),
|
|
DEVMETHOD(device_attach, cxgb_port_attach),
|
|
DEVMETHOD(device_detach, cxgb_port_detach),
|
|
{ 0, 0 }
|
|
};
|
|
|
|
static driver_t cxgb_port_driver = {
|
|
"cxgb",
|
|
cxgb_port_methods,
|
|
0
|
|
};
|
|
|
|
static d_ioctl_t cxgb_extension_ioctl;
|
|
static d_open_t cxgb_extension_open;
|
|
static d_close_t cxgb_extension_close;
|
|
|
|
static struct cdevsw cxgb_cdevsw = {
|
|
.d_version = D_VERSION,
|
|
.d_flags = 0,
|
|
.d_open = cxgb_extension_open,
|
|
.d_close = cxgb_extension_close,
|
|
.d_ioctl = cxgb_extension_ioctl,
|
|
.d_name = "cxgb",
|
|
};
|
|
|
|
static devclass_t cxgb_port_devclass;
|
|
DRIVER_MODULE(cxgb, cxgbc, cxgb_port_driver, cxgb_port_devclass, 0, 0);
|
|
MODULE_VERSION(cxgb, 1);
|
|
|
|
static struct mtx t3_list_lock;
|
|
static SLIST_HEAD(, adapter) t3_list;
|
|
#ifdef TCP_OFFLOAD
|
|
static struct mtx t3_uld_list_lock;
|
|
static SLIST_HEAD(, uld_info) t3_uld_list;
|
|
#endif
|
|
|
|
/*
|
|
* The driver uses the best interrupt scheme available on a platform in the
|
|
* order MSI-X, MSI, legacy pin interrupts. This parameter determines which
|
|
* of these schemes the driver may consider as follows:
|
|
*
|
|
* msi = 2: choose from among all three options
|
|
* msi = 1 : only consider MSI and pin interrupts
|
|
* msi = 0: force pin interrupts
|
|
*/
|
|
static int msi_allowed = 2;
|
|
|
|
SYSCTL_NODE(_hw, OID_AUTO, cxgb, CTLFLAG_RD, 0, "CXGB driver parameters");
|
|
SYSCTL_INT(_hw_cxgb, OID_AUTO, msi_allowed, CTLFLAG_RDTUN, &msi_allowed, 0,
|
|
"MSI-X, MSI, INTx selector");
|
|
|
|
/*
|
|
* The driver uses an auto-queue algorithm by default.
|
|
* To disable it and force a single queue-set per port, use multiq = 0
|
|
*/
|
|
static int multiq = 1;
|
|
SYSCTL_INT(_hw_cxgb, OID_AUTO, multiq, CTLFLAG_RDTUN, &multiq, 0,
|
|
"use min(ncpus/ports, 8) queue-sets per port");
|
|
|
|
/*
|
|
* By default the driver will not update the firmware unless
|
|
* it was compiled against a newer version
|
|
*
|
|
*/
|
|
static int force_fw_update = 0;
|
|
SYSCTL_INT(_hw_cxgb, OID_AUTO, force_fw_update, CTLFLAG_RDTUN, &force_fw_update, 0,
|
|
"update firmware even if up to date");
|
|
|
|
int cxgb_use_16k_clusters = -1;
|
|
SYSCTL_INT(_hw_cxgb, OID_AUTO, use_16k_clusters, CTLFLAG_RDTUN,
|
|
&cxgb_use_16k_clusters, 0, "use 16kB clusters for the jumbo queue ");
|
|
|
|
static int nfilters = -1;
|
|
SYSCTL_INT(_hw_cxgb, OID_AUTO, nfilters, CTLFLAG_RDTUN,
|
|
&nfilters, 0, "max number of entries in the filter table");
|
|
|
|
enum {
|
|
MAX_TXQ_ENTRIES = 16384,
|
|
MAX_CTRL_TXQ_ENTRIES = 1024,
|
|
MAX_RSPQ_ENTRIES = 16384,
|
|
MAX_RX_BUFFERS = 16384,
|
|
MAX_RX_JUMBO_BUFFERS = 16384,
|
|
MIN_TXQ_ENTRIES = 4,
|
|
MIN_CTRL_TXQ_ENTRIES = 4,
|
|
MIN_RSPQ_ENTRIES = 32,
|
|
MIN_FL_ENTRIES = 32,
|
|
MIN_FL_JUMBO_ENTRIES = 32
|
|
};
|
|
|
|
struct filter_info {
|
|
u32 sip;
|
|
u32 sip_mask;
|
|
u32 dip;
|
|
u16 sport;
|
|
u16 dport;
|
|
u32 vlan:12;
|
|
u32 vlan_prio:3;
|
|
u32 mac_hit:1;
|
|
u32 mac_idx:4;
|
|
u32 mac_vld:1;
|
|
u32 pkt_type:2;
|
|
u32 report_filter_id:1;
|
|
u32 pass:1;
|
|
u32 rss:1;
|
|
u32 qset:3;
|
|
u32 locked:1;
|
|
u32 valid:1;
|
|
};
|
|
|
|
enum { FILTER_NO_VLAN_PRI = 7 };
|
|
|
|
#define EEPROM_MAGIC 0x38E2F10C
|
|
|
|
#define PORT_MASK ((1 << MAX_NPORTS) - 1)
|
|
|
|
/* Table for probing the cards. The desc field isn't actually used */
|
|
struct cxgb_ident {
|
|
uint16_t vendor;
|
|
uint16_t device;
|
|
int index;
|
|
char *desc;
|
|
} cxgb_identifiers[] = {
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0020, 0, "PE9000"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0021, 1, "T302E"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0022, 2, "T310E"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0023, 3, "T320X"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0024, 1, "T302X"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0025, 3, "T320E"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0026, 2, "T310X"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0030, 2, "T3B10"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0031, 3, "T3B20"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0032, 1, "T3B02"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0033, 4, "T3B04"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0035, 6, "T3C10"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0036, 3, "S320E-CR"},
|
|
{PCI_VENDOR_ID_CHELSIO, 0x0037, 7, "N320E-G2"},
|
|
{0, 0, 0, NULL}
|
|
};
|
|
|
|
static int set_eeprom(struct port_info *pi, const uint8_t *data, int len, int offset);
|
|
|
|
|
|
static __inline char
|
|
t3rev2char(struct adapter *adapter)
|
|
{
|
|
char rev = 'z';
|
|
|
|
switch(adapter->params.rev) {
|
|
case T3_REV_A:
|
|
rev = 'a';
|
|
break;
|
|
case T3_REV_B:
|
|
case T3_REV_B2:
|
|
rev = 'b';
|
|
break;
|
|
case T3_REV_C:
|
|
rev = 'c';
|
|
break;
|
|
}
|
|
return rev;
|
|
}
|
|
|
|
static struct cxgb_ident *
|
|
cxgb_get_ident(device_t dev)
|
|
{
|
|
struct cxgb_ident *id;
|
|
|
|
for (id = cxgb_identifiers; id->desc != NULL; id++) {
|
|
if ((id->vendor == pci_get_vendor(dev)) &&
|
|
(id->device == pci_get_device(dev))) {
|
|
return (id);
|
|
}
|
|
}
|
|
return (NULL);
|
|
}
|
|
|
|
static const struct adapter_info *
|
|
cxgb_get_adapter_info(device_t dev)
|
|
{
|
|
struct cxgb_ident *id;
|
|
const struct adapter_info *ai;
|
|
|
|
id = cxgb_get_ident(dev);
|
|
if (id == NULL)
|
|
return (NULL);
|
|
|
|
ai = t3_get_adapter_info(id->index);
|
|
|
|
return (ai);
|
|
}
|
|
|
|
static int
|
|
cxgb_controller_probe(device_t dev)
|
|
{
|
|
const struct adapter_info *ai;
|
|
char *ports, buf[80];
|
|
int nports;
|
|
|
|
ai = cxgb_get_adapter_info(dev);
|
|
if (ai == NULL)
|
|
return (ENXIO);
|
|
|
|
nports = ai->nports0 + ai->nports1;
|
|
if (nports == 1)
|
|
ports = "port";
|
|
else
|
|
ports = "ports";
|
|
|
|
snprintf(buf, sizeof(buf), "%s, %d %s", ai->desc, nports, ports);
|
|
device_set_desc_copy(dev, buf);
|
|
return (BUS_PROBE_DEFAULT);
|
|
}
|
|
|
|
#define FW_FNAME "cxgb_t3fw"
|
|
#define TPEEPROM_NAME "cxgb_t3%c_tp_eeprom"
|
|
#define TPSRAM_NAME "cxgb_t3%c_protocol_sram"
|
|
|
|
static int
|
|
upgrade_fw(adapter_t *sc)
|
|
{
|
|
const struct firmware *fw;
|
|
int status;
|
|
u32 vers;
|
|
|
|
if ((fw = firmware_get(FW_FNAME)) == NULL) {
|
|
device_printf(sc->dev, "Could not find firmware image %s\n", FW_FNAME);
|
|
return (ENOENT);
|
|
} else
|
|
device_printf(sc->dev, "installing firmware on card\n");
|
|
status = t3_load_fw(sc, (const uint8_t *)fw->data, fw->datasize);
|
|
|
|
if (status != 0) {
|
|
device_printf(sc->dev, "failed to install firmware: %d\n",
|
|
status);
|
|
} else {
|
|
t3_get_fw_version(sc, &vers);
|
|
snprintf(&sc->fw_version[0], sizeof(sc->fw_version), "%d.%d.%d",
|
|
G_FW_VERSION_MAJOR(vers), G_FW_VERSION_MINOR(vers),
|
|
G_FW_VERSION_MICRO(vers));
|
|
}
|
|
|
|
firmware_put(fw, FIRMWARE_UNLOAD);
|
|
|
|
return (status);
|
|
}
|
|
|
|
/*
|
|
* The cxgb_controller_attach function is responsible for the initial
|
|
* bringup of the device. Its responsibilities include:
|
|
*
|
|
* 1. Determine if the device supports MSI or MSI-X.
|
|
* 2. Allocate bus resources so that we can access the Base Address Register
|
|
* 3. Create and initialize mutexes for the controller and its control
|
|
* logic such as SGE and MDIO.
|
|
* 4. Call hardware specific setup routine for the adapter as a whole.
|
|
* 5. Allocate the BAR for doing MSI-X.
|
|
* 6. Setup the line interrupt iff MSI-X is not supported.
|
|
* 7. Create the driver's taskq.
|
|
* 8. Start one task queue service thread.
|
|
* 9. Check if the firmware and SRAM are up-to-date. They will be
|
|
* auto-updated later (before FULL_INIT_DONE), if required.
|
|
* 10. Create a child device for each MAC (port)
|
|
* 11. Initialize T3 private state.
|
|
* 12. Trigger the LED
|
|
* 13. Setup offload iff supported.
|
|
* 14. Reset/restart the tick callout.
|
|
* 15. Attach sysctls
|
|
*
|
|
* NOTE: Any modification or deviation from this list MUST be reflected in
|
|
* the above comment. Failure to do so will result in problems on various
|
|
* error conditions including link flapping.
|
|
*/
|
|
static int
|
|
cxgb_controller_attach(device_t dev)
|
|
{
|
|
device_t child;
|
|
const struct adapter_info *ai;
|
|
struct adapter *sc;
|
|
int i, error = 0;
|
|
uint32_t vers;
|
|
int port_qsets = 1;
|
|
int msi_needed, reg;
|
|
char buf[80];
|
|
|
|
sc = device_get_softc(dev);
|
|
sc->dev = dev;
|
|
sc->msi_count = 0;
|
|
ai = cxgb_get_adapter_info(dev);
|
|
|
|
snprintf(sc->lockbuf, ADAPTER_LOCK_NAME_LEN, "cxgb controller lock %d",
|
|
device_get_unit(dev));
|
|
ADAPTER_LOCK_INIT(sc, sc->lockbuf);
|
|
|
|
snprintf(sc->reglockbuf, ADAPTER_LOCK_NAME_LEN, "SGE reg lock %d",
|
|
device_get_unit(dev));
|
|
snprintf(sc->mdiolockbuf, ADAPTER_LOCK_NAME_LEN, "cxgb mdio lock %d",
|
|
device_get_unit(dev));
|
|
snprintf(sc->elmerlockbuf, ADAPTER_LOCK_NAME_LEN, "cxgb elmer lock %d",
|
|
device_get_unit(dev));
|
|
|
|
MTX_INIT(&sc->sge.reg_lock, sc->reglockbuf, NULL, MTX_SPIN);
|
|
MTX_INIT(&sc->mdio_lock, sc->mdiolockbuf, NULL, MTX_DEF);
|
|
MTX_INIT(&sc->elmer_lock, sc->elmerlockbuf, NULL, MTX_DEF);
|
|
|
|
mtx_lock(&t3_list_lock);
|
|
SLIST_INSERT_HEAD(&t3_list, sc, link);
|
|
mtx_unlock(&t3_list_lock);
|
|
|
|
/* find the PCIe link width and set max read request to 4KB*/
|
|
if (pci_find_cap(dev, PCIY_EXPRESS, ®) == 0) {
|
|
uint16_t lnk;
|
|
|
|
lnk = pci_read_config(dev, reg + PCIER_LINK_STA, 2);
|
|
sc->link_width = (lnk & PCIEM_LINK_STA_WIDTH) >> 4;
|
|
if (sc->link_width < 8 &&
|
|
(ai->caps & SUPPORTED_10000baseT_Full)) {
|
|
device_printf(sc->dev,
|
|
"PCIe x%d Link, expect reduced performance\n",
|
|
sc->link_width);
|
|
}
|
|
|
|
pci_set_max_read_req(dev, 4096);
|
|
}
|
|
|
|
touch_bars(dev);
|
|
pci_enable_busmaster(dev);
|
|
/*
|
|
* Allocate the registers and make them available to the driver.
|
|
* The registers that we care about for NIC mode are in BAR 0
|
|
*/
|
|
sc->regs_rid = PCIR_BAR(0);
|
|
if ((sc->regs_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY,
|
|
&sc->regs_rid, RF_ACTIVE)) == NULL) {
|
|
device_printf(dev, "Cannot allocate BAR region 0\n");
|
|
error = ENXIO;
|
|
goto out;
|
|
}
|
|
|
|
sc->bt = rman_get_bustag(sc->regs_res);
|
|
sc->bh = rman_get_bushandle(sc->regs_res);
|
|
sc->mmio_len = rman_get_size(sc->regs_res);
|
|
|
|
for (i = 0; i < MAX_NPORTS; i++)
|
|
sc->port[i].adapter = sc;
|
|
|
|
if (t3_prep_adapter(sc, ai, 1) < 0) {
|
|
printf("prep adapter failed\n");
|
|
error = ENODEV;
|
|
goto out;
|
|
}
|
|
|
|
sc->udbs_rid = PCIR_BAR(2);
|
|
sc->udbs_res = NULL;
|
|
if (is_offload(sc) &&
|
|
((sc->udbs_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY,
|
|
&sc->udbs_rid, RF_ACTIVE)) == NULL)) {
|
|
device_printf(dev, "Cannot allocate BAR region 1\n");
|
|
error = ENXIO;
|
|
goto out;
|
|
}
|
|
|
|
/* Allocate the BAR for doing MSI-X. If it succeeds, try to allocate
|
|
* enough messages for the queue sets. If that fails, try falling
|
|
* back to MSI. If that fails, then try falling back to the legacy
|
|
* interrupt pin model.
|
|
*/
|
|
sc->msix_regs_rid = 0x20;
|
|
if ((msi_allowed >= 2) &&
|
|
(sc->msix_regs_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY,
|
|
&sc->msix_regs_rid, RF_ACTIVE)) != NULL) {
|
|
|
|
if (multiq)
|
|
port_qsets = min(SGE_QSETS/sc->params.nports, mp_ncpus);
|
|
msi_needed = sc->msi_count = sc->params.nports * port_qsets + 1;
|
|
|
|
if (pci_msix_count(dev) == 0 ||
|
|
(error = pci_alloc_msix(dev, &sc->msi_count)) != 0 ||
|
|
sc->msi_count != msi_needed) {
|
|
device_printf(dev, "alloc msix failed - "
|
|
"msi_count=%d, msi_needed=%d, err=%d; "
|
|
"will try MSI\n", sc->msi_count,
|
|
msi_needed, error);
|
|
sc->msi_count = 0;
|
|
port_qsets = 1;
|
|
pci_release_msi(dev);
|
|
bus_release_resource(dev, SYS_RES_MEMORY,
|
|
sc->msix_regs_rid, sc->msix_regs_res);
|
|
sc->msix_regs_res = NULL;
|
|
} else {
|
|
sc->flags |= USING_MSIX;
|
|
sc->cxgb_intr = cxgb_async_intr;
|
|
device_printf(dev,
|
|
"using MSI-X interrupts (%u vectors)\n",
|
|
sc->msi_count);
|
|
}
|
|
}
|
|
|
|
if ((msi_allowed >= 1) && (sc->msi_count == 0)) {
|
|
sc->msi_count = 1;
|
|
if ((error = pci_alloc_msi(dev, &sc->msi_count)) != 0) {
|
|
device_printf(dev, "alloc msi failed - "
|
|
"err=%d; will try INTx\n", error);
|
|
sc->msi_count = 0;
|
|
port_qsets = 1;
|
|
pci_release_msi(dev);
|
|
} else {
|
|
sc->flags |= USING_MSI;
|
|
sc->cxgb_intr = t3_intr_msi;
|
|
device_printf(dev, "using MSI interrupts\n");
|
|
}
|
|
}
|
|
if (sc->msi_count == 0) {
|
|
device_printf(dev, "using line interrupts\n");
|
|
sc->cxgb_intr = t3b_intr;
|
|
}
|
|
|
|
/* Create a private taskqueue thread for handling driver events */
|
|
sc->tq = taskqueue_create("cxgb_taskq", M_NOWAIT,
|
|
taskqueue_thread_enqueue, &sc->tq);
|
|
if (sc->tq == NULL) {
|
|
device_printf(dev, "failed to allocate controller task queue\n");
|
|
goto out;
|
|
}
|
|
|
|
taskqueue_start_threads(&sc->tq, 1, PI_NET, "%s taskq",
|
|
device_get_nameunit(dev));
|
|
TASK_INIT(&sc->tick_task, 0, cxgb_tick_handler, sc);
|
|
|
|
|
|
/* Create a periodic callout for checking adapter status */
|
|
callout_init(&sc->cxgb_tick_ch, 1);
|
|
|
|
if (t3_check_fw_version(sc) < 0 || force_fw_update) {
|
|
/*
|
|
* Warn user that a firmware update will be attempted in init.
|
|
*/
|
|
device_printf(dev, "firmware needs to be updated to version %d.%d.%d\n",
|
|
FW_VERSION_MAJOR, FW_VERSION_MINOR, FW_VERSION_MICRO);
|
|
sc->flags &= ~FW_UPTODATE;
|
|
} else {
|
|
sc->flags |= FW_UPTODATE;
|
|
}
|
|
|
|
if (t3_check_tpsram_version(sc) < 0) {
|
|
/*
|
|
* Warn user that a firmware update will be attempted in init.
|
|
*/
|
|
device_printf(dev, "SRAM needs to be updated to version %c-%d.%d.%d\n",
|
|
t3rev2char(sc), TP_VERSION_MAJOR, TP_VERSION_MINOR, TP_VERSION_MICRO);
|
|
sc->flags &= ~TPS_UPTODATE;
|
|
} else {
|
|
sc->flags |= TPS_UPTODATE;
|
|
}
|
|
|
|
/*
|
|
* Create a child device for each MAC. The ethernet attachment
|
|
* will be done in these children.
|
|
*/
|
|
for (i = 0; i < (sc)->params.nports; i++) {
|
|
struct port_info *pi;
|
|
|
|
if ((child = device_add_child(dev, "cxgb", -1)) == NULL) {
|
|
device_printf(dev, "failed to add child port\n");
|
|
error = EINVAL;
|
|
goto out;
|
|
}
|
|
pi = &sc->port[i];
|
|
pi->adapter = sc;
|
|
pi->nqsets = port_qsets;
|
|
pi->first_qset = i*port_qsets;
|
|
pi->port_id = i;
|
|
pi->tx_chan = i >= ai->nports0;
|
|
pi->txpkt_intf = pi->tx_chan ? 2 * (i - ai->nports0) + 1 : 2 * i;
|
|
sc->rxpkt_map[pi->txpkt_intf] = i;
|
|
sc->port[i].tx_chan = i >= ai->nports0;
|
|
sc->portdev[i] = child;
|
|
device_set_softc(child, pi);
|
|
}
|
|
if ((error = bus_generic_attach(dev)) != 0)
|
|
goto out;
|
|
|
|
/* initialize sge private state */
|
|
t3_sge_init_adapter(sc);
|
|
|
|
t3_led_ready(sc);
|
|
|
|
error = t3_get_fw_version(sc, &vers);
|
|
if (error)
|
|
goto out;
|
|
|
|
snprintf(&sc->fw_version[0], sizeof(sc->fw_version), "%d.%d.%d",
|
|
G_FW_VERSION_MAJOR(vers), G_FW_VERSION_MINOR(vers),
|
|
G_FW_VERSION_MICRO(vers));
|
|
|
|
snprintf(buf, sizeof(buf), "%s %sNIC\t E/C: %s S/N: %s",
|
|
ai->desc, is_offload(sc) ? "R" : "",
|
|
sc->params.vpd.ec, sc->params.vpd.sn);
|
|
device_set_desc_copy(dev, buf);
|
|
|
|
snprintf(&sc->port_types[0], sizeof(sc->port_types), "%x%x%x%x",
|
|
sc->params.vpd.port_type[0], sc->params.vpd.port_type[1],
|
|
sc->params.vpd.port_type[2], sc->params.vpd.port_type[3]);
|
|
|
|
device_printf(sc->dev, "Firmware Version %s\n", &sc->fw_version[0]);
|
|
callout_reset(&sc->cxgb_tick_ch, hz, cxgb_tick, sc);
|
|
t3_add_attach_sysctls(sc);
|
|
|
|
#ifdef TCP_OFFLOAD
|
|
for (i = 0; i < NUM_CPL_HANDLERS; i++)
|
|
sc->cpl_handler[i] = cpl_not_handled;
|
|
#endif
|
|
|
|
t3_intr_clear(sc);
|
|
error = cxgb_setup_interrupts(sc);
|
|
out:
|
|
if (error)
|
|
cxgb_free(sc);
|
|
|
|
return (error);
|
|
}
|
|
|
|
/*
|
|
* The cxgb_controller_detach routine is called with the device is
|
|
* unloaded from the system.
|
|
*/
|
|
|
|
static int
|
|
cxgb_controller_detach(device_t dev)
|
|
{
|
|
struct adapter *sc;
|
|
|
|
sc = device_get_softc(dev);
|
|
|
|
cxgb_free(sc);
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* The cxgb_free() is called by the cxgb_controller_detach() routine
|
|
* to tear down the structures that were built up in
|
|
* cxgb_controller_attach(), and should be the final piece of work
|
|
* done when fully unloading the driver.
|
|
*
|
|
*
|
|
* 1. Shutting down the threads started by the cxgb_controller_attach()
|
|
* routine.
|
|
* 2. Stopping the lower level device and all callouts (cxgb_down_locked()).
|
|
* 3. Detaching all of the port devices created during the
|
|
* cxgb_controller_attach() routine.
|
|
* 4. Removing the device children created via cxgb_controller_attach().
|
|
* 5. Releasing PCI resources associated with the device.
|
|
* 6. Turning off the offload support, iff it was turned on.
|
|
* 7. Destroying the mutexes created in cxgb_controller_attach().
|
|
*
|
|
*/
|
|
static void
|
|
cxgb_free(struct adapter *sc)
|
|
{
|
|
int i, nqsets = 0;
|
|
|
|
ADAPTER_LOCK(sc);
|
|
sc->flags |= CXGB_SHUTDOWN;
|
|
ADAPTER_UNLOCK(sc);
|
|
|
|
/*
|
|
* Make sure all child devices are gone.
|
|
*/
|
|
bus_generic_detach(sc->dev);
|
|
for (i = 0; i < (sc)->params.nports; i++) {
|
|
if (sc->portdev[i] &&
|
|
device_delete_child(sc->dev, sc->portdev[i]) != 0)
|
|
device_printf(sc->dev, "failed to delete child port\n");
|
|
nqsets += sc->port[i].nqsets;
|
|
}
|
|
|
|
/*
|
|
* At this point, it is as if cxgb_port_detach has run on all ports, and
|
|
* cxgb_down has run on the adapter. All interrupts have been silenced,
|
|
* all open devices have been closed.
|
|
*/
|
|
KASSERT(sc->open_device_map == 0, ("%s: device(s) still open (%x)",
|
|
__func__, sc->open_device_map));
|
|
for (i = 0; i < sc->params.nports; i++) {
|
|
KASSERT(sc->port[i].ifp == NULL, ("%s: port %i undead!",
|
|
__func__, i));
|
|
}
|
|
|
|
/*
|
|
* Finish off the adapter's callouts.
|
|
*/
|
|
callout_drain(&sc->cxgb_tick_ch);
|
|
callout_drain(&sc->sge_timer_ch);
|
|
|
|
/*
|
|
* Release resources grabbed under FULL_INIT_DONE by cxgb_up. The
|
|
* sysctls are cleaned up by the kernel linker.
|
|
*/
|
|
if (sc->flags & FULL_INIT_DONE) {
|
|
t3_free_sge_resources(sc, nqsets);
|
|
sc->flags &= ~FULL_INIT_DONE;
|
|
}
|
|
|
|
/*
|
|
* Release all interrupt resources.
|
|
*/
|
|
cxgb_teardown_interrupts(sc);
|
|
if (sc->flags & (USING_MSI | USING_MSIX)) {
|
|
device_printf(sc->dev, "releasing msi message(s)\n");
|
|
pci_release_msi(sc->dev);
|
|
} else {
|
|
device_printf(sc->dev, "no msi message to release\n");
|
|
}
|
|
|
|
if (sc->msix_regs_res != NULL) {
|
|
bus_release_resource(sc->dev, SYS_RES_MEMORY, sc->msix_regs_rid,
|
|
sc->msix_regs_res);
|
|
}
|
|
|
|
/*
|
|
* Free the adapter's taskqueue.
|
|
*/
|
|
if (sc->tq != NULL) {
|
|
taskqueue_free(sc->tq);
|
|
sc->tq = NULL;
|
|
}
|
|
|
|
free(sc->filters, M_DEVBUF);
|
|
t3_sge_free(sc);
|
|
|
|
if (sc->udbs_res != NULL)
|
|
bus_release_resource(sc->dev, SYS_RES_MEMORY, sc->udbs_rid,
|
|
sc->udbs_res);
|
|
|
|
if (sc->regs_res != NULL)
|
|
bus_release_resource(sc->dev, SYS_RES_MEMORY, sc->regs_rid,
|
|
sc->regs_res);
|
|
|
|
MTX_DESTROY(&sc->mdio_lock);
|
|
MTX_DESTROY(&sc->sge.reg_lock);
|
|
MTX_DESTROY(&sc->elmer_lock);
|
|
mtx_lock(&t3_list_lock);
|
|
SLIST_REMOVE(&t3_list, sc, adapter, link);
|
|
mtx_unlock(&t3_list_lock);
|
|
ADAPTER_LOCK_DEINIT(sc);
|
|
}
|
|
|
|
/**
|
|
* setup_sge_qsets - configure SGE Tx/Rx/response queues
|
|
* @sc: the controller softc
|
|
*
|
|
* Determines how many sets of SGE queues to use and initializes them.
|
|
* We support multiple queue sets per port if we have MSI-X, otherwise
|
|
* just one queue set per port.
|
|
*/
|
|
static int
|
|
setup_sge_qsets(adapter_t *sc)
|
|
{
|
|
int i, j, err, irq_idx = 0, qset_idx = 0;
|
|
u_int ntxq = SGE_TXQ_PER_SET;
|
|
|
|
if ((err = t3_sge_alloc(sc)) != 0) {
|
|
device_printf(sc->dev, "t3_sge_alloc returned %d\n", err);
|
|
return (err);
|
|
}
|
|
|
|
if (sc->params.rev > 0 && !(sc->flags & USING_MSI))
|
|
irq_idx = -1;
|
|
|
|
for (i = 0; i < (sc)->params.nports; i++) {
|
|
struct port_info *pi = &sc->port[i];
|
|
|
|
for (j = 0; j < pi->nqsets; j++, qset_idx++) {
|
|
err = t3_sge_alloc_qset(sc, qset_idx, (sc)->params.nports,
|
|
(sc->flags & USING_MSIX) ? qset_idx + 1 : irq_idx,
|
|
&sc->params.sge.qset[qset_idx], ntxq, pi);
|
|
if (err) {
|
|
t3_free_sge_resources(sc, qset_idx);
|
|
device_printf(sc->dev,
|
|
"t3_sge_alloc_qset failed with %d\n", err);
|
|
return (err);
|
|
}
|
|
}
|
|
}
|
|
|
|
return (0);
|
|
}
|
|
|
|
static void
|
|
cxgb_teardown_interrupts(adapter_t *sc)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < SGE_QSETS; i++) {
|
|
if (sc->msix_intr_tag[i] == NULL) {
|
|
|
|
/* Should have been setup fully or not at all */
|
|
KASSERT(sc->msix_irq_res[i] == NULL &&
|
|
sc->msix_irq_rid[i] == 0,
|
|
("%s: half-done interrupt (%d).", __func__, i));
|
|
|
|
continue;
|
|
}
|
|
|
|
bus_teardown_intr(sc->dev, sc->msix_irq_res[i],
|
|
sc->msix_intr_tag[i]);
|
|
bus_release_resource(sc->dev, SYS_RES_IRQ, sc->msix_irq_rid[i],
|
|
sc->msix_irq_res[i]);
|
|
|
|
sc->msix_irq_res[i] = sc->msix_intr_tag[i] = NULL;
|
|
sc->msix_irq_rid[i] = 0;
|
|
}
|
|
|
|
if (sc->intr_tag) {
|
|
KASSERT(sc->irq_res != NULL,
|
|
("%s: half-done interrupt.", __func__));
|
|
|
|
bus_teardown_intr(sc->dev, sc->irq_res, sc->intr_tag);
|
|
bus_release_resource(sc->dev, SYS_RES_IRQ, sc->irq_rid,
|
|
sc->irq_res);
|
|
|
|
sc->irq_res = sc->intr_tag = NULL;
|
|
sc->irq_rid = 0;
|
|
}
|
|
}
|
|
|
|
static int
|
|
cxgb_setup_interrupts(adapter_t *sc)
|
|
{
|
|
struct resource *res;
|
|
void *tag;
|
|
int i, rid, err, intr_flag = sc->flags & (USING_MSI | USING_MSIX);
|
|
|
|
sc->irq_rid = intr_flag ? 1 : 0;
|
|
sc->irq_res = bus_alloc_resource_any(sc->dev, SYS_RES_IRQ, &sc->irq_rid,
|
|
RF_SHAREABLE | RF_ACTIVE);
|
|
if (sc->irq_res == NULL) {
|
|
device_printf(sc->dev, "Cannot allocate interrupt (%x, %u)\n",
|
|
intr_flag, sc->irq_rid);
|
|
err = EINVAL;
|
|
sc->irq_rid = 0;
|
|
} else {
|
|
err = bus_setup_intr(sc->dev, sc->irq_res,
|
|
INTR_MPSAFE | INTR_TYPE_NET, NULL,
|
|
sc->cxgb_intr, sc, &sc->intr_tag);
|
|
|
|
if (err) {
|
|
device_printf(sc->dev,
|
|
"Cannot set up interrupt (%x, %u, %d)\n",
|
|
intr_flag, sc->irq_rid, err);
|
|
bus_release_resource(sc->dev, SYS_RES_IRQ, sc->irq_rid,
|
|
sc->irq_res);
|
|
sc->irq_res = sc->intr_tag = NULL;
|
|
sc->irq_rid = 0;
|
|
}
|
|
}
|
|
|
|
/* That's all for INTx or MSI */
|
|
if (!(intr_flag & USING_MSIX) || err)
|
|
return (err);
|
|
|
|
bus_describe_intr(sc->dev, sc->irq_res, sc->intr_tag, "err");
|
|
for (i = 0; i < sc->msi_count - 1; i++) {
|
|
rid = i + 2;
|
|
res = bus_alloc_resource_any(sc->dev, SYS_RES_IRQ, &rid,
|
|
RF_SHAREABLE | RF_ACTIVE);
|
|
if (res == NULL) {
|
|
device_printf(sc->dev, "Cannot allocate interrupt "
|
|
"for message %d\n", rid);
|
|
err = EINVAL;
|
|
break;
|
|
}
|
|
|
|
err = bus_setup_intr(sc->dev, res, INTR_MPSAFE | INTR_TYPE_NET,
|
|
NULL, t3_intr_msix, &sc->sge.qs[i], &tag);
|
|
if (err) {
|
|
device_printf(sc->dev, "Cannot set up interrupt "
|
|
"for message %d (%d)\n", rid, err);
|
|
bus_release_resource(sc->dev, SYS_RES_IRQ, rid, res);
|
|
break;
|
|
}
|
|
|
|
sc->msix_irq_rid[i] = rid;
|
|
sc->msix_irq_res[i] = res;
|
|
sc->msix_intr_tag[i] = tag;
|
|
bus_describe_intr(sc->dev, res, tag, "qs%d", i);
|
|
}
|
|
|
|
if (err)
|
|
cxgb_teardown_interrupts(sc);
|
|
|
|
return (err);
|
|
}
|
|
|
|
|
|
static int
|
|
cxgb_port_probe(device_t dev)
|
|
{
|
|
struct port_info *p;
|
|
char buf[80];
|
|
const char *desc;
|
|
|
|
p = device_get_softc(dev);
|
|
desc = p->phy.desc;
|
|
snprintf(buf, sizeof(buf), "Port %d %s", p->port_id, desc);
|
|
device_set_desc_copy(dev, buf);
|
|
return (0);
|
|
}
|
|
|
|
|
|
static int
|
|
cxgb_makedev(struct port_info *pi)
|
|
{
|
|
|
|
pi->port_cdev = make_dev(&cxgb_cdevsw, pi->ifp->if_dunit,
|
|
UID_ROOT, GID_WHEEL, 0600, "%s", if_name(pi->ifp));
|
|
|
|
if (pi->port_cdev == NULL)
|
|
return (ENOMEM);
|
|
|
|
pi->port_cdev->si_drv1 = (void *)pi;
|
|
|
|
return (0);
|
|
}
|
|
|
|
#define CXGB_CAP (IFCAP_VLAN_HWTAGGING | IFCAP_VLAN_MTU | IFCAP_HWCSUM | \
|
|
IFCAP_VLAN_HWCSUM | IFCAP_TSO | IFCAP_JUMBO_MTU | IFCAP_LRO | \
|
|
IFCAP_VLAN_HWTSO | IFCAP_LINKSTATE | IFCAP_HWCSUM_IPV6)
|
|
#define CXGB_CAP_ENABLE CXGB_CAP
|
|
|
|
static int
|
|
cxgb_port_attach(device_t dev)
|
|
{
|
|
struct port_info *p;
|
|
struct ifnet *ifp;
|
|
int err;
|
|
struct adapter *sc;
|
|
|
|
p = device_get_softc(dev);
|
|
sc = p->adapter;
|
|
snprintf(p->lockbuf, PORT_NAME_LEN, "cxgb port lock %d:%d",
|
|
device_get_unit(device_get_parent(dev)), p->port_id);
|
|
PORT_LOCK_INIT(p, p->lockbuf);
|
|
|
|
callout_init(&p->link_check_ch, 1);
|
|
TASK_INIT(&p->link_check_task, 0, check_link_status, p);
|
|
|
|
/* Allocate an ifnet object and set it up */
|
|
ifp = p->ifp = if_alloc(IFT_ETHER);
|
|
if (ifp == NULL) {
|
|
device_printf(dev, "Cannot allocate ifnet\n");
|
|
return (ENOMEM);
|
|
}
|
|
|
|
if_initname(ifp, device_get_name(dev), device_get_unit(dev));
|
|
ifp->if_init = cxgb_init;
|
|
ifp->if_softc = p;
|
|
ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
|
|
ifp->if_ioctl = cxgb_ioctl;
|
|
ifp->if_transmit = cxgb_transmit;
|
|
ifp->if_qflush = cxgb_qflush;
|
|
ifp->if_get_counter = cxgb_get_counter;
|
|
|
|
ifp->if_capabilities = CXGB_CAP;
|
|
#ifdef TCP_OFFLOAD
|
|
if (is_offload(sc))
|
|
ifp->if_capabilities |= IFCAP_TOE4;
|
|
#endif
|
|
ifp->if_capenable = CXGB_CAP_ENABLE;
|
|
ifp->if_hwassist = CSUM_TCP | CSUM_UDP | CSUM_IP | CSUM_TSO |
|
|
CSUM_UDP_IPV6 | CSUM_TCP_IPV6;
|
|
|
|
/*
|
|
* Disable TSO on 4-port - it isn't supported by the firmware.
|
|
*/
|
|
if (sc->params.nports > 2) {
|
|
ifp->if_capabilities &= ~(IFCAP_TSO | IFCAP_VLAN_HWTSO);
|
|
ifp->if_capenable &= ~(IFCAP_TSO | IFCAP_VLAN_HWTSO);
|
|
ifp->if_hwassist &= ~CSUM_TSO;
|
|
}
|
|
|
|
ether_ifattach(ifp, p->hw_addr);
|
|
|
|
#ifdef DEFAULT_JUMBO
|
|
if (sc->params.nports <= 2)
|
|
ifp->if_mtu = ETHERMTU_JUMBO;
|
|
#endif
|
|
if ((err = cxgb_makedev(p)) != 0) {
|
|
printf("makedev failed %d\n", err);
|
|
return (err);
|
|
}
|
|
|
|
/* Create a list of media supported by this port */
|
|
ifmedia_init(&p->media, IFM_IMASK, cxgb_media_change,
|
|
cxgb_media_status);
|
|
cxgb_build_medialist(p);
|
|
|
|
t3_sge_init_port(p);
|
|
|
|
return (err);
|
|
}
|
|
|
|
/*
|
|
* cxgb_port_detach() is called via the device_detach methods when
|
|
* cxgb_free() calls the bus_generic_detach. It is responsible for
|
|
* removing the device from the view of the kernel, i.e. from all
|
|
* interfaces lists etc. This routine is only called when the driver is
|
|
* being unloaded, not when the link goes down.
|
|
*/
|
|
static int
|
|
cxgb_port_detach(device_t dev)
|
|
{
|
|
struct port_info *p;
|
|
struct adapter *sc;
|
|
int i;
|
|
|
|
p = device_get_softc(dev);
|
|
sc = p->adapter;
|
|
|
|
/* Tell cxgb_ioctl and if_init that the port is going away */
|
|
ADAPTER_LOCK(sc);
|
|
SET_DOOMED(p);
|
|
wakeup(&sc->flags);
|
|
while (IS_BUSY(sc))
|
|
mtx_sleep(&sc->flags, &sc->lock, 0, "cxgbdtch", 0);
|
|
SET_BUSY(sc);
|
|
ADAPTER_UNLOCK(sc);
|
|
|
|
if (p->port_cdev != NULL)
|
|
destroy_dev(p->port_cdev);
|
|
|
|
cxgb_uninit_synchronized(p);
|
|
ether_ifdetach(p->ifp);
|
|
|
|
for (i = p->first_qset; i < p->first_qset + p->nqsets; i++) {
|
|
struct sge_qset *qs = &sc->sge.qs[i];
|
|
struct sge_txq *txq = &qs->txq[TXQ_ETH];
|
|
|
|
callout_drain(&txq->txq_watchdog);
|
|
callout_drain(&txq->txq_timer);
|
|
}
|
|
|
|
PORT_LOCK_DEINIT(p);
|
|
if_free(p->ifp);
|
|
p->ifp = NULL;
|
|
|
|
ADAPTER_LOCK(sc);
|
|
CLR_BUSY(sc);
|
|
wakeup_one(&sc->flags);
|
|
ADAPTER_UNLOCK(sc);
|
|
return (0);
|
|
}
|
|
|
|
void
|
|
t3_fatal_err(struct adapter *sc)
|
|
{
|
|
u_int fw_status[4];
|
|
|
|
if (sc->flags & FULL_INIT_DONE) {
|
|
t3_sge_stop(sc);
|
|
t3_write_reg(sc, A_XGM_TX_CTRL, 0);
|
|
t3_write_reg(sc, A_XGM_RX_CTRL, 0);
|
|
t3_write_reg(sc, XGM_REG(A_XGM_TX_CTRL, 1), 0);
|
|
t3_write_reg(sc, XGM_REG(A_XGM_RX_CTRL, 1), 0);
|
|
t3_intr_disable(sc);
|
|
}
|
|
device_printf(sc->dev,"encountered fatal error, operation suspended\n");
|
|
if (!t3_cim_ctl_blk_read(sc, 0xa0, 4, fw_status))
|
|
device_printf(sc->dev, "FW_ status: 0x%x, 0x%x, 0x%x, 0x%x\n",
|
|
fw_status[0], fw_status[1], fw_status[2], fw_status[3]);
|
|
}
|
|
|
|
int
|
|
t3_os_find_pci_capability(adapter_t *sc, int cap)
|
|
{
|
|
device_t dev;
|
|
struct pci_devinfo *dinfo;
|
|
pcicfgregs *cfg;
|
|
uint32_t status;
|
|
uint8_t ptr;
|
|
|
|
dev = sc->dev;
|
|
dinfo = device_get_ivars(dev);
|
|
cfg = &dinfo->cfg;
|
|
|
|
status = pci_read_config(dev, PCIR_STATUS, 2);
|
|
if (!(status & PCIM_STATUS_CAPPRESENT))
|
|
return (0);
|
|
|
|
switch (cfg->hdrtype & PCIM_HDRTYPE) {
|
|
case 0:
|
|
case 1:
|
|
ptr = PCIR_CAP_PTR;
|
|
break;
|
|
case 2:
|
|
ptr = PCIR_CAP_PTR_2;
|
|
break;
|
|
default:
|
|
return (0);
|
|
break;
|
|
}
|
|
ptr = pci_read_config(dev, ptr, 1);
|
|
|
|
while (ptr != 0) {
|
|
if (pci_read_config(dev, ptr + PCICAP_ID, 1) == cap)
|
|
return (ptr);
|
|
ptr = pci_read_config(dev, ptr + PCICAP_NEXTPTR, 1);
|
|
}
|
|
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
t3_os_pci_save_state(struct adapter *sc)
|
|
{
|
|
device_t dev;
|
|
struct pci_devinfo *dinfo;
|
|
|
|
dev = sc->dev;
|
|
dinfo = device_get_ivars(dev);
|
|
|
|
pci_cfg_save(dev, dinfo, 0);
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
t3_os_pci_restore_state(struct adapter *sc)
|
|
{
|
|
device_t dev;
|
|
struct pci_devinfo *dinfo;
|
|
|
|
dev = sc->dev;
|
|
dinfo = device_get_ivars(dev);
|
|
|
|
pci_cfg_restore(dev, dinfo);
|
|
return (0);
|
|
}
|
|
|
|
/**
|
|
* t3_os_link_changed - handle link status changes
|
|
* @sc: the adapter associated with the link change
|
|
* @port_id: the port index whose link status has changed
|
|
* @link_status: the new status of the link
|
|
* @speed: the new speed setting
|
|
* @duplex: the new duplex setting
|
|
* @fc: the new flow-control setting
|
|
*
|
|
* This is the OS-dependent handler for link status changes. The OS
|
|
* neutral handler takes care of most of the processing for these events,
|
|
* then calls this handler for any OS-specific processing.
|
|
*/
|
|
void
|
|
t3_os_link_changed(adapter_t *adapter, int port_id, int link_status, int speed,
|
|
int duplex, int fc, int mac_was_reset)
|
|
{
|
|
struct port_info *pi = &adapter->port[port_id];
|
|
struct ifnet *ifp = pi->ifp;
|
|
|
|
/* no race with detach, so ifp should always be good */
|
|
KASSERT(ifp, ("%s: if detached.", __func__));
|
|
|
|
/* Reapply mac settings if they were lost due to a reset */
|
|
if (mac_was_reset) {
|
|
PORT_LOCK(pi);
|
|
cxgb_update_mac_settings(pi);
|
|
PORT_UNLOCK(pi);
|
|
}
|
|
|
|
if (link_status) {
|
|
ifp->if_baudrate = IF_Mbps(speed);
|
|
if_link_state_change(ifp, LINK_STATE_UP);
|
|
} else
|
|
if_link_state_change(ifp, LINK_STATE_DOWN);
|
|
}
|
|
|
|
/**
|
|
* t3_os_phymod_changed - handle PHY module changes
|
|
* @phy: the PHY reporting the module change
|
|
* @mod_type: new module type
|
|
*
|
|
* This is the OS-dependent handler for PHY module changes. It is
|
|
* invoked when a PHY module is removed or inserted for any OS-specific
|
|
* processing.
|
|
*/
|
|
void t3_os_phymod_changed(struct adapter *adap, int port_id)
|
|
{
|
|
static const char *mod_str[] = {
|
|
NULL, "SR", "LR", "LRM", "TWINAX", "TWINAX-L", "unknown"
|
|
};
|
|
struct port_info *pi = &adap->port[port_id];
|
|
int mod = pi->phy.modtype;
|
|
|
|
if (mod != pi->media.ifm_cur->ifm_data)
|
|
cxgb_build_medialist(pi);
|
|
|
|
if (mod == phy_modtype_none)
|
|
if_printf(pi->ifp, "PHY module unplugged\n");
|
|
else {
|
|
KASSERT(mod < ARRAY_SIZE(mod_str),
|
|
("invalid PHY module type %d", mod));
|
|
if_printf(pi->ifp, "%s PHY module inserted\n", mod_str[mod]);
|
|
}
|
|
}
|
|
|
|
void
|
|
t3_os_set_hw_addr(adapter_t *adapter, int port_idx, u8 hw_addr[])
|
|
{
|
|
|
|
/*
|
|
* The ifnet might not be allocated before this gets called,
|
|
* as this is called early on in attach by t3_prep_adapter
|
|
* save the address off in the port structure
|
|
*/
|
|
if (cxgb_debug)
|
|
printf("set_hw_addr on idx %d addr %6D\n", port_idx, hw_addr, ":");
|
|
bcopy(hw_addr, adapter->port[port_idx].hw_addr, ETHER_ADDR_LEN);
|
|
}
|
|
|
|
/*
|
|
* Programs the XGMAC based on the settings in the ifnet. These settings
|
|
* include MTU, MAC address, mcast addresses, etc.
|
|
*/
|
|
static void
|
|
cxgb_update_mac_settings(struct port_info *p)
|
|
{
|
|
struct ifnet *ifp = p->ifp;
|
|
struct t3_rx_mode rm;
|
|
struct cmac *mac = &p->mac;
|
|
int mtu, hwtagging;
|
|
|
|
PORT_LOCK_ASSERT_OWNED(p);
|
|
|
|
bcopy(IF_LLADDR(ifp), p->hw_addr, ETHER_ADDR_LEN);
|
|
|
|
mtu = ifp->if_mtu;
|
|
if (ifp->if_capenable & IFCAP_VLAN_MTU)
|
|
mtu += ETHER_VLAN_ENCAP_LEN;
|
|
|
|
hwtagging = (ifp->if_capenable & IFCAP_VLAN_HWTAGGING) != 0;
|
|
|
|
t3_mac_set_mtu(mac, mtu);
|
|
t3_set_vlan_accel(p->adapter, 1 << p->tx_chan, hwtagging);
|
|
t3_mac_set_address(mac, 0, p->hw_addr);
|
|
t3_init_rx_mode(&rm, p);
|
|
t3_mac_set_rx_mode(mac, &rm);
|
|
}
|
|
|
|
|
|
static int
|
|
await_mgmt_replies(struct adapter *adap, unsigned long init_cnt,
|
|
unsigned long n)
|
|
{
|
|
int attempts = 5;
|
|
|
|
while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) {
|
|
if (!--attempts)
|
|
return (ETIMEDOUT);
|
|
t3_os_sleep(10);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
init_tp_parity(struct adapter *adap)
|
|
{
|
|
int i;
|
|
struct mbuf *m;
|
|
struct cpl_set_tcb_field *greq;
|
|
unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts;
|
|
|
|
t3_tp_set_offload_mode(adap, 1);
|
|
|
|
for (i = 0; i < 16; i++) {
|
|
struct cpl_smt_write_req *req;
|
|
|
|
m = m_gethdr(M_WAITOK, MT_DATA);
|
|
req = mtod(m, struct cpl_smt_write_req *);
|
|
m->m_len = m->m_pkthdr.len = sizeof(*req);
|
|
memset(req, 0, sizeof(*req));
|
|
req->wr.wrh_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
|
|
OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_SMT_WRITE_REQ, i));
|
|
req->iff = i;
|
|
t3_mgmt_tx(adap, m);
|
|
}
|
|
|
|
for (i = 0; i < 2048; i++) {
|
|
struct cpl_l2t_write_req *req;
|
|
|
|
m = m_gethdr(M_WAITOK, MT_DATA);
|
|
req = mtod(m, struct cpl_l2t_write_req *);
|
|
m->m_len = m->m_pkthdr.len = sizeof(*req);
|
|
memset(req, 0, sizeof(*req));
|
|
req->wr.wrh_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
|
|
OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_L2T_WRITE_REQ, i));
|
|
req->params = htonl(V_L2T_W_IDX(i));
|
|
t3_mgmt_tx(adap, m);
|
|
}
|
|
|
|
for (i = 0; i < 2048; i++) {
|
|
struct cpl_rte_write_req *req;
|
|
|
|
m = m_gethdr(M_WAITOK, MT_DATA);
|
|
req = mtod(m, struct cpl_rte_write_req *);
|
|
m->m_len = m->m_pkthdr.len = sizeof(*req);
|
|
memset(req, 0, sizeof(*req));
|
|
req->wr.wrh_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
|
|
OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_RTE_WRITE_REQ, i));
|
|
req->l2t_idx = htonl(V_L2T_W_IDX(i));
|
|
t3_mgmt_tx(adap, m);
|
|
}
|
|
|
|
m = m_gethdr(M_WAITOK, MT_DATA);
|
|
greq = mtod(m, struct cpl_set_tcb_field *);
|
|
m->m_len = m->m_pkthdr.len = sizeof(*greq);
|
|
memset(greq, 0, sizeof(*greq));
|
|
greq->wr.wrh_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
|
|
OPCODE_TID(greq) = htonl(MK_OPCODE_TID(CPL_SET_TCB_FIELD, 0));
|
|
greq->mask = htobe64(1);
|
|
t3_mgmt_tx(adap, m);
|
|
|
|
i = await_mgmt_replies(adap, cnt, 16 + 2048 + 2048 + 1);
|
|
t3_tp_set_offload_mode(adap, 0);
|
|
return (i);
|
|
}
|
|
|
|
/**
|
|
* setup_rss - configure Receive Side Steering (per-queue connection demux)
|
|
* @adap: the adapter
|
|
*
|
|
* Sets up RSS to distribute packets to multiple receive queues. We
|
|
* configure the RSS CPU lookup table to distribute to the number of HW
|
|
* receive queues, and the response queue lookup table to narrow that
|
|
* down to the response queues actually configured for each port.
|
|
* We always configure the RSS mapping for two ports since the mapping
|
|
* table has plenty of entries.
|
|
*/
|
|
static void
|
|
setup_rss(adapter_t *adap)
|
|
{
|
|
int i;
|
|
u_int nq[2];
|
|
uint8_t cpus[SGE_QSETS + 1];
|
|
uint16_t rspq_map[RSS_TABLE_SIZE];
|
|
|
|
for (i = 0; i < SGE_QSETS; ++i)
|
|
cpus[i] = i;
|
|
cpus[SGE_QSETS] = 0xff;
|
|
|
|
nq[0] = nq[1] = 0;
|
|
for_each_port(adap, i) {
|
|
const struct port_info *pi = adap2pinfo(adap, i);
|
|
|
|
nq[pi->tx_chan] += pi->nqsets;
|
|
}
|
|
for (i = 0; i < RSS_TABLE_SIZE / 2; ++i) {
|
|
rspq_map[i] = nq[0] ? i % nq[0] : 0;
|
|
rspq_map[i + RSS_TABLE_SIZE / 2] = nq[1] ? i % nq[1] + nq[0] : 0;
|
|
}
|
|
|
|
/* Calculate the reverse RSS map table */
|
|
for (i = 0; i < SGE_QSETS; ++i)
|
|
adap->rrss_map[i] = 0xff;
|
|
for (i = 0; i < RSS_TABLE_SIZE; ++i)
|
|
if (adap->rrss_map[rspq_map[i]] == 0xff)
|
|
adap->rrss_map[rspq_map[i]] = i;
|
|
|
|
t3_config_rss(adap, F_RQFEEDBACKENABLE | F_TNLLKPEN | F_TNLMAPEN |
|
|
F_TNLPRTEN | F_TNL2TUPEN | F_TNL4TUPEN | F_OFDMAPEN |
|
|
F_RRCPLMAPEN | V_RRCPLCPUSIZE(6) | F_HASHTOEPLITZ,
|
|
cpus, rspq_map);
|
|
|
|
}
|
|
static void
|
|
send_pktsched_cmd(struct adapter *adap, int sched, int qidx, int lo,
|
|
int hi, int port)
|
|
{
|
|
struct mbuf *m;
|
|
struct mngt_pktsched_wr *req;
|
|
|
|
m = m_gethdr(M_NOWAIT, MT_DATA);
|
|
if (m) {
|
|
req = mtod(m, struct mngt_pktsched_wr *);
|
|
req->wr.wrh_hi = htonl(V_WR_OP(FW_WROPCODE_MNGT));
|
|
req->mngt_opcode = FW_MNGTOPCODE_PKTSCHED_SET;
|
|
req->sched = sched;
|
|
req->idx = qidx;
|
|
req->min = lo;
|
|
req->max = hi;
|
|
req->binding = port;
|
|
m->m_len = m->m_pkthdr.len = sizeof(*req);
|
|
t3_mgmt_tx(adap, m);
|
|
}
|
|
}
|
|
|
|
static void
|
|
bind_qsets(adapter_t *sc)
|
|
{
|
|
int i, j;
|
|
|
|
for (i = 0; i < (sc)->params.nports; ++i) {
|
|
const struct port_info *pi = adap2pinfo(sc, i);
|
|
|
|
for (j = 0; j < pi->nqsets; ++j) {
|
|
send_pktsched_cmd(sc, 1, pi->first_qset + j, -1,
|
|
-1, pi->tx_chan);
|
|
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
update_tpeeprom(struct adapter *adap)
|
|
{
|
|
const struct firmware *tpeeprom;
|
|
|
|
uint32_t version;
|
|
unsigned int major, minor;
|
|
int ret, len;
|
|
char rev, name[32];
|
|
|
|
t3_seeprom_read(adap, TP_SRAM_OFFSET, &version);
|
|
|
|
major = G_TP_VERSION_MAJOR(version);
|
|
minor = G_TP_VERSION_MINOR(version);
|
|
if (major == TP_VERSION_MAJOR && minor == TP_VERSION_MINOR)
|
|
return;
|
|
|
|
rev = t3rev2char(adap);
|
|
snprintf(name, sizeof(name), TPEEPROM_NAME, rev);
|
|
|
|
tpeeprom = firmware_get(name);
|
|
if (tpeeprom == NULL) {
|
|
device_printf(adap->dev,
|
|
"could not load TP EEPROM: unable to load %s\n",
|
|
name);
|
|
return;
|
|
}
|
|
|
|
len = tpeeprom->datasize - 4;
|
|
|
|
ret = t3_check_tpsram(adap, tpeeprom->data, tpeeprom->datasize);
|
|
if (ret)
|
|
goto release_tpeeprom;
|
|
|
|
if (len != TP_SRAM_LEN) {
|
|
device_printf(adap->dev,
|
|
"%s length is wrong len=%d expected=%d\n", name,
|
|
len, TP_SRAM_LEN);
|
|
return;
|
|
}
|
|
|
|
ret = set_eeprom(&adap->port[0], tpeeprom->data, tpeeprom->datasize,
|
|
TP_SRAM_OFFSET);
|
|
|
|
if (!ret) {
|
|
device_printf(adap->dev,
|
|
"Protocol SRAM image updated in EEPROM to %d.%d.%d\n",
|
|
TP_VERSION_MAJOR, TP_VERSION_MINOR, TP_VERSION_MICRO);
|
|
} else
|
|
device_printf(adap->dev,
|
|
"Protocol SRAM image update in EEPROM failed\n");
|
|
|
|
release_tpeeprom:
|
|
firmware_put(tpeeprom, FIRMWARE_UNLOAD);
|
|
|
|
return;
|
|
}
|
|
|
|
static int
|
|
update_tpsram(struct adapter *adap)
|
|
{
|
|
const struct firmware *tpsram;
|
|
int ret;
|
|
char rev, name[32];
|
|
|
|
rev = t3rev2char(adap);
|
|
snprintf(name, sizeof(name), TPSRAM_NAME, rev);
|
|
|
|
update_tpeeprom(adap);
|
|
|
|
tpsram = firmware_get(name);
|
|
if (tpsram == NULL){
|
|
device_printf(adap->dev, "could not load TP SRAM\n");
|
|
return (EINVAL);
|
|
} else
|
|
device_printf(adap->dev, "updating TP SRAM\n");
|
|
|
|
ret = t3_check_tpsram(adap, tpsram->data, tpsram->datasize);
|
|
if (ret)
|
|
goto release_tpsram;
|
|
|
|
ret = t3_set_proto_sram(adap, tpsram->data);
|
|
if (ret)
|
|
device_printf(adap->dev, "loading protocol SRAM failed\n");
|
|
|
|
release_tpsram:
|
|
firmware_put(tpsram, FIRMWARE_UNLOAD);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* cxgb_up - enable the adapter
|
|
* @adap: adapter being enabled
|
|
*
|
|
* Called when the first port is enabled, this function performs the
|
|
* actions necessary to make an adapter operational, such as completing
|
|
* the initialization of HW modules, and enabling interrupts.
|
|
*/
|
|
static int
|
|
cxgb_up(struct adapter *sc)
|
|
{
|
|
int err = 0;
|
|
unsigned int mxf = t3_mc5_size(&sc->mc5) - MC5_MIN_TIDS;
|
|
|
|
KASSERT(sc->open_device_map == 0, ("%s: device(s) already open (%x)",
|
|
__func__, sc->open_device_map));
|
|
|
|
if ((sc->flags & FULL_INIT_DONE) == 0) {
|
|
|
|
ADAPTER_LOCK_ASSERT_NOTOWNED(sc);
|
|
|
|
if ((sc->flags & FW_UPTODATE) == 0)
|
|
if ((err = upgrade_fw(sc)))
|
|
goto out;
|
|
|
|
if ((sc->flags & TPS_UPTODATE) == 0)
|
|
if ((err = update_tpsram(sc)))
|
|
goto out;
|
|
|
|
if (is_offload(sc) && nfilters != 0) {
|
|
sc->params.mc5.nservers = 0;
|
|
|
|
if (nfilters < 0)
|
|
sc->params.mc5.nfilters = mxf;
|
|
else
|
|
sc->params.mc5.nfilters = min(nfilters, mxf);
|
|
}
|
|
|
|
err = t3_init_hw(sc, 0);
|
|
if (err)
|
|
goto out;
|
|
|
|
t3_set_reg_field(sc, A_TP_PARA_REG5, 0, F_RXDDPOFFINIT);
|
|
t3_write_reg(sc, A_ULPRX_TDDP_PSZ, V_HPZ0(PAGE_SHIFT - 12));
|
|
|
|
err = setup_sge_qsets(sc);
|
|
if (err)
|
|
goto out;
|
|
|
|
alloc_filters(sc);
|
|
setup_rss(sc);
|
|
|
|
t3_add_configured_sysctls(sc);
|
|
sc->flags |= FULL_INIT_DONE;
|
|
}
|
|
|
|
t3_intr_clear(sc);
|
|
t3_sge_start(sc);
|
|
t3_intr_enable(sc);
|
|
|
|
if (sc->params.rev >= T3_REV_C && !(sc->flags & TP_PARITY_INIT) &&
|
|
is_offload(sc) && init_tp_parity(sc) == 0)
|
|
sc->flags |= TP_PARITY_INIT;
|
|
|
|
if (sc->flags & TP_PARITY_INIT) {
|
|
t3_write_reg(sc, A_TP_INT_CAUSE, F_CMCACHEPERR | F_ARPLUTPERR);
|
|
t3_write_reg(sc, A_TP_INT_ENABLE, 0x7fbfffff);
|
|
}
|
|
|
|
if (!(sc->flags & QUEUES_BOUND)) {
|
|
bind_qsets(sc);
|
|
setup_hw_filters(sc);
|
|
sc->flags |= QUEUES_BOUND;
|
|
}
|
|
|
|
t3_sge_reset_adapter(sc);
|
|
out:
|
|
return (err);
|
|
}
|
|
|
|
/*
|
|
* Called when the last open device is closed. Does NOT undo all of cxgb_up's
|
|
* work. Specifically, the resources grabbed under FULL_INIT_DONE are released
|
|
* during controller_detach, not here.
|
|
*/
|
|
static void
|
|
cxgb_down(struct adapter *sc)
|
|
{
|
|
t3_sge_stop(sc);
|
|
t3_intr_disable(sc);
|
|
}
|
|
|
|
/*
|
|
* if_init for cxgb ports.
|
|
*/
|
|
static void
|
|
cxgb_init(void *arg)
|
|
{
|
|
struct port_info *p = arg;
|
|
struct adapter *sc = p->adapter;
|
|
|
|
ADAPTER_LOCK(sc);
|
|
cxgb_init_locked(p); /* releases adapter lock */
|
|
ADAPTER_LOCK_ASSERT_NOTOWNED(sc);
|
|
}
|
|
|
|
static int
|
|
cxgb_init_locked(struct port_info *p)
|
|
{
|
|
struct adapter *sc = p->adapter;
|
|
struct ifnet *ifp = p->ifp;
|
|
struct cmac *mac = &p->mac;
|
|
int i, rc = 0, may_sleep = 0, gave_up_lock = 0;
|
|
|
|
ADAPTER_LOCK_ASSERT_OWNED(sc);
|
|
|
|
while (!IS_DOOMED(p) && IS_BUSY(sc)) {
|
|
gave_up_lock = 1;
|
|
if (mtx_sleep(&sc->flags, &sc->lock, PCATCH, "cxgbinit", 0)) {
|
|
rc = EINTR;
|
|
goto done;
|
|
}
|
|
}
|
|
if (IS_DOOMED(p)) {
|
|
rc = ENXIO;
|
|
goto done;
|
|
}
|
|
KASSERT(!IS_BUSY(sc), ("%s: controller busy.", __func__));
|
|
|
|
/*
|
|
* The code that runs during one-time adapter initialization can sleep
|
|
* so it's important not to hold any locks across it.
|
|
*/
|
|
may_sleep = sc->flags & FULL_INIT_DONE ? 0 : 1;
|
|
|
|
if (may_sleep) {
|
|
SET_BUSY(sc);
|
|
gave_up_lock = 1;
|
|
ADAPTER_UNLOCK(sc);
|
|
}
|
|
|
|
if (sc->open_device_map == 0 && ((rc = cxgb_up(sc)) != 0))
|
|
goto done;
|
|
|
|
PORT_LOCK(p);
|
|
if (isset(&sc->open_device_map, p->port_id) &&
|
|
(ifp->if_drv_flags & IFF_DRV_RUNNING)) {
|
|
PORT_UNLOCK(p);
|
|
goto done;
|
|
}
|
|
t3_port_intr_enable(sc, p->port_id);
|
|
if (!mac->multiport)
|
|
t3_mac_init(mac);
|
|
cxgb_update_mac_settings(p);
|
|
t3_link_start(&p->phy, mac, &p->link_config);
|
|
t3_mac_enable(mac, MAC_DIRECTION_RX | MAC_DIRECTION_TX);
|
|
ifp->if_drv_flags |= IFF_DRV_RUNNING;
|
|
ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
|
|
PORT_UNLOCK(p);
|
|
|
|
for (i = p->first_qset; i < p->first_qset + p->nqsets; i++) {
|
|
struct sge_qset *qs = &sc->sge.qs[i];
|
|
struct sge_txq *txq = &qs->txq[TXQ_ETH];
|
|
|
|
callout_reset_on(&txq->txq_watchdog, hz, cxgb_tx_watchdog, qs,
|
|
txq->txq_watchdog.c_cpu);
|
|
}
|
|
|
|
/* all ok */
|
|
setbit(&sc->open_device_map, p->port_id);
|
|
callout_reset(&p->link_check_ch,
|
|
p->phy.caps & SUPPORTED_LINK_IRQ ? hz * 3 : hz / 4,
|
|
link_check_callout, p);
|
|
|
|
done:
|
|
if (may_sleep) {
|
|
ADAPTER_LOCK(sc);
|
|
KASSERT(IS_BUSY(sc), ("%s: controller not busy.", __func__));
|
|
CLR_BUSY(sc);
|
|
}
|
|
if (gave_up_lock)
|
|
wakeup_one(&sc->flags);
|
|
ADAPTER_UNLOCK(sc);
|
|
return (rc);
|
|
}
|
|
|
|
static int
|
|
cxgb_uninit_locked(struct port_info *p)
|
|
{
|
|
struct adapter *sc = p->adapter;
|
|
int rc;
|
|
|
|
ADAPTER_LOCK_ASSERT_OWNED(sc);
|
|
|
|
while (!IS_DOOMED(p) && IS_BUSY(sc)) {
|
|
if (mtx_sleep(&sc->flags, &sc->lock, PCATCH, "cxgbunin", 0)) {
|
|
rc = EINTR;
|
|
goto done;
|
|
}
|
|
}
|
|
if (IS_DOOMED(p)) {
|
|
rc = ENXIO;
|
|
goto done;
|
|
}
|
|
KASSERT(!IS_BUSY(sc), ("%s: controller busy.", __func__));
|
|
SET_BUSY(sc);
|
|
ADAPTER_UNLOCK(sc);
|
|
|
|
rc = cxgb_uninit_synchronized(p);
|
|
|
|
ADAPTER_LOCK(sc);
|
|
KASSERT(IS_BUSY(sc), ("%s: controller not busy.", __func__));
|
|
CLR_BUSY(sc);
|
|
wakeup_one(&sc->flags);
|
|
done:
|
|
ADAPTER_UNLOCK(sc);
|
|
return (rc);
|
|
}
|
|
|
|
/*
|
|
* Called on "ifconfig down", and from port_detach
|
|
*/
|
|
static int
|
|
cxgb_uninit_synchronized(struct port_info *pi)
|
|
{
|
|
struct adapter *sc = pi->adapter;
|
|
struct ifnet *ifp = pi->ifp;
|
|
|
|
/*
|
|
* taskqueue_drain may cause a deadlock if the adapter lock is held.
|
|
*/
|
|
ADAPTER_LOCK_ASSERT_NOTOWNED(sc);
|
|
|
|
/*
|
|
* Clear this port's bit from the open device map, and then drain all
|
|
* the tasks that can access/manipulate this port's port_info or ifp.
|
|
* We disable this port's interrupts here and so the slow/ext
|
|
* interrupt tasks won't be enqueued. The tick task will continue to
|
|
* be enqueued every second but the runs after this drain will not see
|
|
* this port in the open device map.
|
|
*
|
|
* A well behaved task must take open_device_map into account and ignore
|
|
* ports that are not open.
|
|
*/
|
|
clrbit(&sc->open_device_map, pi->port_id);
|
|
t3_port_intr_disable(sc, pi->port_id);
|
|
taskqueue_drain(sc->tq, &sc->slow_intr_task);
|
|
taskqueue_drain(sc->tq, &sc->tick_task);
|
|
|
|
callout_drain(&pi->link_check_ch);
|
|
taskqueue_drain(sc->tq, &pi->link_check_task);
|
|
|
|
PORT_LOCK(pi);
|
|
ifp->if_drv_flags &= ~(IFF_DRV_RUNNING | IFF_DRV_OACTIVE);
|
|
|
|
/* disable pause frames */
|
|
t3_set_reg_field(sc, A_XGM_TX_CFG + pi->mac.offset, F_TXPAUSEEN, 0);
|
|
|
|
/* Reset RX FIFO HWM */
|
|
t3_set_reg_field(sc, A_XGM_RXFIFO_CFG + pi->mac.offset,
|
|
V_RXFIFOPAUSEHWM(M_RXFIFOPAUSEHWM), 0);
|
|
|
|
DELAY(100 * 1000);
|
|
|
|
/* Wait for TXFIFO empty */
|
|
t3_wait_op_done(sc, A_XGM_TXFIFO_CFG + pi->mac.offset,
|
|
F_TXFIFO_EMPTY, 1, 20, 5);
|
|
|
|
DELAY(100 * 1000);
|
|
t3_mac_disable(&pi->mac, MAC_DIRECTION_RX);
|
|
|
|
pi->phy.ops->power_down(&pi->phy, 1);
|
|
|
|
PORT_UNLOCK(pi);
|
|
|
|
pi->link_config.link_ok = 0;
|
|
t3_os_link_changed(sc, pi->port_id, 0, 0, 0, 0, 0);
|
|
|
|
if (sc->open_device_map == 0)
|
|
cxgb_down(pi->adapter);
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Mark lro enabled or disabled in all qsets for this port
|
|
*/
|
|
static int
|
|
cxgb_set_lro(struct port_info *p, int enabled)
|
|
{
|
|
int i;
|
|
struct adapter *adp = p->adapter;
|
|
struct sge_qset *q;
|
|
|
|
for (i = 0; i < p->nqsets; i++) {
|
|
q = &adp->sge.qs[p->first_qset + i];
|
|
q->lro.enabled = (enabled != 0);
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
static int
|
|
cxgb_ioctl(struct ifnet *ifp, unsigned long command, caddr_t data)
|
|
{
|
|
struct port_info *p = ifp->if_softc;
|
|
struct adapter *sc = p->adapter;
|
|
struct ifreq *ifr = (struct ifreq *)data;
|
|
int flags, error = 0, mtu;
|
|
uint32_t mask;
|
|
|
|
switch (command) {
|
|
case SIOCSIFMTU:
|
|
ADAPTER_LOCK(sc);
|
|
error = IS_DOOMED(p) ? ENXIO : (IS_BUSY(sc) ? EBUSY : 0);
|
|
if (error) {
|
|
fail:
|
|
ADAPTER_UNLOCK(sc);
|
|
return (error);
|
|
}
|
|
|
|
mtu = ifr->ifr_mtu;
|
|
if ((mtu < ETHERMIN) || (mtu > ETHERMTU_JUMBO)) {
|
|
error = EINVAL;
|
|
} else {
|
|
ifp->if_mtu = mtu;
|
|
PORT_LOCK(p);
|
|
cxgb_update_mac_settings(p);
|
|
PORT_UNLOCK(p);
|
|
}
|
|
ADAPTER_UNLOCK(sc);
|
|
break;
|
|
case SIOCSIFFLAGS:
|
|
ADAPTER_LOCK(sc);
|
|
if (IS_DOOMED(p)) {
|
|
error = ENXIO;
|
|
goto fail;
|
|
}
|
|
if (ifp->if_flags & IFF_UP) {
|
|
if (ifp->if_drv_flags & IFF_DRV_RUNNING) {
|
|
flags = p->if_flags;
|
|
if (((ifp->if_flags ^ flags) & IFF_PROMISC) ||
|
|
((ifp->if_flags ^ flags) & IFF_ALLMULTI)) {
|
|
if (IS_BUSY(sc)) {
|
|
error = EBUSY;
|
|
goto fail;
|
|
}
|
|
PORT_LOCK(p);
|
|
cxgb_update_mac_settings(p);
|
|
PORT_UNLOCK(p);
|
|
}
|
|
ADAPTER_UNLOCK(sc);
|
|
} else
|
|
error = cxgb_init_locked(p);
|
|
p->if_flags = ifp->if_flags;
|
|
} else if (ifp->if_drv_flags & IFF_DRV_RUNNING)
|
|
error = cxgb_uninit_locked(p);
|
|
else
|
|
ADAPTER_UNLOCK(sc);
|
|
|
|
ADAPTER_LOCK_ASSERT_NOTOWNED(sc);
|
|
break;
|
|
case SIOCADDMULTI:
|
|
case SIOCDELMULTI:
|
|
ADAPTER_LOCK(sc);
|
|
error = IS_DOOMED(p) ? ENXIO : (IS_BUSY(sc) ? EBUSY : 0);
|
|
if (error)
|
|
goto fail;
|
|
|
|
if (ifp->if_drv_flags & IFF_DRV_RUNNING) {
|
|
PORT_LOCK(p);
|
|
cxgb_update_mac_settings(p);
|
|
PORT_UNLOCK(p);
|
|
}
|
|
ADAPTER_UNLOCK(sc);
|
|
|
|
break;
|
|
case SIOCSIFCAP:
|
|
ADAPTER_LOCK(sc);
|
|
error = IS_DOOMED(p) ? ENXIO : (IS_BUSY(sc) ? EBUSY : 0);
|
|
if (error)
|
|
goto fail;
|
|
|
|
mask = ifr->ifr_reqcap ^ ifp->if_capenable;
|
|
if (mask & IFCAP_TXCSUM) {
|
|
ifp->if_capenable ^= IFCAP_TXCSUM;
|
|
ifp->if_hwassist ^= (CSUM_TCP | CSUM_UDP | CSUM_IP);
|
|
|
|
if (IFCAP_TSO4 & ifp->if_capenable &&
|
|
!(IFCAP_TXCSUM & ifp->if_capenable)) {
|
|
ifp->if_capenable &= ~IFCAP_TSO4;
|
|
if_printf(ifp,
|
|
"tso4 disabled due to -txcsum.\n");
|
|
}
|
|
}
|
|
if (mask & IFCAP_TXCSUM_IPV6) {
|
|
ifp->if_capenable ^= IFCAP_TXCSUM_IPV6;
|
|
ifp->if_hwassist ^= (CSUM_UDP_IPV6 | CSUM_TCP_IPV6);
|
|
|
|
if (IFCAP_TSO6 & ifp->if_capenable &&
|
|
!(IFCAP_TXCSUM_IPV6 & ifp->if_capenable)) {
|
|
ifp->if_capenable &= ~IFCAP_TSO6;
|
|
if_printf(ifp,
|
|
"tso6 disabled due to -txcsum6.\n");
|
|
}
|
|
}
|
|
if (mask & IFCAP_RXCSUM)
|
|
ifp->if_capenable ^= IFCAP_RXCSUM;
|
|
if (mask & IFCAP_RXCSUM_IPV6)
|
|
ifp->if_capenable ^= IFCAP_RXCSUM_IPV6;
|
|
|
|
/*
|
|
* Note that we leave CSUM_TSO alone (it is always set). The
|
|
* kernel takes both IFCAP_TSOx and CSUM_TSO into account before
|
|
* sending a TSO request our way, so it's sufficient to toggle
|
|
* IFCAP_TSOx only.
|
|
*/
|
|
if (mask & IFCAP_TSO4) {
|
|
if (!(IFCAP_TSO4 & ifp->if_capenable) &&
|
|
!(IFCAP_TXCSUM & ifp->if_capenable)) {
|
|
if_printf(ifp, "enable txcsum first.\n");
|
|
error = EAGAIN;
|
|
goto fail;
|
|
}
|
|
ifp->if_capenable ^= IFCAP_TSO4;
|
|
}
|
|
if (mask & IFCAP_TSO6) {
|
|
if (!(IFCAP_TSO6 & ifp->if_capenable) &&
|
|
!(IFCAP_TXCSUM_IPV6 & ifp->if_capenable)) {
|
|
if_printf(ifp, "enable txcsum6 first.\n");
|
|
error = EAGAIN;
|
|
goto fail;
|
|
}
|
|
ifp->if_capenable ^= IFCAP_TSO6;
|
|
}
|
|
if (mask & IFCAP_LRO) {
|
|
ifp->if_capenable ^= IFCAP_LRO;
|
|
|
|
/* Safe to do this even if cxgb_up not called yet */
|
|
cxgb_set_lro(p, ifp->if_capenable & IFCAP_LRO);
|
|
}
|
|
#ifdef TCP_OFFLOAD
|
|
if (mask & IFCAP_TOE4) {
|
|
int enable = (ifp->if_capenable ^ mask) & IFCAP_TOE4;
|
|
|
|
error = toe_capability(p, enable);
|
|
if (error == 0)
|
|
ifp->if_capenable ^= mask;
|
|
}
|
|
#endif
|
|
if (mask & IFCAP_VLAN_HWTAGGING) {
|
|
ifp->if_capenable ^= IFCAP_VLAN_HWTAGGING;
|
|
if (ifp->if_drv_flags & IFF_DRV_RUNNING) {
|
|
PORT_LOCK(p);
|
|
cxgb_update_mac_settings(p);
|
|
PORT_UNLOCK(p);
|
|
}
|
|
}
|
|
if (mask & IFCAP_VLAN_MTU) {
|
|
ifp->if_capenable ^= IFCAP_VLAN_MTU;
|
|
if (ifp->if_drv_flags & IFF_DRV_RUNNING) {
|
|
PORT_LOCK(p);
|
|
cxgb_update_mac_settings(p);
|
|
PORT_UNLOCK(p);
|
|
}
|
|
}
|
|
if (mask & IFCAP_VLAN_HWTSO)
|
|
ifp->if_capenable ^= IFCAP_VLAN_HWTSO;
|
|
if (mask & IFCAP_VLAN_HWCSUM)
|
|
ifp->if_capenable ^= IFCAP_VLAN_HWCSUM;
|
|
|
|
#ifdef VLAN_CAPABILITIES
|
|
VLAN_CAPABILITIES(ifp);
|
|
#endif
|
|
ADAPTER_UNLOCK(sc);
|
|
break;
|
|
case SIOCSIFMEDIA:
|
|
case SIOCGIFMEDIA:
|
|
error = ifmedia_ioctl(ifp, ifr, &p->media, command);
|
|
break;
|
|
default:
|
|
error = ether_ioctl(ifp, command, data);
|
|
}
|
|
|
|
return (error);
|
|
}
|
|
|
|
static int
|
|
cxgb_media_change(struct ifnet *ifp)
|
|
{
|
|
return (EOPNOTSUPP);
|
|
}
|
|
|
|
/*
|
|
* Translates phy->modtype to the correct Ethernet media subtype.
|
|
*/
|
|
static int
|
|
cxgb_ifm_type(int mod)
|
|
{
|
|
switch (mod) {
|
|
case phy_modtype_sr:
|
|
return (IFM_10G_SR);
|
|
case phy_modtype_lr:
|
|
return (IFM_10G_LR);
|
|
case phy_modtype_lrm:
|
|
return (IFM_10G_LRM);
|
|
case phy_modtype_twinax:
|
|
return (IFM_10G_TWINAX);
|
|
case phy_modtype_twinax_long:
|
|
return (IFM_10G_TWINAX_LONG);
|
|
case phy_modtype_none:
|
|
return (IFM_NONE);
|
|
case phy_modtype_unknown:
|
|
return (IFM_UNKNOWN);
|
|
}
|
|
|
|
KASSERT(0, ("%s: modtype %d unknown", __func__, mod));
|
|
return (IFM_UNKNOWN);
|
|
}
|
|
|
|
/*
|
|
* Rebuilds the ifmedia list for this port, and sets the current media.
|
|
*/
|
|
static void
|
|
cxgb_build_medialist(struct port_info *p)
|
|
{
|
|
struct cphy *phy = &p->phy;
|
|
struct ifmedia *media = &p->media;
|
|
int mod = phy->modtype;
|
|
int m = IFM_ETHER | IFM_FDX;
|
|
|
|
PORT_LOCK(p);
|
|
|
|
ifmedia_removeall(media);
|
|
if (phy->caps & SUPPORTED_TP && phy->caps & SUPPORTED_Autoneg) {
|
|
/* Copper (RJ45) */
|
|
|
|
if (phy->caps & SUPPORTED_10000baseT_Full)
|
|
ifmedia_add(media, m | IFM_10G_T, mod, NULL);
|
|
|
|
if (phy->caps & SUPPORTED_1000baseT_Full)
|
|
ifmedia_add(media, m | IFM_1000_T, mod, NULL);
|
|
|
|
if (phy->caps & SUPPORTED_100baseT_Full)
|
|
ifmedia_add(media, m | IFM_100_TX, mod, NULL);
|
|
|
|
if (phy->caps & SUPPORTED_10baseT_Full)
|
|
ifmedia_add(media, m | IFM_10_T, mod, NULL);
|
|
|
|
ifmedia_add(media, IFM_ETHER | IFM_AUTO, mod, NULL);
|
|
ifmedia_set(media, IFM_ETHER | IFM_AUTO);
|
|
|
|
} else if (phy->caps & SUPPORTED_TP) {
|
|
/* Copper (CX4) */
|
|
|
|
KASSERT(phy->caps & SUPPORTED_10000baseT_Full,
|
|
("%s: unexpected cap 0x%x", __func__, phy->caps));
|
|
|
|
ifmedia_add(media, m | IFM_10G_CX4, mod, NULL);
|
|
ifmedia_set(media, m | IFM_10G_CX4);
|
|
|
|
} else if (phy->caps & SUPPORTED_FIBRE &&
|
|
phy->caps & SUPPORTED_10000baseT_Full) {
|
|
/* 10G optical (but includes SFP+ twinax) */
|
|
|
|
m |= cxgb_ifm_type(mod);
|
|
if (IFM_SUBTYPE(m) == IFM_NONE)
|
|
m &= ~IFM_FDX;
|
|
|
|
ifmedia_add(media, m, mod, NULL);
|
|
ifmedia_set(media, m);
|
|
|
|
} else if (phy->caps & SUPPORTED_FIBRE &&
|
|
phy->caps & SUPPORTED_1000baseT_Full) {
|
|
/* 1G optical */
|
|
|
|
/* XXX: Lie and claim to be SX, could actually be any 1G-X */
|
|
ifmedia_add(media, m | IFM_1000_SX, mod, NULL);
|
|
ifmedia_set(media, m | IFM_1000_SX);
|
|
|
|
} else {
|
|
KASSERT(0, ("%s: don't know how to handle 0x%x.", __func__,
|
|
phy->caps));
|
|
}
|
|
|
|
PORT_UNLOCK(p);
|
|
}
|
|
|
|
static void
|
|
cxgb_media_status(struct ifnet *ifp, struct ifmediareq *ifmr)
|
|
{
|
|
struct port_info *p = ifp->if_softc;
|
|
struct ifmedia_entry *cur = p->media.ifm_cur;
|
|
int speed = p->link_config.speed;
|
|
|
|
if (cur->ifm_data != p->phy.modtype) {
|
|
cxgb_build_medialist(p);
|
|
cur = p->media.ifm_cur;
|
|
}
|
|
|
|
ifmr->ifm_status = IFM_AVALID;
|
|
if (!p->link_config.link_ok)
|
|
return;
|
|
|
|
ifmr->ifm_status |= IFM_ACTIVE;
|
|
|
|
/*
|
|
* active and current will differ iff current media is autoselect. That
|
|
* can happen only for copper RJ45.
|
|
*/
|
|
if (IFM_SUBTYPE(cur->ifm_media) != IFM_AUTO)
|
|
return;
|
|
KASSERT(p->phy.caps & SUPPORTED_TP && p->phy.caps & SUPPORTED_Autoneg,
|
|
("%s: unexpected PHY caps 0x%x", __func__, p->phy.caps));
|
|
|
|
ifmr->ifm_active = IFM_ETHER | IFM_FDX;
|
|
if (speed == SPEED_10000)
|
|
ifmr->ifm_active |= IFM_10G_T;
|
|
else if (speed == SPEED_1000)
|
|
ifmr->ifm_active |= IFM_1000_T;
|
|
else if (speed == SPEED_100)
|
|
ifmr->ifm_active |= IFM_100_TX;
|
|
else if (speed == SPEED_10)
|
|
ifmr->ifm_active |= IFM_10_T;
|
|
else
|
|
KASSERT(0, ("%s: link up but speed unknown (%u)", __func__,
|
|
speed));
|
|
}
|
|
|
|
static uint64_t
|
|
cxgb_get_counter(struct ifnet *ifp, ift_counter c)
|
|
{
|
|
struct port_info *pi = ifp->if_softc;
|
|
struct adapter *sc = pi->adapter;
|
|
struct cmac *mac = &pi->mac;
|
|
struct mac_stats *mstats = &mac->stats;
|
|
|
|
cxgb_refresh_stats(pi);
|
|
|
|
switch (c) {
|
|
case IFCOUNTER_IPACKETS:
|
|
return (mstats->rx_frames);
|
|
|
|
case IFCOUNTER_IERRORS:
|
|
return (mstats->rx_jabber + mstats->rx_data_errs +
|
|
mstats->rx_sequence_errs + mstats->rx_runt +
|
|
mstats->rx_too_long + mstats->rx_mac_internal_errs +
|
|
mstats->rx_short + mstats->rx_fcs_errs);
|
|
|
|
case IFCOUNTER_OPACKETS:
|
|
return (mstats->tx_frames);
|
|
|
|
case IFCOUNTER_OERRORS:
|
|
return (mstats->tx_excess_collisions + mstats->tx_underrun +
|
|
mstats->tx_len_errs + mstats->tx_mac_internal_errs +
|
|
mstats->tx_excess_deferral + mstats->tx_fcs_errs);
|
|
|
|
case IFCOUNTER_COLLISIONS:
|
|
return (mstats->tx_total_collisions);
|
|
|
|
case IFCOUNTER_IBYTES:
|
|
return (mstats->rx_octets);
|
|
|
|
case IFCOUNTER_OBYTES:
|
|
return (mstats->tx_octets);
|
|
|
|
case IFCOUNTER_IMCASTS:
|
|
return (mstats->rx_mcast_frames);
|
|
|
|
case IFCOUNTER_OMCASTS:
|
|
return (mstats->tx_mcast_frames);
|
|
|
|
case IFCOUNTER_IQDROPS:
|
|
return (mstats->rx_cong_drops);
|
|
|
|
case IFCOUNTER_OQDROPS: {
|
|
int i;
|
|
uint64_t drops;
|
|
|
|
drops = 0;
|
|
if (sc->flags & FULL_INIT_DONE) {
|
|
for (i = pi->first_qset; i < pi->first_qset + pi->nqsets; i++)
|
|
drops += sc->sge.qs[i].txq[TXQ_ETH].txq_mr->br_drops;
|
|
}
|
|
|
|
return (drops);
|
|
|
|
}
|
|
|
|
default:
|
|
return (if_get_counter_default(ifp, c));
|
|
}
|
|
}
|
|
|
|
static void
|
|
cxgb_async_intr(void *data)
|
|
{
|
|
adapter_t *sc = data;
|
|
|
|
t3_write_reg(sc, A_PL_INT_ENABLE0, 0);
|
|
(void) t3_read_reg(sc, A_PL_INT_ENABLE0);
|
|
taskqueue_enqueue(sc->tq, &sc->slow_intr_task);
|
|
}
|
|
|
|
static void
|
|
link_check_callout(void *arg)
|
|
{
|
|
struct port_info *pi = arg;
|
|
struct adapter *sc = pi->adapter;
|
|
|
|
if (!isset(&sc->open_device_map, pi->port_id))
|
|
return;
|
|
|
|
taskqueue_enqueue(sc->tq, &pi->link_check_task);
|
|
}
|
|
|
|
static void
|
|
check_link_status(void *arg, int pending)
|
|
{
|
|
struct port_info *pi = arg;
|
|
struct adapter *sc = pi->adapter;
|
|
|
|
if (!isset(&sc->open_device_map, pi->port_id))
|
|
return;
|
|
|
|
t3_link_changed(sc, pi->port_id);
|
|
|
|
if (pi->link_fault || !(pi->phy.caps & SUPPORTED_LINK_IRQ) ||
|
|
pi->link_config.link_ok == 0)
|
|
callout_reset(&pi->link_check_ch, hz, link_check_callout, pi);
|
|
}
|
|
|
|
void
|
|
t3_os_link_intr(struct port_info *pi)
|
|
{
|
|
/*
|
|
* Schedule a link check in the near future. If the link is flapping
|
|
* rapidly we'll keep resetting the callout and delaying the check until
|
|
* things stabilize a bit.
|
|
*/
|
|
callout_reset(&pi->link_check_ch, hz / 4, link_check_callout, pi);
|
|
}
|
|
|
|
static void
|
|
check_t3b2_mac(struct adapter *sc)
|
|
{
|
|
int i;
|
|
|
|
if (sc->flags & CXGB_SHUTDOWN)
|
|
return;
|
|
|
|
for_each_port(sc, i) {
|
|
struct port_info *p = &sc->port[i];
|
|
int status;
|
|
#ifdef INVARIANTS
|
|
struct ifnet *ifp = p->ifp;
|
|
#endif
|
|
|
|
if (!isset(&sc->open_device_map, p->port_id) || p->link_fault ||
|
|
!p->link_config.link_ok)
|
|
continue;
|
|
|
|
KASSERT(ifp->if_drv_flags & IFF_DRV_RUNNING,
|
|
("%s: state mismatch (drv_flags %x, device_map %x)",
|
|
__func__, ifp->if_drv_flags, sc->open_device_map));
|
|
|
|
PORT_LOCK(p);
|
|
status = t3b2_mac_watchdog_task(&p->mac);
|
|
if (status == 1)
|
|
p->mac.stats.num_toggled++;
|
|
else if (status == 2) {
|
|
struct cmac *mac = &p->mac;
|
|
|
|
cxgb_update_mac_settings(p);
|
|
t3_link_start(&p->phy, mac, &p->link_config);
|
|
t3_mac_enable(mac, MAC_DIRECTION_RX | MAC_DIRECTION_TX);
|
|
t3_port_intr_enable(sc, p->port_id);
|
|
p->mac.stats.num_resets++;
|
|
}
|
|
PORT_UNLOCK(p);
|
|
}
|
|
}
|
|
|
|
static void
|
|
cxgb_tick(void *arg)
|
|
{
|
|
adapter_t *sc = (adapter_t *)arg;
|
|
|
|
if (sc->flags & CXGB_SHUTDOWN)
|
|
return;
|
|
|
|
taskqueue_enqueue(sc->tq, &sc->tick_task);
|
|
callout_reset(&sc->cxgb_tick_ch, hz, cxgb_tick, sc);
|
|
}
|
|
|
|
void
|
|
cxgb_refresh_stats(struct port_info *pi)
|
|
{
|
|
struct timeval tv;
|
|
const struct timeval interval = {0, 250000}; /* 250ms */
|
|
|
|
getmicrotime(&tv);
|
|
timevalsub(&tv, &interval);
|
|
if (timevalcmp(&tv, &pi->last_refreshed, <))
|
|
return;
|
|
|
|
PORT_LOCK(pi);
|
|
t3_mac_update_stats(&pi->mac);
|
|
PORT_UNLOCK(pi);
|
|
getmicrotime(&pi->last_refreshed);
|
|
}
|
|
|
|
static void
|
|
cxgb_tick_handler(void *arg, int count)
|
|
{
|
|
adapter_t *sc = (adapter_t *)arg;
|
|
const struct adapter_params *p = &sc->params;
|
|
int i;
|
|
uint32_t cause, reset;
|
|
|
|
if (sc->flags & CXGB_SHUTDOWN || !(sc->flags & FULL_INIT_DONE))
|
|
return;
|
|
|
|
if (p->rev == T3_REV_B2 && p->nports < 4 && sc->open_device_map)
|
|
check_t3b2_mac(sc);
|
|
|
|
cause = t3_read_reg(sc, A_SG_INT_CAUSE) & (F_RSPQSTARVE | F_FLEMPTY);
|
|
if (cause) {
|
|
struct sge_qset *qs = &sc->sge.qs[0];
|
|
uint32_t mask, v;
|
|
|
|
v = t3_read_reg(sc, A_SG_RSPQ_FL_STATUS) & ~0xff00;
|
|
|
|
mask = 1;
|
|
for (i = 0; i < SGE_QSETS; i++) {
|
|
if (v & mask)
|
|
qs[i].rspq.starved++;
|
|
mask <<= 1;
|
|
}
|
|
|
|
mask <<= SGE_QSETS; /* skip RSPQXDISABLED */
|
|
|
|
for (i = 0; i < SGE_QSETS * 2; i++) {
|
|
if (v & mask) {
|
|
qs[i / 2].fl[i % 2].empty++;
|
|
}
|
|
mask <<= 1;
|
|
}
|
|
|
|
/* clear */
|
|
t3_write_reg(sc, A_SG_RSPQ_FL_STATUS, v);
|
|
t3_write_reg(sc, A_SG_INT_CAUSE, cause);
|
|
}
|
|
|
|
for (i = 0; i < sc->params.nports; i++) {
|
|
struct port_info *pi = &sc->port[i];
|
|
struct cmac *mac = &pi->mac;
|
|
|
|
if (!isset(&sc->open_device_map, pi->port_id))
|
|
continue;
|
|
|
|
cxgb_refresh_stats(pi);
|
|
|
|
if (mac->multiport)
|
|
continue;
|
|
|
|
/* Count rx fifo overflows, once per second */
|
|
cause = t3_read_reg(sc, A_XGM_INT_CAUSE + mac->offset);
|
|
reset = 0;
|
|
if (cause & F_RXFIFO_OVERFLOW) {
|
|
mac->stats.rx_fifo_ovfl++;
|
|
reset |= F_RXFIFO_OVERFLOW;
|
|
}
|
|
t3_write_reg(sc, A_XGM_INT_CAUSE + mac->offset, reset);
|
|
}
|
|
}
|
|
|
|
static void
|
|
touch_bars(device_t dev)
|
|
{
|
|
/*
|
|
* Don't enable yet
|
|
*/
|
|
#if !defined(__LP64__) && 0
|
|
u32 v;
|
|
|
|
pci_read_config_dword(pdev, PCI_BASE_ADDRESS_1, &v);
|
|
pci_write_config_dword(pdev, PCI_BASE_ADDRESS_1, v);
|
|
pci_read_config_dword(pdev, PCI_BASE_ADDRESS_3, &v);
|
|
pci_write_config_dword(pdev, PCI_BASE_ADDRESS_3, v);
|
|
pci_read_config_dword(pdev, PCI_BASE_ADDRESS_5, &v);
|
|
pci_write_config_dword(pdev, PCI_BASE_ADDRESS_5, v);
|
|
#endif
|
|
}
|
|
|
|
static int
|
|
set_eeprom(struct port_info *pi, const uint8_t *data, int len, int offset)
|
|
{
|
|
uint8_t *buf;
|
|
int err = 0;
|
|
u32 aligned_offset, aligned_len, *p;
|
|
struct adapter *adapter = pi->adapter;
|
|
|
|
|
|
aligned_offset = offset & ~3;
|
|
aligned_len = (len + (offset & 3) + 3) & ~3;
|
|
|
|
if (aligned_offset != offset || aligned_len != len) {
|
|
buf = malloc(aligned_len, M_DEVBUF, M_WAITOK|M_ZERO);
|
|
if (!buf)
|
|
return (ENOMEM);
|
|
err = t3_seeprom_read(adapter, aligned_offset, (u32 *)buf);
|
|
if (!err && aligned_len > 4)
|
|
err = t3_seeprom_read(adapter,
|
|
aligned_offset + aligned_len - 4,
|
|
(u32 *)&buf[aligned_len - 4]);
|
|
if (err)
|
|
goto out;
|
|
memcpy(buf + (offset & 3), data, len);
|
|
} else
|
|
buf = (uint8_t *)(uintptr_t)data;
|
|
|
|
err = t3_seeprom_wp(adapter, 0);
|
|
if (err)
|
|
goto out;
|
|
|
|
for (p = (u32 *)buf; !err && aligned_len; aligned_len -= 4, p++) {
|
|
err = t3_seeprom_write(adapter, aligned_offset, *p);
|
|
aligned_offset += 4;
|
|
}
|
|
|
|
if (!err)
|
|
err = t3_seeprom_wp(adapter, 1);
|
|
out:
|
|
if (buf != data)
|
|
free(buf, M_DEVBUF);
|
|
return err;
|
|
}
|
|
|
|
|
|
static int
|
|
in_range(int val, int lo, int hi)
|
|
{
|
|
return val < 0 || (val <= hi && val >= lo);
|
|
}
|
|
|
|
static int
|
|
cxgb_extension_open(struct cdev *dev, int flags, int fmp, struct thread *td)
|
|
{
|
|
return (0);
|
|
}
|
|
|
|
static int
|
|
cxgb_extension_close(struct cdev *dev, int flags, int fmt, struct thread *td)
|
|
{
|
|
return (0);
|
|
}
|
|
|
|
static int
|
|
cxgb_extension_ioctl(struct cdev *dev, unsigned long cmd, caddr_t data,
|
|
int fflag, struct thread *td)
|
|
{
|
|
int mmd, error = 0;
|
|
struct port_info *pi = dev->si_drv1;
|
|
adapter_t *sc = pi->adapter;
|
|
|
|
#ifdef PRIV_SUPPORTED
|
|
if (priv_check(td, PRIV_DRIVER)) {
|
|
if (cxgb_debug)
|
|
printf("user does not have access to privileged ioctls\n");
|
|
return (EPERM);
|
|
}
|
|
#else
|
|
if (suser(td)) {
|
|
if (cxgb_debug)
|
|
printf("user does not have access to privileged ioctls\n");
|
|
return (EPERM);
|
|
}
|
|
#endif
|
|
|
|
switch (cmd) {
|
|
case CHELSIO_GET_MIIREG: {
|
|
uint32_t val;
|
|
struct cphy *phy = &pi->phy;
|
|
struct ch_mii_data *mid = (struct ch_mii_data *)data;
|
|
|
|
if (!phy->mdio_read)
|
|
return (EOPNOTSUPP);
|
|
if (is_10G(sc)) {
|
|
mmd = mid->phy_id >> 8;
|
|
if (!mmd)
|
|
mmd = MDIO_DEV_PCS;
|
|
else if (mmd > MDIO_DEV_VEND2)
|
|
return (EINVAL);
|
|
|
|
error = phy->mdio_read(sc, mid->phy_id & 0x1f, mmd,
|
|
mid->reg_num, &val);
|
|
} else
|
|
error = phy->mdio_read(sc, mid->phy_id & 0x1f, 0,
|
|
mid->reg_num & 0x1f, &val);
|
|
if (error == 0)
|
|
mid->val_out = val;
|
|
break;
|
|
}
|
|
case CHELSIO_SET_MIIREG: {
|
|
struct cphy *phy = &pi->phy;
|
|
struct ch_mii_data *mid = (struct ch_mii_data *)data;
|
|
|
|
if (!phy->mdio_write)
|
|
return (EOPNOTSUPP);
|
|
if (is_10G(sc)) {
|
|
mmd = mid->phy_id >> 8;
|
|
if (!mmd)
|
|
mmd = MDIO_DEV_PCS;
|
|
else if (mmd > MDIO_DEV_VEND2)
|
|
return (EINVAL);
|
|
|
|
error = phy->mdio_write(sc, mid->phy_id & 0x1f,
|
|
mmd, mid->reg_num, mid->val_in);
|
|
} else
|
|
error = phy->mdio_write(sc, mid->phy_id & 0x1f, 0,
|
|
mid->reg_num & 0x1f,
|
|
mid->val_in);
|
|
break;
|
|
}
|
|
case CHELSIO_SETREG: {
|
|
struct ch_reg *edata = (struct ch_reg *)data;
|
|
if ((edata->addr & 0x3) != 0 || edata->addr >= sc->mmio_len)
|
|
return (EFAULT);
|
|
t3_write_reg(sc, edata->addr, edata->val);
|
|
break;
|
|
}
|
|
case CHELSIO_GETREG: {
|
|
struct ch_reg *edata = (struct ch_reg *)data;
|
|
if ((edata->addr & 0x3) != 0 || edata->addr >= sc->mmio_len)
|
|
return (EFAULT);
|
|
edata->val = t3_read_reg(sc, edata->addr);
|
|
break;
|
|
}
|
|
case CHELSIO_GET_SGE_CONTEXT: {
|
|
struct ch_cntxt *ecntxt = (struct ch_cntxt *)data;
|
|
mtx_lock_spin(&sc->sge.reg_lock);
|
|
switch (ecntxt->cntxt_type) {
|
|
case CNTXT_TYPE_EGRESS:
|
|
error = -t3_sge_read_ecntxt(sc, ecntxt->cntxt_id,
|
|
ecntxt->data);
|
|
break;
|
|
case CNTXT_TYPE_FL:
|
|
error = -t3_sge_read_fl(sc, ecntxt->cntxt_id,
|
|
ecntxt->data);
|
|
break;
|
|
case CNTXT_TYPE_RSP:
|
|
error = -t3_sge_read_rspq(sc, ecntxt->cntxt_id,
|
|
ecntxt->data);
|
|
break;
|
|
case CNTXT_TYPE_CQ:
|
|
error = -t3_sge_read_cq(sc, ecntxt->cntxt_id,
|
|
ecntxt->data);
|
|
break;
|
|
default:
|
|
error = EINVAL;
|
|
break;
|
|
}
|
|
mtx_unlock_spin(&sc->sge.reg_lock);
|
|
break;
|
|
}
|
|
case CHELSIO_GET_SGE_DESC: {
|
|
struct ch_desc *edesc = (struct ch_desc *)data;
|
|
int ret;
|
|
if (edesc->queue_num >= SGE_QSETS * 6)
|
|
return (EINVAL);
|
|
ret = t3_get_desc(&sc->sge.qs[edesc->queue_num / 6],
|
|
edesc->queue_num % 6, edesc->idx, edesc->data);
|
|
if (ret < 0)
|
|
return (EINVAL);
|
|
edesc->size = ret;
|
|
break;
|
|
}
|
|
case CHELSIO_GET_QSET_PARAMS: {
|
|
struct qset_params *q;
|
|
struct ch_qset_params *t = (struct ch_qset_params *)data;
|
|
int q1 = pi->first_qset;
|
|
int nqsets = pi->nqsets;
|
|
int i;
|
|
|
|
if (t->qset_idx >= nqsets)
|
|
return EINVAL;
|
|
|
|
i = q1 + t->qset_idx;
|
|
q = &sc->params.sge.qset[i];
|
|
t->rspq_size = q->rspq_size;
|
|
t->txq_size[0] = q->txq_size[0];
|
|
t->txq_size[1] = q->txq_size[1];
|
|
t->txq_size[2] = q->txq_size[2];
|
|
t->fl_size[0] = q->fl_size;
|
|
t->fl_size[1] = q->jumbo_size;
|
|
t->polling = q->polling;
|
|
t->lro = q->lro;
|
|
t->intr_lat = q->coalesce_usecs;
|
|
t->cong_thres = q->cong_thres;
|
|
t->qnum = i;
|
|
|
|
if ((sc->flags & FULL_INIT_DONE) == 0)
|
|
t->vector = 0;
|
|
else if (sc->flags & USING_MSIX)
|
|
t->vector = rman_get_start(sc->msix_irq_res[i]);
|
|
else
|
|
t->vector = rman_get_start(sc->irq_res);
|
|
|
|
break;
|
|
}
|
|
case CHELSIO_GET_QSET_NUM: {
|
|
struct ch_reg *edata = (struct ch_reg *)data;
|
|
edata->val = pi->nqsets;
|
|
break;
|
|
}
|
|
case CHELSIO_LOAD_FW: {
|
|
uint8_t *fw_data;
|
|
uint32_t vers;
|
|
struct ch_mem_range *t = (struct ch_mem_range *)data;
|
|
|
|
/*
|
|
* You're allowed to load a firmware only before FULL_INIT_DONE
|
|
*
|
|
* FW_UPTODATE is also set so the rest of the initialization
|
|
* will not overwrite what was loaded here. This gives you the
|
|
* flexibility to load any firmware (and maybe shoot yourself in
|
|
* the foot).
|
|
*/
|
|
|
|
ADAPTER_LOCK(sc);
|
|
if (sc->open_device_map || sc->flags & FULL_INIT_DONE) {
|
|
ADAPTER_UNLOCK(sc);
|
|
return (EBUSY);
|
|
}
|
|
|
|
fw_data = malloc(t->len, M_DEVBUF, M_NOWAIT);
|
|
if (!fw_data)
|
|
error = ENOMEM;
|
|
else
|
|
error = copyin(t->buf, fw_data, t->len);
|
|
|
|
if (!error)
|
|
error = -t3_load_fw(sc, fw_data, t->len);
|
|
|
|
if (t3_get_fw_version(sc, &vers) == 0) {
|
|
snprintf(&sc->fw_version[0], sizeof(sc->fw_version),
|
|
"%d.%d.%d", G_FW_VERSION_MAJOR(vers),
|
|
G_FW_VERSION_MINOR(vers), G_FW_VERSION_MICRO(vers));
|
|
}
|
|
|
|
if (!error)
|
|
sc->flags |= FW_UPTODATE;
|
|
|
|
free(fw_data, M_DEVBUF);
|
|
ADAPTER_UNLOCK(sc);
|
|
break;
|
|
}
|
|
case CHELSIO_LOAD_BOOT: {
|
|
uint8_t *boot_data;
|
|
struct ch_mem_range *t = (struct ch_mem_range *)data;
|
|
|
|
boot_data = malloc(t->len, M_DEVBUF, M_NOWAIT);
|
|
if (!boot_data)
|
|
return ENOMEM;
|
|
|
|
error = copyin(t->buf, boot_data, t->len);
|
|
if (!error)
|
|
error = -t3_load_boot(sc, boot_data, t->len);
|
|
|
|
free(boot_data, M_DEVBUF);
|
|
break;
|
|
}
|
|
case CHELSIO_GET_PM: {
|
|
struct ch_pm *m = (struct ch_pm *)data;
|
|
struct tp_params *p = &sc->params.tp;
|
|
|
|
if (!is_offload(sc))
|
|
return (EOPNOTSUPP);
|
|
|
|
m->tx_pg_sz = p->tx_pg_size;
|
|
m->tx_num_pg = p->tx_num_pgs;
|
|
m->rx_pg_sz = p->rx_pg_size;
|
|
m->rx_num_pg = p->rx_num_pgs;
|
|
m->pm_total = p->pmtx_size + p->chan_rx_size * p->nchan;
|
|
|
|
break;
|
|
}
|
|
case CHELSIO_SET_PM: {
|
|
struct ch_pm *m = (struct ch_pm *)data;
|
|
struct tp_params *p = &sc->params.tp;
|
|
|
|
if (!is_offload(sc))
|
|
return (EOPNOTSUPP);
|
|
if (sc->flags & FULL_INIT_DONE)
|
|
return (EBUSY);
|
|
|
|
if (!m->rx_pg_sz || (m->rx_pg_sz & (m->rx_pg_sz - 1)) ||
|
|
!m->tx_pg_sz || (m->tx_pg_sz & (m->tx_pg_sz - 1)))
|
|
return (EINVAL); /* not power of 2 */
|
|
if (!(m->rx_pg_sz & 0x14000))
|
|
return (EINVAL); /* not 16KB or 64KB */
|
|
if (!(m->tx_pg_sz & 0x1554000))
|
|
return (EINVAL);
|
|
if (m->tx_num_pg == -1)
|
|
m->tx_num_pg = p->tx_num_pgs;
|
|
if (m->rx_num_pg == -1)
|
|
m->rx_num_pg = p->rx_num_pgs;
|
|
if (m->tx_num_pg % 24 || m->rx_num_pg % 24)
|
|
return (EINVAL);
|
|
if (m->rx_num_pg * m->rx_pg_sz > p->chan_rx_size ||
|
|
m->tx_num_pg * m->tx_pg_sz > p->chan_tx_size)
|
|
return (EINVAL);
|
|
|
|
p->rx_pg_size = m->rx_pg_sz;
|
|
p->tx_pg_size = m->tx_pg_sz;
|
|
p->rx_num_pgs = m->rx_num_pg;
|
|
p->tx_num_pgs = m->tx_num_pg;
|
|
break;
|
|
}
|
|
case CHELSIO_SETMTUTAB: {
|
|
struct ch_mtus *m = (struct ch_mtus *)data;
|
|
int i;
|
|
|
|
if (!is_offload(sc))
|
|
return (EOPNOTSUPP);
|
|
if (offload_running(sc))
|
|
return (EBUSY);
|
|
if (m->nmtus != NMTUS)
|
|
return (EINVAL);
|
|
if (m->mtus[0] < 81) /* accommodate SACK */
|
|
return (EINVAL);
|
|
|
|
/*
|
|
* MTUs must be in ascending order
|
|
*/
|
|
for (i = 1; i < NMTUS; ++i)
|
|
if (m->mtus[i] < m->mtus[i - 1])
|
|
return (EINVAL);
|
|
|
|
memcpy(sc->params.mtus, m->mtus, sizeof(sc->params.mtus));
|
|
break;
|
|
}
|
|
case CHELSIO_GETMTUTAB: {
|
|
struct ch_mtus *m = (struct ch_mtus *)data;
|
|
|
|
if (!is_offload(sc))
|
|
return (EOPNOTSUPP);
|
|
|
|
memcpy(m->mtus, sc->params.mtus, sizeof(m->mtus));
|
|
m->nmtus = NMTUS;
|
|
break;
|
|
}
|
|
case CHELSIO_GET_MEM: {
|
|
struct ch_mem_range *t = (struct ch_mem_range *)data;
|
|
struct mc7 *mem;
|
|
uint8_t *useraddr;
|
|
u64 buf[32];
|
|
|
|
/*
|
|
* Use these to avoid modifying len/addr in the return
|
|
* struct
|
|
*/
|
|
uint32_t len = t->len, addr = t->addr;
|
|
|
|
if (!is_offload(sc))
|
|
return (EOPNOTSUPP);
|
|
if (!(sc->flags & FULL_INIT_DONE))
|
|
return (EIO); /* need the memory controllers */
|
|
if ((addr & 0x7) || (len & 0x7))
|
|
return (EINVAL);
|
|
if (t->mem_id == MEM_CM)
|
|
mem = &sc->cm;
|
|
else if (t->mem_id == MEM_PMRX)
|
|
mem = &sc->pmrx;
|
|
else if (t->mem_id == MEM_PMTX)
|
|
mem = &sc->pmtx;
|
|
else
|
|
return (EINVAL);
|
|
|
|
/*
|
|
* Version scheme:
|
|
* bits 0..9: chip version
|
|
* bits 10..15: chip revision
|
|
*/
|
|
t->version = 3 | (sc->params.rev << 10);
|
|
|
|
/*
|
|
* Read 256 bytes at a time as len can be large and we don't
|
|
* want to use huge intermediate buffers.
|
|
*/
|
|
useraddr = (uint8_t *)t->buf;
|
|
while (len) {
|
|
unsigned int chunk = min(len, sizeof(buf));
|
|
|
|
error = t3_mc7_bd_read(mem, addr / 8, chunk / 8, buf);
|
|
if (error)
|
|
return (-error);
|
|
if (copyout(buf, useraddr, chunk))
|
|
return (EFAULT);
|
|
useraddr += chunk;
|
|
addr += chunk;
|
|
len -= chunk;
|
|
}
|
|
break;
|
|
}
|
|
case CHELSIO_READ_TCAM_WORD: {
|
|
struct ch_tcam_word *t = (struct ch_tcam_word *)data;
|
|
|
|
if (!is_offload(sc))
|
|
return (EOPNOTSUPP);
|
|
if (!(sc->flags & FULL_INIT_DONE))
|
|
return (EIO); /* need MC5 */
|
|
return -t3_read_mc5_range(&sc->mc5, t->addr, 1, t->buf);
|
|
break;
|
|
}
|
|
case CHELSIO_SET_TRACE_FILTER: {
|
|
struct ch_trace *t = (struct ch_trace *)data;
|
|
const struct trace_params *tp;
|
|
|
|
tp = (const struct trace_params *)&t->sip;
|
|
if (t->config_tx)
|
|
t3_config_trace_filter(sc, tp, 0, t->invert_match,
|
|
t->trace_tx);
|
|
if (t->config_rx)
|
|
t3_config_trace_filter(sc, tp, 1, t->invert_match,
|
|
t->trace_rx);
|
|
break;
|
|
}
|
|
case CHELSIO_SET_PKTSCHED: {
|
|
struct ch_pktsched_params *p = (struct ch_pktsched_params *)data;
|
|
if (sc->open_device_map == 0)
|
|
return (EAGAIN);
|
|
send_pktsched_cmd(sc, p->sched, p->idx, p->min, p->max,
|
|
p->binding);
|
|
break;
|
|
}
|
|
case CHELSIO_IFCONF_GETREGS: {
|
|
struct ch_ifconf_regs *regs = (struct ch_ifconf_regs *)data;
|
|
int reglen = cxgb_get_regs_len();
|
|
uint8_t *buf = malloc(reglen, M_DEVBUF, M_NOWAIT);
|
|
if (buf == NULL) {
|
|
return (ENOMEM);
|
|
}
|
|
if (regs->len > reglen)
|
|
regs->len = reglen;
|
|
else if (regs->len < reglen)
|
|
error = ENOBUFS;
|
|
|
|
if (!error) {
|
|
cxgb_get_regs(sc, regs, buf);
|
|
error = copyout(buf, regs->data, reglen);
|
|
}
|
|
free(buf, M_DEVBUF);
|
|
|
|
break;
|
|
}
|
|
case CHELSIO_SET_HW_SCHED: {
|
|
struct ch_hw_sched *t = (struct ch_hw_sched *)data;
|
|
unsigned int ticks_per_usec = core_ticks_per_usec(sc);
|
|
|
|
if ((sc->flags & FULL_INIT_DONE) == 0)
|
|
return (EAGAIN); /* need TP to be initialized */
|
|
if (t->sched >= NTX_SCHED || !in_range(t->mode, 0, 1) ||
|
|
!in_range(t->channel, 0, 1) ||
|
|
!in_range(t->kbps, 0, 10000000) ||
|
|
!in_range(t->class_ipg, 0, 10000 * 65535 / ticks_per_usec) ||
|
|
!in_range(t->flow_ipg, 0,
|
|
dack_ticks_to_usec(sc, 0x7ff)))
|
|
return (EINVAL);
|
|
|
|
if (t->kbps >= 0) {
|
|
error = t3_config_sched(sc, t->kbps, t->sched);
|
|
if (error < 0)
|
|
return (-error);
|
|
}
|
|
if (t->class_ipg >= 0)
|
|
t3_set_sched_ipg(sc, t->sched, t->class_ipg);
|
|
if (t->flow_ipg >= 0) {
|
|
t->flow_ipg *= 1000; /* us -> ns */
|
|
t3_set_pace_tbl(sc, &t->flow_ipg, t->sched, 1);
|
|
}
|
|
if (t->mode >= 0) {
|
|
int bit = 1 << (S_TX_MOD_TIMER_MODE + t->sched);
|
|
|
|
t3_set_reg_field(sc, A_TP_TX_MOD_QUEUE_REQ_MAP,
|
|
bit, t->mode ? bit : 0);
|
|
}
|
|
if (t->channel >= 0)
|
|
t3_set_reg_field(sc, A_TP_TX_MOD_QUEUE_REQ_MAP,
|
|
1 << t->sched, t->channel << t->sched);
|
|
break;
|
|
}
|
|
case CHELSIO_GET_EEPROM: {
|
|
int i;
|
|
struct ch_eeprom *e = (struct ch_eeprom *)data;
|
|
uint8_t *buf = malloc(EEPROMSIZE, M_DEVBUF, M_NOWAIT);
|
|
|
|
if (buf == NULL) {
|
|
return (ENOMEM);
|
|
}
|
|
e->magic = EEPROM_MAGIC;
|
|
for (i = e->offset & ~3; !error && i < e->offset + e->len; i += 4)
|
|
error = -t3_seeprom_read(sc, i, (uint32_t *)&buf[i]);
|
|
|
|
if (!error)
|
|
error = copyout(buf + e->offset, e->data, e->len);
|
|
|
|
free(buf, M_DEVBUF);
|
|
break;
|
|
}
|
|
case CHELSIO_CLEAR_STATS: {
|
|
if (!(sc->flags & FULL_INIT_DONE))
|
|
return EAGAIN;
|
|
|
|
PORT_LOCK(pi);
|
|
t3_mac_update_stats(&pi->mac);
|
|
memset(&pi->mac.stats, 0, sizeof(pi->mac.stats));
|
|
PORT_UNLOCK(pi);
|
|
break;
|
|
}
|
|
case CHELSIO_GET_UP_LA: {
|
|
struct ch_up_la *la = (struct ch_up_la *)data;
|
|
uint8_t *buf = malloc(LA_BUFSIZE, M_DEVBUF, M_NOWAIT);
|
|
if (buf == NULL) {
|
|
return (ENOMEM);
|
|
}
|
|
if (la->bufsize < LA_BUFSIZE)
|
|
error = ENOBUFS;
|
|
|
|
if (!error)
|
|
error = -t3_get_up_la(sc, &la->stopped, &la->idx,
|
|
&la->bufsize, buf);
|
|
if (!error)
|
|
error = copyout(buf, la->data, la->bufsize);
|
|
|
|
free(buf, M_DEVBUF);
|
|
break;
|
|
}
|
|
case CHELSIO_GET_UP_IOQS: {
|
|
struct ch_up_ioqs *ioqs = (struct ch_up_ioqs *)data;
|
|
uint8_t *buf = malloc(IOQS_BUFSIZE, M_DEVBUF, M_NOWAIT);
|
|
uint32_t *v;
|
|
|
|
if (buf == NULL) {
|
|
return (ENOMEM);
|
|
}
|
|
if (ioqs->bufsize < IOQS_BUFSIZE)
|
|
error = ENOBUFS;
|
|
|
|
if (!error)
|
|
error = -t3_get_up_ioqs(sc, &ioqs->bufsize, buf);
|
|
|
|
if (!error) {
|
|
v = (uint32_t *)buf;
|
|
|
|
ioqs->ioq_rx_enable = *v++;
|
|
ioqs->ioq_tx_enable = *v++;
|
|
ioqs->ioq_rx_status = *v++;
|
|
ioqs->ioq_tx_status = *v++;
|
|
|
|
error = copyout(v, ioqs->data, ioqs->bufsize);
|
|
}
|
|
|
|
free(buf, M_DEVBUF);
|
|
break;
|
|
}
|
|
case CHELSIO_SET_FILTER: {
|
|
struct ch_filter *f = (struct ch_filter *)data;
|
|
struct filter_info *p;
|
|
unsigned int nfilters = sc->params.mc5.nfilters;
|
|
|
|
if (!is_offload(sc))
|
|
return (EOPNOTSUPP); /* No TCAM */
|
|
if (!(sc->flags & FULL_INIT_DONE))
|
|
return (EAGAIN); /* mc5 not setup yet */
|
|
if (nfilters == 0)
|
|
return (EBUSY); /* TOE will use TCAM */
|
|
|
|
/* sanity checks */
|
|
if (f->filter_id >= nfilters ||
|
|
(f->val.dip && f->mask.dip != 0xffffffff) ||
|
|
(f->val.sport && f->mask.sport != 0xffff) ||
|
|
(f->val.dport && f->mask.dport != 0xffff) ||
|
|
(f->val.vlan && f->mask.vlan != 0xfff) ||
|
|
(f->val.vlan_prio &&
|
|
f->mask.vlan_prio != FILTER_NO_VLAN_PRI) ||
|
|
(f->mac_addr_idx != 0xffff && f->mac_addr_idx > 15) ||
|
|
f->qset >= SGE_QSETS ||
|
|
sc->rrss_map[f->qset] >= RSS_TABLE_SIZE)
|
|
return (EINVAL);
|
|
|
|
/* Was allocated with M_WAITOK */
|
|
KASSERT(sc->filters, ("filter table NULL\n"));
|
|
|
|
p = &sc->filters[f->filter_id];
|
|
if (p->locked)
|
|
return (EPERM);
|
|
|
|
bzero(p, sizeof(*p));
|
|
p->sip = f->val.sip;
|
|
p->sip_mask = f->mask.sip;
|
|
p->dip = f->val.dip;
|
|
p->sport = f->val.sport;
|
|
p->dport = f->val.dport;
|
|
p->vlan = f->mask.vlan ? f->val.vlan : 0xfff;
|
|
p->vlan_prio = f->mask.vlan_prio ? (f->val.vlan_prio & 6) :
|
|
FILTER_NO_VLAN_PRI;
|
|
p->mac_hit = f->mac_hit;
|
|
p->mac_vld = f->mac_addr_idx != 0xffff;
|
|
p->mac_idx = f->mac_addr_idx;
|
|
p->pkt_type = f->proto;
|
|
p->report_filter_id = f->want_filter_id;
|
|
p->pass = f->pass;
|
|
p->rss = f->rss;
|
|
p->qset = f->qset;
|
|
|
|
error = set_filter(sc, f->filter_id, p);
|
|
if (error == 0)
|
|
p->valid = 1;
|
|
break;
|
|
}
|
|
case CHELSIO_DEL_FILTER: {
|
|
struct ch_filter *f = (struct ch_filter *)data;
|
|
struct filter_info *p;
|
|
unsigned int nfilters = sc->params.mc5.nfilters;
|
|
|
|
if (!is_offload(sc))
|
|
return (EOPNOTSUPP);
|
|
if (!(sc->flags & FULL_INIT_DONE))
|
|
return (EAGAIN);
|
|
if (nfilters == 0 || sc->filters == NULL)
|
|
return (EINVAL);
|
|
if (f->filter_id >= nfilters)
|
|
return (EINVAL);
|
|
|
|
p = &sc->filters[f->filter_id];
|
|
if (p->locked)
|
|
return (EPERM);
|
|
if (!p->valid)
|
|
return (EFAULT); /* Read "Bad address" as "Bad index" */
|
|
|
|
bzero(p, sizeof(*p));
|
|
p->sip = p->sip_mask = 0xffffffff;
|
|
p->vlan = 0xfff;
|
|
p->vlan_prio = FILTER_NO_VLAN_PRI;
|
|
p->pkt_type = 1;
|
|
error = set_filter(sc, f->filter_id, p);
|
|
break;
|
|
}
|
|
case CHELSIO_GET_FILTER: {
|
|
struct ch_filter *f = (struct ch_filter *)data;
|
|
struct filter_info *p;
|
|
unsigned int i, nfilters = sc->params.mc5.nfilters;
|
|
|
|
if (!is_offload(sc))
|
|
return (EOPNOTSUPP);
|
|
if (!(sc->flags & FULL_INIT_DONE))
|
|
return (EAGAIN);
|
|
if (nfilters == 0 || sc->filters == NULL)
|
|
return (EINVAL);
|
|
|
|
i = f->filter_id == 0xffffffff ? 0 : f->filter_id + 1;
|
|
for (; i < nfilters; i++) {
|
|
p = &sc->filters[i];
|
|
if (!p->valid)
|
|
continue;
|
|
|
|
bzero(f, sizeof(*f));
|
|
|
|
f->filter_id = i;
|
|
f->val.sip = p->sip;
|
|
f->mask.sip = p->sip_mask;
|
|
f->val.dip = p->dip;
|
|
f->mask.dip = p->dip ? 0xffffffff : 0;
|
|
f->val.sport = p->sport;
|
|
f->mask.sport = p->sport ? 0xffff : 0;
|
|
f->val.dport = p->dport;
|
|
f->mask.dport = p->dport ? 0xffff : 0;
|
|
f->val.vlan = p->vlan == 0xfff ? 0 : p->vlan;
|
|
f->mask.vlan = p->vlan == 0xfff ? 0 : 0xfff;
|
|
f->val.vlan_prio = p->vlan_prio == FILTER_NO_VLAN_PRI ?
|
|
0 : p->vlan_prio;
|
|
f->mask.vlan_prio = p->vlan_prio == FILTER_NO_VLAN_PRI ?
|
|
0 : FILTER_NO_VLAN_PRI;
|
|
f->mac_hit = p->mac_hit;
|
|
f->mac_addr_idx = p->mac_vld ? p->mac_idx : 0xffff;
|
|
f->proto = p->pkt_type;
|
|
f->want_filter_id = p->report_filter_id;
|
|
f->pass = p->pass;
|
|
f->rss = p->rss;
|
|
f->qset = p->qset;
|
|
|
|
break;
|
|
}
|
|
|
|
if (i == nfilters)
|
|
f->filter_id = 0xffffffff;
|
|
break;
|
|
}
|
|
default:
|
|
return (EOPNOTSUPP);
|
|
break;
|
|
}
|
|
|
|
return (error);
|
|
}
|
|
|
|
static __inline void
|
|
reg_block_dump(struct adapter *ap, uint8_t *buf, unsigned int start,
|
|
unsigned int end)
|
|
{
|
|
uint32_t *p = (uint32_t *)(buf + start);
|
|
|
|
for ( ; start <= end; start += sizeof(uint32_t))
|
|
*p++ = t3_read_reg(ap, start);
|
|
}
|
|
|
|
#define T3_REGMAP_SIZE (3 * 1024)
|
|
static int
|
|
cxgb_get_regs_len(void)
|
|
{
|
|
return T3_REGMAP_SIZE;
|
|
}
|
|
|
|
static void
|
|
cxgb_get_regs(adapter_t *sc, struct ch_ifconf_regs *regs, uint8_t *buf)
|
|
{
|
|
|
|
/*
|
|
* Version scheme:
|
|
* bits 0..9: chip version
|
|
* bits 10..15: chip revision
|
|
* bit 31: set for PCIe cards
|
|
*/
|
|
regs->version = 3 | (sc->params.rev << 10) | (is_pcie(sc) << 31);
|
|
|
|
/*
|
|
* We skip the MAC statistics registers because they are clear-on-read.
|
|
* Also reading multi-register stats would need to synchronize with the
|
|
* periodic mac stats accumulation. Hard to justify the complexity.
|
|
*/
|
|
memset(buf, 0, cxgb_get_regs_len());
|
|
reg_block_dump(sc, buf, 0, A_SG_RSPQ_CREDIT_RETURN);
|
|
reg_block_dump(sc, buf, A_SG_HI_DRB_HI_THRSH, A_ULPRX_PBL_ULIMIT);
|
|
reg_block_dump(sc, buf, A_ULPTX_CONFIG, A_MPS_INT_CAUSE);
|
|
reg_block_dump(sc, buf, A_CPL_SWITCH_CNTRL, A_CPL_MAP_TBL_DATA);
|
|
reg_block_dump(sc, buf, A_SMB_GLOBAL_TIME_CFG, A_XGM_SERDES_STAT3);
|
|
reg_block_dump(sc, buf, A_XGM_SERDES_STATUS0,
|
|
XGM_REG(A_XGM_SERDES_STAT3, 1));
|
|
reg_block_dump(sc, buf, XGM_REG(A_XGM_SERDES_STATUS0, 1),
|
|
XGM_REG(A_XGM_RX_SPI4_SOP_EOP_CNT, 1));
|
|
}
|
|
|
|
static int
|
|
alloc_filters(struct adapter *sc)
|
|
{
|
|
struct filter_info *p;
|
|
unsigned int nfilters = sc->params.mc5.nfilters;
|
|
|
|
if (nfilters == 0)
|
|
return (0);
|
|
|
|
p = malloc(sizeof(*p) * nfilters, M_DEVBUF, M_WAITOK | M_ZERO);
|
|
sc->filters = p;
|
|
|
|
p = &sc->filters[nfilters - 1];
|
|
p->vlan = 0xfff;
|
|
p->vlan_prio = FILTER_NO_VLAN_PRI;
|
|
p->pass = p->rss = p->valid = p->locked = 1;
|
|
|
|
return (0);
|
|
}
|
|
|
|
static int
|
|
setup_hw_filters(struct adapter *sc)
|
|
{
|
|
int i, rc;
|
|
unsigned int nfilters = sc->params.mc5.nfilters;
|
|
|
|
if (!sc->filters)
|
|
return (0);
|
|
|
|
t3_enable_filters(sc);
|
|
|
|
for (i = rc = 0; i < nfilters && !rc; i++) {
|
|
if (sc->filters[i].locked)
|
|
rc = set_filter(sc, i, &sc->filters[i]);
|
|
}
|
|
|
|
return (rc);
|
|
}
|
|
|
|
static int
|
|
set_filter(struct adapter *sc, int id, const struct filter_info *f)
|
|
{
|
|
int len;
|
|
struct mbuf *m;
|
|
struct ulp_txpkt *txpkt;
|
|
struct work_request_hdr *wr;
|
|
struct cpl_pass_open_req *oreq;
|
|
struct cpl_set_tcb_field *sreq;
|
|
|
|
len = sizeof(*wr) + sizeof(*oreq) + 2 * sizeof(*sreq);
|
|
KASSERT(len <= MHLEN, ("filter request too big for an mbuf"));
|
|
|
|
id += t3_mc5_size(&sc->mc5) - sc->params.mc5.nroutes -
|
|
sc->params.mc5.nfilters;
|
|
|
|
m = m_gethdr(M_WAITOK, MT_DATA);
|
|
m->m_len = m->m_pkthdr.len = len;
|
|
bzero(mtod(m, char *), len);
|
|
|
|
wr = mtod(m, struct work_request_hdr *);
|
|
wr->wrh_hi = htonl(V_WR_OP(FW_WROPCODE_BYPASS) | F_WR_ATOMIC);
|
|
|
|
oreq = (struct cpl_pass_open_req *)(wr + 1);
|
|
txpkt = (struct ulp_txpkt *)oreq;
|
|
txpkt->cmd_dest = htonl(V_ULPTX_CMD(ULP_TXPKT));
|
|
txpkt->len = htonl(V_ULPTX_NFLITS(sizeof(*oreq) / 8));
|
|
OPCODE_TID(oreq) = htonl(MK_OPCODE_TID(CPL_PASS_OPEN_REQ, id));
|
|
oreq->local_port = htons(f->dport);
|
|
oreq->peer_port = htons(f->sport);
|
|
oreq->local_ip = htonl(f->dip);
|
|
oreq->peer_ip = htonl(f->sip);
|
|
oreq->peer_netmask = htonl(f->sip_mask);
|
|
oreq->opt0h = 0;
|
|
oreq->opt0l = htonl(F_NO_OFFLOAD);
|
|
oreq->opt1 = htonl(V_MAC_MATCH_VALID(f->mac_vld) |
|
|
V_CONN_POLICY(CPL_CONN_POLICY_FILTER) |
|
|
V_VLAN_PRI(f->vlan_prio >> 1) |
|
|
V_VLAN_PRI_VALID(f->vlan_prio != FILTER_NO_VLAN_PRI) |
|
|
V_PKT_TYPE(f->pkt_type) | V_OPT1_VLAN(f->vlan) |
|
|
V_MAC_MATCH(f->mac_idx | (f->mac_hit << 4)));
|
|
|
|
sreq = (struct cpl_set_tcb_field *)(oreq + 1);
|
|
set_tcb_field_ulp(sreq, id, 1, 0x1800808000ULL,
|
|
(f->report_filter_id << 15) | (1 << 23) |
|
|
((u64)f->pass << 35) | ((u64)!f->rss << 36));
|
|
set_tcb_field_ulp(sreq + 1, id, 0, 0xffffffff, (2 << 19) | 1);
|
|
t3_mgmt_tx(sc, m);
|
|
|
|
if (f->pass && !f->rss) {
|
|
len = sizeof(*sreq);
|
|
m = m_gethdr(M_WAITOK, MT_DATA);
|
|
m->m_len = m->m_pkthdr.len = len;
|
|
bzero(mtod(m, char *), len);
|
|
sreq = mtod(m, struct cpl_set_tcb_field *);
|
|
sreq->wr.wrh_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD));
|
|
mk_set_tcb_field(sreq, id, 25, 0x3f80000,
|
|
(u64)sc->rrss_map[f->qset] << 19);
|
|
t3_mgmt_tx(sc, m);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static inline void
|
|
mk_set_tcb_field(struct cpl_set_tcb_field *req, unsigned int tid,
|
|
unsigned int word, u64 mask, u64 val)
|
|
{
|
|
OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_SET_TCB_FIELD, tid));
|
|
req->reply = V_NO_REPLY(1);
|
|
req->cpu_idx = 0;
|
|
req->word = htons(word);
|
|
req->mask = htobe64(mask);
|
|
req->val = htobe64(val);
|
|
}
|
|
|
|
static inline void
|
|
set_tcb_field_ulp(struct cpl_set_tcb_field *req, unsigned int tid,
|
|
unsigned int word, u64 mask, u64 val)
|
|
{
|
|
struct ulp_txpkt *txpkt = (struct ulp_txpkt *)req;
|
|
|
|
txpkt->cmd_dest = htonl(V_ULPTX_CMD(ULP_TXPKT));
|
|
txpkt->len = htonl(V_ULPTX_NFLITS(sizeof(*req) / 8));
|
|
mk_set_tcb_field(req, tid, word, mask, val);
|
|
}
|
|
|
|
void
|
|
t3_iterate(void (*func)(struct adapter *, void *), void *arg)
|
|
{
|
|
struct adapter *sc;
|
|
|
|
mtx_lock(&t3_list_lock);
|
|
SLIST_FOREACH(sc, &t3_list, link) {
|
|
/*
|
|
* func should not make any assumptions about what state sc is
|
|
* in - the only guarantee is that sc->sc_lock is a valid lock.
|
|
*/
|
|
func(sc, arg);
|
|
}
|
|
mtx_unlock(&t3_list_lock);
|
|
}
|
|
|
|
#ifdef TCP_OFFLOAD
|
|
static int
|
|
toe_capability(struct port_info *pi, int enable)
|
|
{
|
|
int rc;
|
|
struct adapter *sc = pi->adapter;
|
|
|
|
ADAPTER_LOCK_ASSERT_OWNED(sc);
|
|
|
|
if (!is_offload(sc))
|
|
return (ENODEV);
|
|
|
|
if (enable) {
|
|
if (!(sc->flags & FULL_INIT_DONE)) {
|
|
log(LOG_WARNING,
|
|
"You must enable a cxgb interface first\n");
|
|
return (EAGAIN);
|
|
}
|
|
|
|
if (isset(&sc->offload_map, pi->port_id))
|
|
return (0);
|
|
|
|
if (!(sc->flags & TOM_INIT_DONE)) {
|
|
rc = t3_activate_uld(sc, ULD_TOM);
|
|
if (rc == EAGAIN) {
|
|
log(LOG_WARNING,
|
|
"You must kldload t3_tom.ko before trying "
|
|
"to enable TOE on a cxgb interface.\n");
|
|
}
|
|
if (rc != 0)
|
|
return (rc);
|
|
KASSERT(sc->tom_softc != NULL,
|
|
("%s: TOM activated but softc NULL", __func__));
|
|
KASSERT(sc->flags & TOM_INIT_DONE,
|
|
("%s: TOM activated but flag not set", __func__));
|
|
}
|
|
|
|
setbit(&sc->offload_map, pi->port_id);
|
|
|
|
/*
|
|
* XXX: Temporary code to allow iWARP to be enabled when TOE is
|
|
* enabled on any port. Need to figure out how to enable,
|
|
* disable, load, and unload iWARP cleanly.
|
|
*/
|
|
if (!isset(&sc->offload_map, MAX_NPORTS) &&
|
|
t3_activate_uld(sc, ULD_IWARP) == 0)
|
|
setbit(&sc->offload_map, MAX_NPORTS);
|
|
} else {
|
|
if (!isset(&sc->offload_map, pi->port_id))
|
|
return (0);
|
|
|
|
KASSERT(sc->flags & TOM_INIT_DONE,
|
|
("%s: TOM never initialized?", __func__));
|
|
clrbit(&sc->offload_map, pi->port_id);
|
|
}
|
|
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Add an upper layer driver to the global list.
|
|
*/
|
|
int
|
|
t3_register_uld(struct uld_info *ui)
|
|
{
|
|
int rc = 0;
|
|
struct uld_info *u;
|
|
|
|
mtx_lock(&t3_uld_list_lock);
|
|
SLIST_FOREACH(u, &t3_uld_list, link) {
|
|
if (u->uld_id == ui->uld_id) {
|
|
rc = EEXIST;
|
|
goto done;
|
|
}
|
|
}
|
|
|
|
SLIST_INSERT_HEAD(&t3_uld_list, ui, link);
|
|
ui->refcount = 0;
|
|
done:
|
|
mtx_unlock(&t3_uld_list_lock);
|
|
return (rc);
|
|
}
|
|
|
|
int
|
|
t3_unregister_uld(struct uld_info *ui)
|
|
{
|
|
int rc = EINVAL;
|
|
struct uld_info *u;
|
|
|
|
mtx_lock(&t3_uld_list_lock);
|
|
|
|
SLIST_FOREACH(u, &t3_uld_list, link) {
|
|
if (u == ui) {
|
|
if (ui->refcount > 0) {
|
|
rc = EBUSY;
|
|
goto done;
|
|
}
|
|
|
|
SLIST_REMOVE(&t3_uld_list, ui, uld_info, link);
|
|
rc = 0;
|
|
goto done;
|
|
}
|
|
}
|
|
done:
|
|
mtx_unlock(&t3_uld_list_lock);
|
|
return (rc);
|
|
}
|
|
|
|
int
|
|
t3_activate_uld(struct adapter *sc, int id)
|
|
{
|
|
int rc = EAGAIN;
|
|
struct uld_info *ui;
|
|
|
|
mtx_lock(&t3_uld_list_lock);
|
|
|
|
SLIST_FOREACH(ui, &t3_uld_list, link) {
|
|
if (ui->uld_id == id) {
|
|
rc = ui->activate(sc);
|
|
if (rc == 0)
|
|
ui->refcount++;
|
|
goto done;
|
|
}
|
|
}
|
|
done:
|
|
mtx_unlock(&t3_uld_list_lock);
|
|
|
|
return (rc);
|
|
}
|
|
|
|
int
|
|
t3_deactivate_uld(struct adapter *sc, int id)
|
|
{
|
|
int rc = EINVAL;
|
|
struct uld_info *ui;
|
|
|
|
mtx_lock(&t3_uld_list_lock);
|
|
|
|
SLIST_FOREACH(ui, &t3_uld_list, link) {
|
|
if (ui->uld_id == id) {
|
|
rc = ui->deactivate(sc);
|
|
if (rc == 0)
|
|
ui->refcount--;
|
|
goto done;
|
|
}
|
|
}
|
|
done:
|
|
mtx_unlock(&t3_uld_list_lock);
|
|
|
|
return (rc);
|
|
}
|
|
|
|
static int
|
|
cpl_not_handled(struct sge_qset *qs __unused, struct rsp_desc *r __unused,
|
|
struct mbuf *m)
|
|
{
|
|
m_freem(m);
|
|
return (EDOOFUS);
|
|
}
|
|
|
|
int
|
|
t3_register_cpl_handler(struct adapter *sc, int opcode, cpl_handler_t h)
|
|
{
|
|
uintptr_t *loc, new;
|
|
|
|
if (opcode >= NUM_CPL_HANDLERS)
|
|
return (EINVAL);
|
|
|
|
new = h ? (uintptr_t)h : (uintptr_t)cpl_not_handled;
|
|
loc = (uintptr_t *) &sc->cpl_handler[opcode];
|
|
atomic_store_rel_ptr(loc, new);
|
|
|
|
return (0);
|
|
}
|
|
#endif
|
|
|
|
static int
|
|
cxgbc_mod_event(module_t mod, int cmd, void *arg)
|
|
{
|
|
int rc = 0;
|
|
|
|
switch (cmd) {
|
|
case MOD_LOAD:
|
|
mtx_init(&t3_list_lock, "T3 adapters", 0, MTX_DEF);
|
|
SLIST_INIT(&t3_list);
|
|
#ifdef TCP_OFFLOAD
|
|
mtx_init(&t3_uld_list_lock, "T3 ULDs", 0, MTX_DEF);
|
|
SLIST_INIT(&t3_uld_list);
|
|
#endif
|
|
break;
|
|
|
|
case MOD_UNLOAD:
|
|
#ifdef TCP_OFFLOAD
|
|
mtx_lock(&t3_uld_list_lock);
|
|
if (!SLIST_EMPTY(&t3_uld_list)) {
|
|
rc = EBUSY;
|
|
mtx_unlock(&t3_uld_list_lock);
|
|
break;
|
|
}
|
|
mtx_unlock(&t3_uld_list_lock);
|
|
mtx_destroy(&t3_uld_list_lock);
|
|
#endif
|
|
mtx_lock(&t3_list_lock);
|
|
if (!SLIST_EMPTY(&t3_list)) {
|
|
rc = EBUSY;
|
|
mtx_unlock(&t3_list_lock);
|
|
break;
|
|
}
|
|
mtx_unlock(&t3_list_lock);
|
|
mtx_destroy(&t3_list_lock);
|
|
break;
|
|
}
|
|
|
|
return (rc);
|
|
}
|