freebsd-skq/sys/dev/rt/if_rt.c
Pedro F. Giffuni 718cf2ccb9 sys/dev: further adoption of SPDX licensing ID tags.
Mainly focus on files that use BSD 2-Clause license, however the tool I
was using misidentified many licenses so this was mostly a manual - error
prone - task.

The Software Package Data Exchange (SPDX) group provides a specification
to make it easier for automated tools to detect and summarize well known
opensource licenses. We are gradually adopting the specification, noting
that the tags are considered only advisory and do not, in any way,
superceed or replace the license texts.
2017-11-27 14:52:40 +00:00

2951 lines
72 KiB
C

/*-
* SPDX-License-Identifier: BSD-2-Clause-FreeBSD
*
* Copyright (c) 2015-2016, Stanislav Galabov
* Copyright (c) 2014, Aleksandr A. Mityaev
* Copyright (c) 2011, Aleksandr Rybalko
* based on hard work
* by Alexander Egorenkov <egorenar@gmail.com>
* and by Damien Bergamini <damien.bergamini@free.fr>
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice unmodified, this list of conditions, and the following
* disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*/
#include <sys/cdefs.h>
__FBSDID("$FreeBSD$");
#include "if_rtvar.h"
#include "if_rtreg.h"
#include <net/if.h>
#include <net/if_var.h>
#include <net/if_arp.h>
#include <net/ethernet.h>
#include <net/if_dl.h>
#include <net/if_media.h>
#include <net/if_types.h>
#include <net/if_vlan_var.h>
#include <net/bpf.h>
#include <machine/bus.h>
#include <machine/cache.h>
#include <machine/cpufunc.h>
#include <machine/resource.h>
#include <vm/vm_param.h>
#include <vm/vm.h>
#include <vm/pmap.h>
#include <machine/pmap.h>
#include <sys/bus.h>
#include <sys/rman.h>
#include "opt_platform.h"
#include "opt_rt305x.h"
#ifdef FDT
#include <dev/ofw/openfirm.h>
#include <dev/ofw/ofw_bus.h>
#include <dev/ofw/ofw_bus_subr.h>
#endif
#include <dev/mii/mii.h>
#include <dev/mii/miivar.h>
#ifdef RT_MDIO
#include <dev/mdio/mdio.h>
#include <dev/etherswitch/miiproxy.h>
#include "mdio_if.h"
#endif
#if 0
#include <mips/rt305x/rt305x_sysctlvar.h>
#include <mips/rt305x/rt305xreg.h>
#endif
#ifdef IF_RT_PHY_SUPPORT
#include "miibus_if.h"
#endif
/*
* Defines and macros
*/
#define RT_MAX_AGG_SIZE 3840
#define RT_TX_DATA_SEG0_SIZE MJUMPAGESIZE
#define RT_MS(_v, _f) (((_v) & _f) >> _f##_S)
#define RT_SM(_v, _f) (((_v) << _f##_S) & _f)
#define RT_TX_WATCHDOG_TIMEOUT 5
#define RT_CHIPID_RT2880 0x2880
#define RT_CHIPID_RT3050 0x3050
#define RT_CHIPID_RT5350 0x5350
#define RT_CHIPID_MT7620 0x7620
#define RT_CHIPID_MT7621 0x7621
#ifdef FDT
/* more specific and new models should go first */
static const struct ofw_compat_data rt_compat_data[] = {
{ "ralink,rt2880-eth", RT_CHIPID_RT2880 },
{ "ralink,rt3050-eth", RT_CHIPID_RT3050 },
{ "ralink,rt3352-eth", RT_CHIPID_RT3050 },
{ "ralink,rt3883-eth", RT_CHIPID_RT3050 },
{ "ralink,rt5350-eth", RT_CHIPID_RT5350 },
{ "ralink,mt7620a-eth", RT_CHIPID_MT7620 },
{ "mediatek,mt7620-eth", RT_CHIPID_MT7620 },
{ "ralink,mt7621-eth", RT_CHIPID_MT7621 },
{ "mediatek,mt7621-eth", RT_CHIPID_MT7621 },
{ NULL, 0 }
};
#endif
/*
* Static function prototypes
*/
static int rt_probe(device_t dev);
static int rt_attach(device_t dev);
static int rt_detach(device_t dev);
static int rt_shutdown(device_t dev);
static int rt_suspend(device_t dev);
static int rt_resume(device_t dev);
static void rt_init_locked(void *priv);
static void rt_init(void *priv);
static void rt_stop_locked(void *priv);
static void rt_stop(void *priv);
static void rt_start(struct ifnet *ifp);
static int rt_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data);
static void rt_periodic(void *arg);
static void rt_tx_watchdog(void *arg);
static void rt_intr(void *arg);
static void rt_rt5350_intr(void *arg);
static void rt_tx_coherent_intr(struct rt_softc *sc);
static void rt_rx_coherent_intr(struct rt_softc *sc);
static void rt_rx_delay_intr(struct rt_softc *sc);
static void rt_tx_delay_intr(struct rt_softc *sc);
static void rt_rx_intr(struct rt_softc *sc, int qid);
static void rt_tx_intr(struct rt_softc *sc, int qid);
static void rt_rx_done_task(void *context, int pending);
static void rt_tx_done_task(void *context, int pending);
static void rt_periodic_task(void *context, int pending);
static int rt_rx_eof(struct rt_softc *sc,
struct rt_softc_rx_ring *ring, int limit);
static void rt_tx_eof(struct rt_softc *sc,
struct rt_softc_tx_ring *ring);
static void rt_update_stats(struct rt_softc *sc);
static void rt_watchdog(struct rt_softc *sc);
static void rt_update_raw_counters(struct rt_softc *sc);
static void rt_intr_enable(struct rt_softc *sc, uint32_t intr_mask);
static void rt_intr_disable(struct rt_softc *sc, uint32_t intr_mask);
static int rt_txrx_enable(struct rt_softc *sc);
static int rt_alloc_rx_ring(struct rt_softc *sc,
struct rt_softc_rx_ring *ring, int qid);
static void rt_reset_rx_ring(struct rt_softc *sc,
struct rt_softc_rx_ring *ring);
static void rt_free_rx_ring(struct rt_softc *sc,
struct rt_softc_rx_ring *ring);
static int rt_alloc_tx_ring(struct rt_softc *sc,
struct rt_softc_tx_ring *ring, int qid);
static void rt_reset_tx_ring(struct rt_softc *sc,
struct rt_softc_tx_ring *ring);
static void rt_free_tx_ring(struct rt_softc *sc,
struct rt_softc_tx_ring *ring);
static void rt_dma_map_addr(void *arg, bus_dma_segment_t *segs,
int nseg, int error);
static void rt_sysctl_attach(struct rt_softc *sc);
#ifdef IF_RT_PHY_SUPPORT
void rt_miibus_statchg(device_t);
#endif
#if defined(IF_RT_PHY_SUPPORT) || defined(RT_MDIO)
static int rt_miibus_readreg(device_t, int, int);
static int rt_miibus_writereg(device_t, int, int, int);
#endif
static int rt_ifmedia_upd(struct ifnet *);
static void rt_ifmedia_sts(struct ifnet *, struct ifmediareq *);
static SYSCTL_NODE(_hw, OID_AUTO, rt, CTLFLAG_RD, 0, "RT driver parameters");
#ifdef IF_RT_DEBUG
static int rt_debug = 0;
SYSCTL_INT(_hw_rt, OID_AUTO, debug, CTLFLAG_RWTUN, &rt_debug, 0,
"RT debug level");
#endif
static int
rt_probe(device_t dev)
{
struct rt_softc *sc = device_get_softc(dev);
char buf[80];
#ifdef FDT
const struct ofw_compat_data * cd;
cd = ofw_bus_search_compatible(dev, rt_compat_data);
if (cd->ocd_data == 0)
return (ENXIO);
sc->rt_chipid = (unsigned int)(cd->ocd_data);
#else
#if defined(MT7620)
sc->rt_chipid = RT_CHIPID_MT7620;
#elif defined(MT7621)
sc->rt_chipid = RT_CHIPID_MT7621;
#elif defined(RT5350)
sc->rt_chipid = RT_CHIPID_RT5350;
#else
sc->rt_chipid = RT_CHIPID_RT3050;
#endif
#endif
snprintf(buf, sizeof(buf), "Ralink %cT%x onChip Ethernet driver",
sc->rt_chipid >= 0x7600 ? 'M' : 'R', sc->rt_chipid);
device_set_desc_copy(dev, buf);
return (BUS_PROBE_GENERIC);
}
/*
* macaddr_atoi - translate string MAC address to uint8_t array
*/
static int
macaddr_atoi(const char *str, uint8_t *mac)
{
int count, i;
unsigned int amac[ETHER_ADDR_LEN]; /* Aligned version */
count = sscanf(str, "%x%*c%x%*c%x%*c%x%*c%x%*c%x",
&amac[0], &amac[1], &amac[2],
&amac[3], &amac[4], &amac[5]);
if (count < ETHER_ADDR_LEN) {
memset(mac, 0, ETHER_ADDR_LEN);
return (1);
}
/* Copy aligned to result */
for (i = 0; i < ETHER_ADDR_LEN; i ++)
mac[i] = (amac[i] & 0xff);
return (0);
}
#ifdef USE_GENERATED_MAC_ADDRESS
/*
* generate_mac(uin8_t *mac)
* This is MAC address generator for cases when real device MAC address
* unknown or not yet accessible.
* Use 'b','s','d' signature and 3 octets from CRC32 on kenv.
* MAC = 'b', 's', 'd', CRC[3]^CRC[2], CRC[1], CRC[0]
*
* Output - MAC address, that do not change between reboots, if hints or
* bootloader info unchange.
*/
static void
generate_mac(uint8_t *mac)
{
unsigned char *cp;
int i = 0;
uint32_t crc = 0xffffffff;
/* Generate CRC32 on kenv */
for (cp = kenvp[0]; cp != NULL; cp = kenvp[++i]) {
crc = calculate_crc32c(crc, cp, strlen(cp) + 1);
}
crc = ~crc;
mac[0] = 'b';
mac[1] = 's';
mac[2] = 'd';
mac[3] = (crc >> 24) ^ ((crc >> 16) & 0xff);
mac[4] = (crc >> 8) & 0xff;
mac[5] = crc & 0xff;
}
#endif
/*
* ether_request_mac - try to find usable MAC address.
*/
static int
ether_request_mac(device_t dev, uint8_t *mac)
{
char *var;
/*
* "ethaddr" is passed via envp on RedBoot platforms
* "kmac" is passed via argv on RouterBOOT platforms
*/
#if defined(RT305X_UBOOT) || defined(__REDBOOT__) || defined(__ROUTERBOOT__)
if ((var = kern_getenv("ethaddr")) != NULL ||
(var = kern_getenv("kmac")) != NULL ) {
if(!macaddr_atoi(var, mac)) {
printf("%s: use %s macaddr from KENV\n",
device_get_nameunit(dev), var);
freeenv(var);
return (0);
}
freeenv(var);
}
#endif
/*
* Try from hints
* hint.[dev].[unit].macaddr
*/
if (!resource_string_value(device_get_name(dev),
device_get_unit(dev), "macaddr", (const char **)&var)) {
if(!macaddr_atoi(var, mac)) {
printf("%s: use %s macaddr from hints\n",
device_get_nameunit(dev), var);
return (0);
}
}
#ifdef USE_GENERATED_MAC_ADDRESS
generate_mac(mac);
device_printf(dev, "use generated %02x:%02x:%02x:%02x:%02x:%02x "
"macaddr\n", mac[0], mac[1], mac[2], mac[3], mac[4], mac[5]);
#else
/* Hardcoded */
mac[0] = 0x00;
mac[1] = 0x18;
mac[2] = 0xe7;
mac[3] = 0xd5;
mac[4] = 0x83;
mac[5] = 0x90;
device_printf(dev, "use hardcoded 00:18:e7:d5:83:90 macaddr\n");
#endif
return (0);
}
/*
* Reset hardware
*/
static void
reset_freng(struct rt_softc *sc)
{
/* XXX hard reset kills everything so skip it ... */
return;
}
static int
rt_attach(device_t dev)
{
struct rt_softc *sc;
struct ifnet *ifp;
int error, i;
sc = device_get_softc(dev);
sc->dev = dev;
mtx_init(&sc->lock, device_get_nameunit(dev), MTX_NETWORK_LOCK,
MTX_DEF | MTX_RECURSE);
sc->mem_rid = 0;
sc->mem = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &sc->mem_rid,
RF_ACTIVE | RF_SHAREABLE);
if (sc->mem == NULL) {
device_printf(dev, "could not allocate memory resource\n");
error = ENXIO;
goto fail;
}
sc->bst = rman_get_bustag(sc->mem);
sc->bsh = rman_get_bushandle(sc->mem);
sc->irq_rid = 0;
sc->irq = bus_alloc_resource_any(dev, SYS_RES_IRQ, &sc->irq_rid,
RF_ACTIVE);
if (sc->irq == NULL) {
device_printf(dev,
"could not allocate interrupt resource\n");
error = ENXIO;
goto fail;
}
#ifdef IF_RT_DEBUG
sc->debug = rt_debug;
SYSCTL_ADD_INT(device_get_sysctl_ctx(dev),
SYSCTL_CHILDREN(device_get_sysctl_tree(dev)), OID_AUTO,
"debug", CTLFLAG_RW, &sc->debug, 0, "rt debug level");
#endif
/* Reset hardware */
reset_freng(sc);
if (sc->rt_chipid == RT_CHIPID_MT7620) {
sc->csum_fail_ip = MT7620_RXD_SRC_IP_CSUM_FAIL;
sc->csum_fail_l4 = MT7620_RXD_SRC_L4_CSUM_FAIL;
} else if (sc->rt_chipid == RT_CHIPID_MT7621) {
sc->csum_fail_ip = MT7621_RXD_SRC_IP_CSUM_FAIL;
sc->csum_fail_l4 = MT7621_RXD_SRC_L4_CSUM_FAIL;
} else {
sc->csum_fail_ip = RT305X_RXD_SRC_IP_CSUM_FAIL;
sc->csum_fail_l4 = RT305X_RXD_SRC_L4_CSUM_FAIL;
}
/* Fill in soc-specific registers map */
switch(sc->rt_chipid) {
case RT_CHIPID_MT7620:
case RT_CHIPID_MT7621:
sc->gdma1_base = MT7620_GDMA1_BASE;
/* fallthrough */
case RT_CHIPID_RT5350:
device_printf(dev, "%cT%x Ethernet MAC (rev 0x%08x)\n",
sc->rt_chipid >= 0x7600 ? 'M' : 'R',
sc->rt_chipid, sc->mac_rev);
/* RT5350: No GDMA, PSE, CDMA, PPE */
RT_WRITE(sc, GE_PORT_BASE + 0x0C00, // UDPCS, TCPCS, IPCS=1
RT_READ(sc, GE_PORT_BASE + 0x0C00) | (0x7<<16));
sc->delay_int_cfg=RT5350_PDMA_BASE+RT5350_DELAY_INT_CFG;
sc->fe_int_status=RT5350_FE_INT_STATUS;
sc->fe_int_enable=RT5350_FE_INT_ENABLE;
sc->pdma_glo_cfg=RT5350_PDMA_BASE+RT5350_PDMA_GLO_CFG;
sc->pdma_rst_idx=RT5350_PDMA_BASE+RT5350_PDMA_RST_IDX;
for (i = 0; i < RT_SOFTC_TX_RING_COUNT; i++) {
sc->tx_base_ptr[i]=RT5350_PDMA_BASE+RT5350_TX_BASE_PTR(i);
sc->tx_max_cnt[i]=RT5350_PDMA_BASE+RT5350_TX_MAX_CNT(i);
sc->tx_ctx_idx[i]=RT5350_PDMA_BASE+RT5350_TX_CTX_IDX(i);
sc->tx_dtx_idx[i]=RT5350_PDMA_BASE+RT5350_TX_DTX_IDX(i);
}
sc->rx_ring_count=2;
sc->rx_base_ptr[0]=RT5350_PDMA_BASE+RT5350_RX_BASE_PTR0;
sc->rx_max_cnt[0]=RT5350_PDMA_BASE+RT5350_RX_MAX_CNT0;
sc->rx_calc_idx[0]=RT5350_PDMA_BASE+RT5350_RX_CALC_IDX0;
sc->rx_drx_idx[0]=RT5350_PDMA_BASE+RT5350_RX_DRX_IDX0;
sc->rx_base_ptr[1]=RT5350_PDMA_BASE+RT5350_RX_BASE_PTR1;
sc->rx_max_cnt[1]=RT5350_PDMA_BASE+RT5350_RX_MAX_CNT1;
sc->rx_calc_idx[1]=RT5350_PDMA_BASE+RT5350_RX_CALC_IDX1;
sc->rx_drx_idx[1]=RT5350_PDMA_BASE+RT5350_RX_DRX_IDX1;
sc->int_rx_done_mask=RT5350_INT_RXQ0_DONE;
sc->int_tx_done_mask=RT5350_INT_TXQ0_DONE;
break;
default:
device_printf(dev, "RT305XF Ethernet MAC (rev 0x%08x)\n",
sc->mac_rev);
sc->gdma1_base = GDMA1_BASE;
sc->delay_int_cfg=PDMA_BASE+DELAY_INT_CFG;
sc->fe_int_status=GE_PORT_BASE+FE_INT_STATUS;
sc->fe_int_enable=GE_PORT_BASE+FE_INT_ENABLE;
sc->pdma_glo_cfg=PDMA_BASE+PDMA_GLO_CFG;
sc->pdma_rst_idx=PDMA_BASE+PDMA_RST_IDX;
for (i = 0; i < RT_SOFTC_TX_RING_COUNT; i++) {
sc->tx_base_ptr[i]=PDMA_BASE+TX_BASE_PTR(i);
sc->tx_max_cnt[i]=PDMA_BASE+TX_MAX_CNT(i);
sc->tx_ctx_idx[i]=PDMA_BASE+TX_CTX_IDX(i);
sc->tx_dtx_idx[i]=PDMA_BASE+TX_DTX_IDX(i);
}
sc->rx_ring_count=1;
sc->rx_base_ptr[0]=PDMA_BASE+RX_BASE_PTR0;
sc->rx_max_cnt[0]=PDMA_BASE+RX_MAX_CNT0;
sc->rx_calc_idx[0]=PDMA_BASE+RX_CALC_IDX0;
sc->rx_drx_idx[0]=PDMA_BASE+RX_DRX_IDX0;
sc->int_rx_done_mask=INT_RX_DONE;
sc->int_tx_done_mask=INT_TXQ0_DONE;
}
if (sc->gdma1_base != 0)
RT_WRITE(sc, sc->gdma1_base + GDMA_FWD_CFG,
(
GDM_ICS_EN | /* Enable IP Csum */
GDM_TCS_EN | /* Enable TCP Csum */
GDM_UCS_EN | /* Enable UDP Csum */
GDM_STRPCRC | /* Strip CRC from packet */
GDM_DST_PORT_CPU << GDM_UFRC_P_SHIFT | /* fwd UCast to CPU */
GDM_DST_PORT_CPU << GDM_BFRC_P_SHIFT | /* fwd BCast to CPU */
GDM_DST_PORT_CPU << GDM_MFRC_P_SHIFT | /* fwd MCast to CPU */
GDM_DST_PORT_CPU << GDM_OFRC_P_SHIFT /* fwd Other to CPU */
));
if (sc->rt_chipid == RT_CHIPID_RT2880)
RT_WRITE(sc, MDIO_CFG, MDIO_2880_100T_INIT);
/* allocate Tx and Rx rings */
for (i = 0; i < RT_SOFTC_TX_RING_COUNT; i++) {
error = rt_alloc_tx_ring(sc, &sc->tx_ring[i], i);
if (error != 0) {
device_printf(dev, "could not allocate Tx ring #%d\n",
i);
goto fail;
}
}
sc->tx_ring_mgtqid = 5;
for (i = 0; i < sc->rx_ring_count; i++) {
error = rt_alloc_rx_ring(sc, &sc->rx_ring[i], i);
if (error != 0) {
device_printf(dev, "could not allocate Rx ring\n");
goto fail;
}
}
callout_init(&sc->periodic_ch, 0);
callout_init_mtx(&sc->tx_watchdog_ch, &sc->lock, 0);
ifp = sc->ifp = if_alloc(IFT_ETHER);
if (ifp == NULL) {
device_printf(dev, "could not if_alloc()\n");
error = ENOMEM;
goto fail;
}
ifp->if_softc = sc;
if_initname(ifp, device_get_name(sc->dev), device_get_unit(sc->dev));
ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
ifp->if_init = rt_init;
ifp->if_ioctl = rt_ioctl;
ifp->if_start = rt_start;
#define RT_TX_QLEN 256
IFQ_SET_MAXLEN(&ifp->if_snd, RT_TX_QLEN);
ifp->if_snd.ifq_drv_maxlen = RT_TX_QLEN;
IFQ_SET_READY(&ifp->if_snd);
#ifdef IF_RT_PHY_SUPPORT
error = mii_attach(dev, &sc->rt_miibus, ifp, rt_ifmedia_upd,
rt_ifmedia_sts, BMSR_DEFCAPMASK, MII_PHY_ANY, MII_OFFSET_ANY, 0);
if (error != 0) {
device_printf(dev, "attaching PHYs failed\n");
error = ENXIO;
goto fail;
}
#else
ifmedia_init(&sc->rt_ifmedia, 0, rt_ifmedia_upd, rt_ifmedia_sts);
ifmedia_add(&sc->rt_ifmedia, IFM_ETHER | IFM_100_TX | IFM_FDX, 0,
NULL);
ifmedia_set(&sc->rt_ifmedia, IFM_ETHER | IFM_100_TX | IFM_FDX);
#endif /* IF_RT_PHY_SUPPORT */
ether_request_mac(dev, sc->mac_addr);
ether_ifattach(ifp, sc->mac_addr);
/*
* Tell the upper layer(s) we support long frames.
*/
ifp->if_hdrlen = sizeof(struct ether_vlan_header);
ifp->if_capabilities |= IFCAP_VLAN_MTU;
ifp->if_capenable |= IFCAP_VLAN_MTU;
ifp->if_capabilities |= IFCAP_RXCSUM|IFCAP_TXCSUM;
ifp->if_capenable |= IFCAP_RXCSUM|IFCAP_TXCSUM;
/* init task queue */
TASK_INIT(&sc->rx_done_task, 0, rt_rx_done_task, sc);
TASK_INIT(&sc->tx_done_task, 0, rt_tx_done_task, sc);
TASK_INIT(&sc->periodic_task, 0, rt_periodic_task, sc);
sc->rx_process_limit = 100;
sc->taskqueue = taskqueue_create("rt_taskq", M_NOWAIT,
taskqueue_thread_enqueue, &sc->taskqueue);
taskqueue_start_threads(&sc->taskqueue, 1, PI_NET, "%s taskq",
device_get_nameunit(sc->dev));
rt_sysctl_attach(sc);
/* set up interrupt */
error = bus_setup_intr(dev, sc->irq, INTR_TYPE_NET | INTR_MPSAFE,
NULL, (sc->rt_chipid == RT_CHIPID_RT5350 ||
sc->rt_chipid == RT_CHIPID_MT7620 ||
sc->rt_chipid == RT_CHIPID_MT7621) ? rt_rt5350_intr : rt_intr,
sc, &sc->irqh);
if (error != 0) {
printf("%s: could not set up interrupt\n",
device_get_nameunit(dev));
goto fail;
}
#ifdef IF_RT_DEBUG
device_printf(dev, "debug var at %#08x\n", (u_int)&(sc->debug));
#endif
return (0);
fail:
/* free Tx and Rx rings */
for (i = 0; i < RT_SOFTC_TX_RING_COUNT; i++)
rt_free_tx_ring(sc, &sc->tx_ring[i]);
for (i = 0; i < sc->rx_ring_count; i++)
rt_free_rx_ring(sc, &sc->rx_ring[i]);
mtx_destroy(&sc->lock);
if (sc->mem != NULL)
bus_release_resource(dev, SYS_RES_MEMORY, sc->mem_rid,
sc->mem);
if (sc->irq != NULL)
bus_release_resource(dev, SYS_RES_IRQ, sc->irq_rid,
sc->irq);
return (error);
}
/*
* Set media options.
*/
static int
rt_ifmedia_upd(struct ifnet *ifp)
{
struct rt_softc *sc;
#ifdef IF_RT_PHY_SUPPORT
struct mii_data *mii;
struct mii_softc *miisc;
int error = 0;
sc = ifp->if_softc;
RT_SOFTC_LOCK(sc);
mii = device_get_softc(sc->rt_miibus);
LIST_FOREACH(miisc, &mii->mii_phys, mii_list)
PHY_RESET(miisc);
error = mii_mediachg(mii);
RT_SOFTC_UNLOCK(sc);
return (error);
#else /* !IF_RT_PHY_SUPPORT */
struct ifmedia *ifm;
struct ifmedia_entry *ife;
sc = ifp->if_softc;
ifm = &sc->rt_ifmedia;
ife = ifm->ifm_cur;
if (IFM_TYPE(ifm->ifm_media) != IFM_ETHER)
return (EINVAL);
if (IFM_SUBTYPE(ife->ifm_media) == IFM_AUTO) {
device_printf(sc->dev,
"AUTO is not supported for multiphy MAC");
return (EINVAL);
}
/*
* Ignore everything
*/
return (0);
#endif /* IF_RT_PHY_SUPPORT */
}
/*
* Report current media status.
*/
static void
rt_ifmedia_sts(struct ifnet *ifp, struct ifmediareq *ifmr)
{
#ifdef IF_RT_PHY_SUPPORT
struct rt_softc *sc;
struct mii_data *mii;
sc = ifp->if_softc;
RT_SOFTC_LOCK(sc);
mii = device_get_softc(sc->rt_miibus);
mii_pollstat(mii);
ifmr->ifm_active = mii->mii_media_active;
ifmr->ifm_status = mii->mii_media_status;
ifmr->ifm_active = IFM_ETHER | IFM_100_TX | IFM_FDX;
ifmr->ifm_status = IFM_AVALID | IFM_ACTIVE;
RT_SOFTC_UNLOCK(sc);
#else /* !IF_RT_PHY_SUPPORT */
ifmr->ifm_status = IFM_AVALID | IFM_ACTIVE;
ifmr->ifm_active = IFM_ETHER | IFM_100_TX | IFM_FDX;
#endif /* IF_RT_PHY_SUPPORT */
}
static int
rt_detach(device_t dev)
{
struct rt_softc *sc;
struct ifnet *ifp;
int i;
sc = device_get_softc(dev);
ifp = sc->ifp;
RT_DPRINTF(sc, RT_DEBUG_ANY, "detaching\n");
RT_SOFTC_LOCK(sc);
ifp->if_drv_flags &= ~(IFF_DRV_RUNNING | IFF_DRV_OACTIVE);
callout_stop(&sc->periodic_ch);
callout_stop(&sc->tx_watchdog_ch);
taskqueue_drain(sc->taskqueue, &sc->rx_done_task);
taskqueue_drain(sc->taskqueue, &sc->tx_done_task);
taskqueue_drain(sc->taskqueue, &sc->periodic_task);
/* free Tx and Rx rings */
for (i = 0; i < RT_SOFTC_TX_RING_COUNT; i++)
rt_free_tx_ring(sc, &sc->tx_ring[i]);
for (i = 0; i < sc->rx_ring_count; i++)
rt_free_rx_ring(sc, &sc->rx_ring[i]);
RT_SOFTC_UNLOCK(sc);
#ifdef IF_RT_PHY_SUPPORT
if (sc->rt_miibus != NULL)
device_delete_child(dev, sc->rt_miibus);
#endif
ether_ifdetach(ifp);
if_free(ifp);
taskqueue_free(sc->taskqueue);
mtx_destroy(&sc->lock);
bus_generic_detach(dev);
bus_teardown_intr(dev, sc->irq, sc->irqh);
bus_release_resource(dev, SYS_RES_IRQ, sc->irq_rid, sc->irq);
bus_release_resource(dev, SYS_RES_MEMORY, sc->mem_rid, sc->mem);
return (0);
}
static int
rt_shutdown(device_t dev)
{
struct rt_softc *sc;
sc = device_get_softc(dev);
RT_DPRINTF(sc, RT_DEBUG_ANY, "shutting down\n");
rt_stop(sc);
return (0);
}
static int
rt_suspend(device_t dev)
{
struct rt_softc *sc;
sc = device_get_softc(dev);
RT_DPRINTF(sc, RT_DEBUG_ANY, "suspending\n");
rt_stop(sc);
return (0);
}
static int
rt_resume(device_t dev)
{
struct rt_softc *sc;
struct ifnet *ifp;
sc = device_get_softc(dev);
ifp = sc->ifp;
RT_DPRINTF(sc, RT_DEBUG_ANY, "resuming\n");
if (ifp->if_flags & IFF_UP)
rt_init(sc);
return (0);
}
/*
* rt_init_locked - Run initialization process having locked mtx.
*/
static void
rt_init_locked(void *priv)
{
struct rt_softc *sc;
struct ifnet *ifp;
#ifdef IF_RT_PHY_SUPPORT
struct mii_data *mii;
#endif
int i, ntries;
uint32_t tmp;
sc = priv;
ifp = sc->ifp;
#ifdef IF_RT_PHY_SUPPORT
mii = device_get_softc(sc->rt_miibus);
#endif
RT_DPRINTF(sc, RT_DEBUG_ANY, "initializing\n");
RT_SOFTC_ASSERT_LOCKED(sc);
/* hardware reset */
//RT_WRITE(sc, GE_PORT_BASE + FE_RST_GLO, PSE_RESET);
//rt305x_sysctl_set(SYSCTL_RSTCTRL, SYSCTL_RSTCTRL_FRENG);
/* Fwd to CPU (uni|broad|multi)cast and Unknown */
if (sc->gdma1_base != 0)
RT_WRITE(sc, sc->gdma1_base + GDMA_FWD_CFG,
(
GDM_ICS_EN | /* Enable IP Csum */
GDM_TCS_EN | /* Enable TCP Csum */
GDM_UCS_EN | /* Enable UDP Csum */
GDM_STRPCRC | /* Strip CRC from packet */
GDM_DST_PORT_CPU << GDM_UFRC_P_SHIFT | /* fwd UCast to CPU */
GDM_DST_PORT_CPU << GDM_BFRC_P_SHIFT | /* fwd BCast to CPU */
GDM_DST_PORT_CPU << GDM_MFRC_P_SHIFT | /* fwd MCast to CPU */
GDM_DST_PORT_CPU << GDM_OFRC_P_SHIFT /* fwd Other to CPU */
));
/* disable DMA engine */
RT_WRITE(sc, sc->pdma_glo_cfg, 0);
RT_WRITE(sc, sc->pdma_rst_idx, 0xffffffff);
/* wait while DMA engine is busy */
for (ntries = 0; ntries < 100; ntries++) {
tmp = RT_READ(sc, sc->pdma_glo_cfg);
if (!(tmp & (FE_TX_DMA_BUSY | FE_RX_DMA_BUSY)))
break;
DELAY(1000);
}
if (ntries == 100) {
device_printf(sc->dev, "timeout waiting for DMA engine\n");
goto fail;
}
/* reset Rx and Tx rings */
tmp = FE_RST_DRX_IDX0 |
FE_RST_DTX_IDX3 |
FE_RST_DTX_IDX2 |
FE_RST_DTX_IDX1 |
FE_RST_DTX_IDX0;
RT_WRITE(sc, sc->pdma_rst_idx, tmp);
/* XXX switch set mac address */
for (i = 0; i < RT_SOFTC_TX_RING_COUNT; i++)
rt_reset_tx_ring(sc, &sc->tx_ring[i]);
for (i = 0; i < RT_SOFTC_TX_RING_COUNT; i++) {
/* update TX_BASE_PTRx */
RT_WRITE(sc, sc->tx_base_ptr[i],
sc->tx_ring[i].desc_phys_addr);
RT_WRITE(sc, sc->tx_max_cnt[i],
RT_SOFTC_TX_RING_DESC_COUNT);
RT_WRITE(sc, sc->tx_ctx_idx[i], 0);
}
/* init Rx ring */
for (i = 0; i < sc->rx_ring_count; i++)
rt_reset_rx_ring(sc, &sc->rx_ring[i]);
/* update RX_BASE_PTRx */
for (i = 0; i < sc->rx_ring_count; i++) {
RT_WRITE(sc, sc->rx_base_ptr[i],
sc->rx_ring[i].desc_phys_addr);
RT_WRITE(sc, sc->rx_max_cnt[i],
RT_SOFTC_RX_RING_DATA_COUNT);
RT_WRITE(sc, sc->rx_calc_idx[i],
RT_SOFTC_RX_RING_DATA_COUNT - 1);
}
/* write back DDONE, 16byte burst enable RX/TX DMA */
tmp = FE_TX_WB_DDONE | FE_DMA_BT_SIZE16 | FE_RX_DMA_EN | FE_TX_DMA_EN;
if (sc->rt_chipid == RT_CHIPID_MT7620 ||
sc->rt_chipid == RT_CHIPID_MT7621)
tmp |= (1<<31);
RT_WRITE(sc, sc->pdma_glo_cfg, tmp);
/* disable interrupts mitigation */
RT_WRITE(sc, sc->delay_int_cfg, 0);
/* clear pending interrupts */
RT_WRITE(sc, sc->fe_int_status, 0xffffffff);
/* enable interrupts */
if (sc->rt_chipid == RT_CHIPID_RT5350 ||
sc->rt_chipid == RT_CHIPID_MT7620 ||
sc->rt_chipid == RT_CHIPID_MT7621)
tmp = RT5350_INT_TX_COHERENT |
RT5350_INT_RX_COHERENT |
RT5350_INT_TXQ3_DONE |
RT5350_INT_TXQ2_DONE |
RT5350_INT_TXQ1_DONE |
RT5350_INT_TXQ0_DONE |
RT5350_INT_RXQ1_DONE |
RT5350_INT_RXQ0_DONE;
else
tmp = CNT_PPE_AF |
CNT_GDM_AF |
PSE_P2_FC |
GDM_CRC_DROP |
PSE_BUF_DROP |
GDM_OTHER_DROP |
PSE_P1_FC |
PSE_P0_FC |
PSE_FQ_EMPTY |
INT_TX_COHERENT |
INT_RX_COHERENT |
INT_TXQ3_DONE |
INT_TXQ2_DONE |
INT_TXQ1_DONE |
INT_TXQ0_DONE |
INT_RX_DONE;
sc->intr_enable_mask = tmp;
RT_WRITE(sc, sc->fe_int_enable, tmp);
if (rt_txrx_enable(sc) != 0)
goto fail;
#ifdef IF_RT_PHY_SUPPORT
if (mii) mii_mediachg(mii);
#endif /* IF_RT_PHY_SUPPORT */
ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
ifp->if_drv_flags |= IFF_DRV_RUNNING;
sc->periodic_round = 0;
callout_reset(&sc->periodic_ch, hz / 10, rt_periodic, sc);
return;
fail:
rt_stop_locked(sc);
}
/*
* rt_init - lock and initialize device.
*/
static void
rt_init(void *priv)
{
struct rt_softc *sc;
sc = priv;
RT_SOFTC_LOCK(sc);
rt_init_locked(sc);
RT_SOFTC_UNLOCK(sc);
}
/*
* rt_stop_locked - stop TX/RX w/ lock
*/
static void
rt_stop_locked(void *priv)
{
struct rt_softc *sc;
struct ifnet *ifp;
sc = priv;
ifp = sc->ifp;
RT_DPRINTF(sc, RT_DEBUG_ANY, "stopping\n");
RT_SOFTC_ASSERT_LOCKED(sc);
sc->tx_timer = 0;
ifp->if_drv_flags &= ~(IFF_DRV_RUNNING | IFF_DRV_OACTIVE);
callout_stop(&sc->periodic_ch);
callout_stop(&sc->tx_watchdog_ch);
RT_SOFTC_UNLOCK(sc);
taskqueue_block(sc->taskqueue);
/*
* Sometime rt_stop_locked called from isr and we get panic
* When found, I fix it
*/
#ifdef notyet
taskqueue_drain(sc->taskqueue, &sc->rx_done_task);
taskqueue_drain(sc->taskqueue, &sc->tx_done_task);
taskqueue_drain(sc->taskqueue, &sc->periodic_task);
#endif
RT_SOFTC_LOCK(sc);
/* disable interrupts */
RT_WRITE(sc, sc->fe_int_enable, 0);
if(sc->rt_chipid != RT_CHIPID_RT5350 &&
sc->rt_chipid != RT_CHIPID_MT7620 &&
sc->rt_chipid != RT_CHIPID_MT7621) {
/* reset adapter */
RT_WRITE(sc, GE_PORT_BASE + FE_RST_GLO, PSE_RESET);
}
if (sc->gdma1_base != 0)
RT_WRITE(sc, sc->gdma1_base + GDMA_FWD_CFG,
(
GDM_ICS_EN | /* Enable IP Csum */
GDM_TCS_EN | /* Enable TCP Csum */
GDM_UCS_EN | /* Enable UDP Csum */
GDM_STRPCRC | /* Strip CRC from packet */
GDM_DST_PORT_CPU << GDM_UFRC_P_SHIFT | /* fwd UCast to CPU */
GDM_DST_PORT_CPU << GDM_BFRC_P_SHIFT | /* fwd BCast to CPU */
GDM_DST_PORT_CPU << GDM_MFRC_P_SHIFT | /* fwd MCast to CPU */
GDM_DST_PORT_CPU << GDM_OFRC_P_SHIFT /* fwd Other to CPU */
));
}
static void
rt_stop(void *priv)
{
struct rt_softc *sc;
sc = priv;
RT_SOFTC_LOCK(sc);
rt_stop_locked(sc);
RT_SOFTC_UNLOCK(sc);
}
/*
* rt_tx_data - transmit packet.
*/
static int
rt_tx_data(struct rt_softc *sc, struct mbuf *m, int qid)
{
struct ifnet *ifp;
struct rt_softc_tx_ring *ring;
struct rt_softc_tx_data *data;
struct rt_txdesc *desc;
struct mbuf *m_d;
bus_dma_segment_t dma_seg[RT_SOFTC_MAX_SCATTER];
int error, ndmasegs, ndescs, i;
KASSERT(qid >= 0 && qid < RT_SOFTC_TX_RING_COUNT,
("%s: Tx data: invalid qid=%d\n",
device_get_nameunit(sc->dev), qid));
RT_SOFTC_TX_RING_ASSERT_LOCKED(&sc->tx_ring[qid]);
ifp = sc->ifp;
ring = &sc->tx_ring[qid];
desc = &ring->desc[ring->desc_cur];
data = &ring->data[ring->data_cur];
error = bus_dmamap_load_mbuf_sg(ring->data_dma_tag, data->dma_map, m,
dma_seg, &ndmasegs, 0);
if (error != 0) {
/* too many fragments, linearize */
RT_DPRINTF(sc, RT_DEBUG_TX,
"could not load mbuf DMA map, trying to linearize "
"mbuf: ndmasegs=%d, len=%d, error=%d\n",
ndmasegs, m->m_pkthdr.len, error);
m_d = m_collapse(m, M_NOWAIT, 16);
if (m_d == NULL) {
m_freem(m);
m = NULL;
return (ENOMEM);
}
m = m_d;
sc->tx_defrag_packets++;
error = bus_dmamap_load_mbuf_sg(ring->data_dma_tag,
data->dma_map, m, dma_seg, &ndmasegs, 0);
if (error != 0) {
device_printf(sc->dev, "could not load mbuf DMA map: "
"ndmasegs=%d, len=%d, error=%d\n",
ndmasegs, m->m_pkthdr.len, error);
m_freem(m);
return (error);
}
}
if (m->m_pkthdr.len == 0)
ndmasegs = 0;
/* determine how many Tx descs are required */
ndescs = 1 + ndmasegs / 2;
if ((ring->desc_queued + ndescs) >
(RT_SOFTC_TX_RING_DESC_COUNT - 2)) {
RT_DPRINTF(sc, RT_DEBUG_TX,
"there are not enough Tx descs\n");
sc->no_tx_desc_avail++;
bus_dmamap_unload(ring->data_dma_tag, data->dma_map);
m_freem(m);
return (EFBIG);
}
data->m = m;
/* set up Tx descs */
for (i = 0; i < ndmasegs; i += 2) {
/* TODO: this needs to be refined as MT7620 for example has
* a different word3 layout than RT305x and RT5350 (the last
* one doesn't use word3 at all). And so does MT7621...
*/
if (sc->rt_chipid != RT_CHIPID_MT7621) {
/* Set destination */
if (sc->rt_chipid != RT_CHIPID_MT7620)
desc->dst = (TXDSCR_DST_PORT_GDMA1);
if ((ifp->if_capenable & IFCAP_TXCSUM) != 0)
desc->dst |= (TXDSCR_IP_CSUM_GEN |
TXDSCR_UDP_CSUM_GEN | TXDSCR_TCP_CSUM_GEN);
/* Set queue id */
desc->qn = qid;
/* No PPPoE */
desc->pppoe = 0;
/* No VLAN */
desc->vid = 0;
} else {
desc->vid = 0;
desc->pppoe = 0;
desc->qn = 0;
desc->dst = 2;
}
desc->sdp0 = htole32(dma_seg[i].ds_addr);
desc->sdl0 = htole16(dma_seg[i].ds_len |
( ((i+1) == ndmasegs )?RT_TXDESC_SDL0_LASTSEG:0 ));
if ((i+1) < ndmasegs) {
desc->sdp1 = htole32(dma_seg[i+1].ds_addr);
desc->sdl1 = htole16(dma_seg[i+1].ds_len |
( ((i+2) == ndmasegs )?RT_TXDESC_SDL1_LASTSEG:0 ));
} else {
desc->sdp1 = 0;
desc->sdl1 = 0;
}
if ((i+2) < ndmasegs) {
ring->desc_queued++;
ring->desc_cur = (ring->desc_cur + 1) %
RT_SOFTC_TX_RING_DESC_COUNT;
}
desc = &ring->desc[ring->desc_cur];
}
RT_DPRINTF(sc, RT_DEBUG_TX, "sending data: len=%d, ndmasegs=%d, "
"DMA ds_len=%d/%d/%d/%d/%d\n",
m->m_pkthdr.len, ndmasegs,
(int) dma_seg[0].ds_len,
(int) dma_seg[1].ds_len,
(int) dma_seg[2].ds_len,
(int) dma_seg[3].ds_len,
(int) dma_seg[4].ds_len);
bus_dmamap_sync(ring->seg0_dma_tag, ring->seg0_dma_map,
BUS_DMASYNC_PREWRITE);
bus_dmamap_sync(ring->data_dma_tag, data->dma_map,
BUS_DMASYNC_PREWRITE);
bus_dmamap_sync(ring->desc_dma_tag, ring->desc_dma_map,
BUS_DMASYNC_PREWRITE);
ring->desc_queued++;
ring->desc_cur = (ring->desc_cur + 1) % RT_SOFTC_TX_RING_DESC_COUNT;
ring->data_queued++;
ring->data_cur = (ring->data_cur + 1) % RT_SOFTC_TX_RING_DATA_COUNT;
/* kick Tx */
RT_WRITE(sc, sc->tx_ctx_idx[qid], ring->desc_cur);
return (0);
}
/*
* rt_start - start Transmit/Receive
*/
static void
rt_start(struct ifnet *ifp)
{
struct rt_softc *sc;
struct mbuf *m;
int qid = 0 /* XXX must check QoS priority */;
sc = ifp->if_softc;
if (!(ifp->if_drv_flags & IFF_DRV_RUNNING))
return;
for (;;) {
IFQ_DRV_DEQUEUE(&ifp->if_snd, m);
if (m == NULL)
break;
m->m_pkthdr.rcvif = NULL;
RT_SOFTC_TX_RING_LOCK(&sc->tx_ring[qid]);
if (sc->tx_ring[qid].data_queued >=
RT_SOFTC_TX_RING_DATA_COUNT) {
RT_SOFTC_TX_RING_UNLOCK(&sc->tx_ring[qid]);
RT_DPRINTF(sc, RT_DEBUG_TX,
"if_start: Tx ring with qid=%d is full\n", qid);
m_freem(m);
ifp->if_drv_flags |= IFF_DRV_OACTIVE;
if_inc_counter(ifp, IFCOUNTER_OERRORS, 1);
sc->tx_data_queue_full[qid]++;
break;
}
if (rt_tx_data(sc, m, qid) != 0) {
RT_SOFTC_TX_RING_UNLOCK(&sc->tx_ring[qid]);
if_inc_counter(ifp, IFCOUNTER_OERRORS, 1);
break;
}
RT_SOFTC_TX_RING_UNLOCK(&sc->tx_ring[qid]);
sc->tx_timer = RT_TX_WATCHDOG_TIMEOUT;
callout_reset(&sc->tx_watchdog_ch, hz, rt_tx_watchdog, sc);
}
}
/*
* rt_update_promisc - set/clear promiscuous mode. Unused yet, because
* filtering done by attached Ethernet switch.
*/
static void
rt_update_promisc(struct ifnet *ifp)
{
struct rt_softc *sc;
sc = ifp->if_softc;
printf("%s: %s promiscuous mode\n",
device_get_nameunit(sc->dev),
(ifp->if_flags & IFF_PROMISC) ? "entering" : "leaving");
}
/*
* rt_ioctl - ioctl handler.
*/
static int
rt_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data)
{
struct rt_softc *sc;
struct ifreq *ifr;
#ifdef IF_RT_PHY_SUPPORT
struct mii_data *mii;
#endif /* IF_RT_PHY_SUPPORT */
int error, startall;
sc = ifp->if_softc;
ifr = (struct ifreq *) data;
error = 0;
switch (cmd) {
case SIOCSIFFLAGS:
startall = 0;
RT_SOFTC_LOCK(sc);
if (ifp->if_flags & IFF_UP) {
if (ifp->if_drv_flags & IFF_DRV_RUNNING) {
if ((ifp->if_flags ^ sc->if_flags) &
IFF_PROMISC)
rt_update_promisc(ifp);
} else {
rt_init_locked(sc);
startall = 1;
}
} else {
if (ifp->if_drv_flags & IFF_DRV_RUNNING)
rt_stop_locked(sc);
}
sc->if_flags = ifp->if_flags;
RT_SOFTC_UNLOCK(sc);
break;
case SIOCGIFMEDIA:
case SIOCSIFMEDIA:
#ifdef IF_RT_PHY_SUPPORT
mii = device_get_softc(sc->rt_miibus);
error = ifmedia_ioctl(ifp, ifr, &mii->mii_media, cmd);
#else
error = ifmedia_ioctl(ifp, ifr, &sc->rt_ifmedia, cmd);
#endif /* IF_RT_PHY_SUPPORT */
break;
default:
error = ether_ioctl(ifp, cmd, data);
break;
}
return (error);
}
/*
* rt_periodic - Handler of PERIODIC interrupt
*/
static void
rt_periodic(void *arg)
{
struct rt_softc *sc;
sc = arg;
RT_DPRINTF(sc, RT_DEBUG_PERIODIC, "periodic\n");
taskqueue_enqueue(sc->taskqueue, &sc->periodic_task);
}
/*
* rt_tx_watchdog - Handler of TX Watchdog
*/
static void
rt_tx_watchdog(void *arg)
{
struct rt_softc *sc;
struct ifnet *ifp;
sc = arg;
ifp = sc->ifp;
if (sc->tx_timer == 0)
return;
if (--sc->tx_timer == 0) {
device_printf(sc->dev, "Tx watchdog timeout: resetting\n");
#ifdef notyet
/*
* XXX: Commented out, because reset break input.
*/
rt_stop_locked(sc);
rt_init_locked(sc);
#endif
if_inc_counter(ifp, IFCOUNTER_OERRORS, 1);
sc->tx_watchdog_timeouts++;
}
callout_reset(&sc->tx_watchdog_ch, hz, rt_tx_watchdog, sc);
}
/*
* rt_cnt_ppe_af - Handler of PPE Counter Table Almost Full interrupt
*/
static void
rt_cnt_ppe_af(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR, "PPE Counter Table Almost Full\n");
}
/*
* rt_cnt_gdm_af - Handler of GDMA 1 & 2 Counter Table Almost Full interrupt
*/
static void
rt_cnt_gdm_af(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR,
"GDMA 1 & 2 Counter Table Almost Full\n");
}
/*
* rt_pse_p2_fc - Handler of PSE port2 (GDMA 2) flow control interrupt
*/
static void
rt_pse_p2_fc(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR,
"PSE port2 (GDMA 2) flow control asserted.\n");
}
/*
* rt_gdm_crc_drop - Handler of GDMA 1/2 discard a packet due to CRC error
* interrupt
*/
static void
rt_gdm_crc_drop(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR,
"GDMA 1 & 2 discard a packet due to CRC error\n");
}
/*
* rt_pse_buf_drop - Handler of buffer sharing limitation interrupt
*/
static void
rt_pse_buf_drop(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR,
"PSE discards a packet due to buffer sharing limitation\n");
}
/*
* rt_gdm_other_drop - Handler of discard on other reason interrupt
*/
static void
rt_gdm_other_drop(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR,
"GDMA 1 & 2 discard a packet due to other reason\n");
}
/*
* rt_pse_p1_fc - Handler of PSE port1 (GDMA 1) flow control interrupt
*/
static void
rt_pse_p1_fc(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR,
"PSE port1 (GDMA 1) flow control asserted.\n");
}
/*
* rt_pse_p0_fc - Handler of PSE port0 (CDMA) flow control interrupt
*/
static void
rt_pse_p0_fc(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR,
"PSE port0 (CDMA) flow control asserted.\n");
}
/*
* rt_pse_fq_empty - Handler of PSE free Q empty threshold reached interrupt
*/
static void
rt_pse_fq_empty(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR,
"PSE free Q empty threshold reached & forced drop "
"condition occurred.\n");
}
/*
* rt_intr - main ISR
*/
static void
rt_intr(void *arg)
{
struct rt_softc *sc;
struct ifnet *ifp;
uint32_t status;
sc = arg;
ifp = sc->ifp;
/* acknowledge interrupts */
status = RT_READ(sc, sc->fe_int_status);
RT_WRITE(sc, sc->fe_int_status, status);
RT_DPRINTF(sc, RT_DEBUG_INTR, "interrupt: status=0x%08x\n", status);
if (status == 0xffffffff || /* device likely went away */
status == 0) /* not for us */
return;
sc->interrupts++;
if (!(ifp->if_drv_flags & IFF_DRV_RUNNING))
return;
if (status & CNT_PPE_AF)
rt_cnt_ppe_af(sc);
if (status & CNT_GDM_AF)
rt_cnt_gdm_af(sc);
if (status & PSE_P2_FC)
rt_pse_p2_fc(sc);
if (status & GDM_CRC_DROP)
rt_gdm_crc_drop(sc);
if (status & PSE_BUF_DROP)
rt_pse_buf_drop(sc);
if (status & GDM_OTHER_DROP)
rt_gdm_other_drop(sc);
if (status & PSE_P1_FC)
rt_pse_p1_fc(sc);
if (status & PSE_P0_FC)
rt_pse_p0_fc(sc);
if (status & PSE_FQ_EMPTY)
rt_pse_fq_empty(sc);
if (status & INT_TX_COHERENT)
rt_tx_coherent_intr(sc);
if (status & INT_RX_COHERENT)
rt_rx_coherent_intr(sc);
if (status & RX_DLY_INT)
rt_rx_delay_intr(sc);
if (status & TX_DLY_INT)
rt_tx_delay_intr(sc);
if (status & INT_RX_DONE)
rt_rx_intr(sc, 0);
if (status & INT_TXQ3_DONE)
rt_tx_intr(sc, 3);
if (status & INT_TXQ2_DONE)
rt_tx_intr(sc, 2);
if (status & INT_TXQ1_DONE)
rt_tx_intr(sc, 1);
if (status & INT_TXQ0_DONE)
rt_tx_intr(sc, 0);
}
/*
* rt_rt5350_intr - main ISR for Ralink 5350 SoC
*/
static void
rt_rt5350_intr(void *arg)
{
struct rt_softc *sc;
struct ifnet *ifp;
uint32_t status;
sc = arg;
ifp = sc->ifp;
/* acknowledge interrupts */
status = RT_READ(sc, sc->fe_int_status);
RT_WRITE(sc, sc->fe_int_status, status);
RT_DPRINTF(sc, RT_DEBUG_INTR, "interrupt: status=0x%08x\n", status);
if (status == 0xffffffff || /* device likely went away */
status == 0) /* not for us */
return;
sc->interrupts++;
if (!(ifp->if_drv_flags & IFF_DRV_RUNNING))
return;
if (status & RT5350_INT_TX_COHERENT)
rt_tx_coherent_intr(sc);
if (status & RT5350_INT_RX_COHERENT)
rt_rx_coherent_intr(sc);
if (status & RT5350_RX_DLY_INT)
rt_rx_delay_intr(sc);
if (status & RT5350_TX_DLY_INT)
rt_tx_delay_intr(sc);
if (status & RT5350_INT_RXQ1_DONE)
rt_rx_intr(sc, 1);
if (status & RT5350_INT_RXQ0_DONE)
rt_rx_intr(sc, 0);
if (status & RT5350_INT_TXQ3_DONE)
rt_tx_intr(sc, 3);
if (status & RT5350_INT_TXQ2_DONE)
rt_tx_intr(sc, 2);
if (status & RT5350_INT_TXQ1_DONE)
rt_tx_intr(sc, 1);
if (status & RT5350_INT_TXQ0_DONE)
rt_tx_intr(sc, 0);
}
static void
rt_tx_coherent_intr(struct rt_softc *sc)
{
uint32_t tmp;
int i;
RT_DPRINTF(sc, RT_DEBUG_INTR, "Tx coherent interrupt\n");
sc->tx_coherent_interrupts++;
/* restart DMA engine */
tmp = RT_READ(sc, sc->pdma_glo_cfg);
tmp &= ~(FE_TX_WB_DDONE | FE_TX_DMA_EN);
RT_WRITE(sc, sc->pdma_glo_cfg, tmp);
for (i = 0; i < RT_SOFTC_TX_RING_COUNT; i++)
rt_reset_tx_ring(sc, &sc->tx_ring[i]);
for (i = 0; i < RT_SOFTC_TX_RING_COUNT; i++) {
RT_WRITE(sc, sc->tx_base_ptr[i],
sc->tx_ring[i].desc_phys_addr);
RT_WRITE(sc, sc->tx_max_cnt[i],
RT_SOFTC_TX_RING_DESC_COUNT);
RT_WRITE(sc, sc->tx_ctx_idx[i], 0);
}
rt_txrx_enable(sc);
}
/*
* rt_rx_coherent_intr
*/
static void
rt_rx_coherent_intr(struct rt_softc *sc)
{
uint32_t tmp;
int i;
RT_DPRINTF(sc, RT_DEBUG_INTR, "Rx coherent interrupt\n");
sc->rx_coherent_interrupts++;
/* restart DMA engine */
tmp = RT_READ(sc, sc->pdma_glo_cfg);
tmp &= ~(FE_RX_DMA_EN);
RT_WRITE(sc, sc->pdma_glo_cfg, tmp);
/* init Rx ring */
for (i = 0; i < sc->rx_ring_count; i++)
rt_reset_rx_ring(sc, &sc->rx_ring[i]);
for (i = 0; i < sc->rx_ring_count; i++) {
RT_WRITE(sc, sc->rx_base_ptr[i],
sc->rx_ring[i].desc_phys_addr);
RT_WRITE(sc, sc->rx_max_cnt[i],
RT_SOFTC_RX_RING_DATA_COUNT);
RT_WRITE(sc, sc->rx_calc_idx[i],
RT_SOFTC_RX_RING_DATA_COUNT - 1);
}
rt_txrx_enable(sc);
}
/*
* rt_rx_intr - a packet received
*/
static void
rt_rx_intr(struct rt_softc *sc, int qid)
{
KASSERT(qid >= 0 && qid < sc->rx_ring_count,
("%s: Rx interrupt: invalid qid=%d\n",
device_get_nameunit(sc->dev), qid));
RT_DPRINTF(sc, RT_DEBUG_INTR, "Rx interrupt\n");
sc->rx_interrupts[qid]++;
RT_SOFTC_LOCK(sc);
if (!(sc->intr_disable_mask & (sc->int_rx_done_mask << qid))) {
rt_intr_disable(sc, (sc->int_rx_done_mask << qid));
taskqueue_enqueue(sc->taskqueue, &sc->rx_done_task);
}
sc->intr_pending_mask |= (sc->int_rx_done_mask << qid);
RT_SOFTC_UNLOCK(sc);
}
static void
rt_rx_delay_intr(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR, "Rx delay interrupt\n");
sc->rx_delay_interrupts++;
}
static void
rt_tx_delay_intr(struct rt_softc *sc)
{
RT_DPRINTF(sc, RT_DEBUG_INTR, "Tx delay interrupt\n");
sc->tx_delay_interrupts++;
}
/*
* rt_tx_intr - Transsmition of packet done
*/
static void
rt_tx_intr(struct rt_softc *sc, int qid)
{
KASSERT(qid >= 0 && qid < RT_SOFTC_TX_RING_COUNT,
("%s: Tx interrupt: invalid qid=%d\n",
device_get_nameunit(sc->dev), qid));
RT_DPRINTF(sc, RT_DEBUG_INTR, "Tx interrupt: qid=%d\n", qid);
sc->tx_interrupts[qid]++;
RT_SOFTC_LOCK(sc);
if (!(sc->intr_disable_mask & (sc->int_tx_done_mask << qid))) {
rt_intr_disable(sc, (sc->int_tx_done_mask << qid));
taskqueue_enqueue(sc->taskqueue, &sc->tx_done_task);
}
sc->intr_pending_mask |= (sc->int_tx_done_mask << qid);
RT_SOFTC_UNLOCK(sc);
}
/*
* rt_rx_done_task - run RX task
*/
static void
rt_rx_done_task(void *context, int pending)
{
struct rt_softc *sc;
struct ifnet *ifp;
int again;
sc = context;
ifp = sc->ifp;
RT_DPRINTF(sc, RT_DEBUG_RX, "Rx done task\n");
if (!(ifp->if_drv_flags & IFF_DRV_RUNNING))
return;
sc->intr_pending_mask &= ~sc->int_rx_done_mask;
again = rt_rx_eof(sc, &sc->rx_ring[0], sc->rx_process_limit);
RT_SOFTC_LOCK(sc);
if ((sc->intr_pending_mask & sc->int_rx_done_mask) || again) {
RT_DPRINTF(sc, RT_DEBUG_RX,
"Rx done task: scheduling again\n");
taskqueue_enqueue(sc->taskqueue, &sc->rx_done_task);
} else {
rt_intr_enable(sc, sc->int_rx_done_mask);
}
RT_SOFTC_UNLOCK(sc);
}
/*
* rt_tx_done_task - check for pending TX task in all queues
*/
static void
rt_tx_done_task(void *context, int pending)
{
struct rt_softc *sc;
struct ifnet *ifp;
uint32_t intr_mask;
int i;
sc = context;
ifp = sc->ifp;
RT_DPRINTF(sc, RT_DEBUG_TX, "Tx done task\n");
if (!(ifp->if_drv_flags & IFF_DRV_RUNNING))
return;
for (i = RT_SOFTC_TX_RING_COUNT - 1; i >= 0; i--) {
if (sc->intr_pending_mask & (sc->int_tx_done_mask << i)) {
sc->intr_pending_mask &= ~(sc->int_tx_done_mask << i);
rt_tx_eof(sc, &sc->tx_ring[i]);
}
}
sc->tx_timer = 0;
ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
if(sc->rt_chipid == RT_CHIPID_RT5350 ||
sc->rt_chipid == RT_CHIPID_MT7620 ||
sc->rt_chipid == RT_CHIPID_MT7621)
intr_mask = (
RT5350_INT_TXQ3_DONE |
RT5350_INT_TXQ2_DONE |
RT5350_INT_TXQ1_DONE |
RT5350_INT_TXQ0_DONE);
else
intr_mask = (
INT_TXQ3_DONE |
INT_TXQ2_DONE |
INT_TXQ1_DONE |
INT_TXQ0_DONE);
RT_SOFTC_LOCK(sc);
rt_intr_enable(sc, ~sc->intr_pending_mask &
(sc->intr_disable_mask & intr_mask));
if (sc->intr_pending_mask & intr_mask) {
RT_DPRINTF(sc, RT_DEBUG_TX,
"Tx done task: scheduling again\n");
taskqueue_enqueue(sc->taskqueue, &sc->tx_done_task);
}
RT_SOFTC_UNLOCK(sc);
if (!IFQ_IS_EMPTY(&ifp->if_snd))
rt_start(ifp);
}
/*
* rt_periodic_task - run periodic task
*/
static void
rt_periodic_task(void *context, int pending)
{
struct rt_softc *sc;
struct ifnet *ifp;
sc = context;
ifp = sc->ifp;
RT_DPRINTF(sc, RT_DEBUG_PERIODIC, "periodic task: round=%lu\n",
sc->periodic_round);
if (!(ifp->if_drv_flags & IFF_DRV_RUNNING))
return;
RT_SOFTC_LOCK(sc);
sc->periodic_round++;
rt_update_stats(sc);
if ((sc->periodic_round % 10) == 0) {
rt_update_raw_counters(sc);
rt_watchdog(sc);
}
RT_SOFTC_UNLOCK(sc);
callout_reset(&sc->periodic_ch, hz / 10, rt_periodic, sc);
}
/*
* rt_rx_eof - check for frames that done by DMA engine and pass it into
* network subsystem.
*/
static int
rt_rx_eof(struct rt_softc *sc, struct rt_softc_rx_ring *ring, int limit)
{
struct ifnet *ifp;
/* struct rt_softc_rx_ring *ring; */
struct rt_rxdesc *desc;
struct rt_softc_rx_data *data;
struct mbuf *m, *mnew;
bus_dma_segment_t segs[1];
bus_dmamap_t dma_map;
uint32_t index, desc_flags;
int error, nsegs, len, nframes;
ifp = sc->ifp;
/* ring = &sc->rx_ring[0]; */
nframes = 0;
while (limit != 0) {
index = RT_READ(sc, sc->rx_drx_idx[0]);
if (ring->cur == index)
break;
desc = &ring->desc[ring->cur];
data = &ring->data[ring->cur];
bus_dmamap_sync(ring->desc_dma_tag, ring->desc_dma_map,
BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
#ifdef IF_RT_DEBUG
if ( sc->debug & RT_DEBUG_RX ) {
printf("\nRX Descriptor[%#08x] dump:\n", (u_int)desc);
hexdump(desc, 16, 0, 0);
printf("-----------------------------------\n");
}
#endif
/* XXX Sometime device don`t set DDONE bit */
#ifdef DDONE_FIXED
if (!(desc->sdl0 & htole16(RT_RXDESC_SDL0_DDONE))) {
RT_DPRINTF(sc, RT_DEBUG_RX, "DDONE=0, try next\n");
break;
}
#endif
len = le16toh(desc->sdl0) & 0x3fff;
RT_DPRINTF(sc, RT_DEBUG_RX, "new frame len=%d\n", len);
nframes++;
mnew = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR,
MJUMPAGESIZE);
if (mnew == NULL) {
sc->rx_mbuf_alloc_errors++;
if_inc_counter(ifp, IFCOUNTER_IERRORS, 1);
goto skip;
}
mnew->m_len = mnew->m_pkthdr.len = MJUMPAGESIZE;
error = bus_dmamap_load_mbuf_sg(ring->data_dma_tag,
ring->spare_dma_map, mnew, segs, &nsegs, BUS_DMA_NOWAIT);
if (error != 0) {
RT_DPRINTF(sc, RT_DEBUG_RX,
"could not load Rx mbuf DMA map: "
"error=%d, nsegs=%d\n",
error, nsegs);
m_freem(mnew);
sc->rx_mbuf_dmamap_errors++;
if_inc_counter(ifp, IFCOUNTER_IERRORS, 1);
goto skip;
}
KASSERT(nsegs == 1, ("%s: too many DMA segments",
device_get_nameunit(sc->dev)));
bus_dmamap_sync(ring->data_dma_tag, data->dma_map,
BUS_DMASYNC_POSTREAD);
bus_dmamap_unload(ring->data_dma_tag, data->dma_map);
dma_map = data->dma_map;
data->dma_map = ring->spare_dma_map;
ring->spare_dma_map = dma_map;
bus_dmamap_sync(ring->data_dma_tag, data->dma_map,
BUS_DMASYNC_PREREAD);
m = data->m;
desc_flags = desc->word3;
data->m = mnew;
/* Add 2 for proper align of RX IP header */
desc->sdp0 = htole32(segs[0].ds_addr+2);
desc->sdl0 = htole32(segs[0].ds_len-2);
desc->word3 = 0;
RT_DPRINTF(sc, RT_DEBUG_RX,
"Rx frame: rxdesc flags=0x%08x\n", desc_flags);
m->m_pkthdr.rcvif = ifp;
/* Add 2 to fix data align, after sdp0 = addr + 2 */
m->m_data += 2;
m->m_pkthdr.len = m->m_len = len;
/* check for crc errors */
if ((ifp->if_capenable & IFCAP_RXCSUM) != 0) {
/*check for valid checksum*/
if (desc_flags & (sc->csum_fail_ip|sc->csum_fail_l4)) {
RT_DPRINTF(sc, RT_DEBUG_RX,
"rxdesc: crc error\n");
if_inc_counter(ifp, IFCOUNTER_IERRORS, 1);
if (!(ifp->if_flags & IFF_PROMISC)) {
m_freem(m);
goto skip;
}
}
if ((desc_flags & sc->csum_fail_ip) == 0) {
m->m_pkthdr.csum_flags |= CSUM_IP_CHECKED;
m->m_pkthdr.csum_flags |= CSUM_IP_VALID;
m->m_pkthdr.csum_data = 0xffff;
}
m->m_flags &= ~M_HASFCS;
}
(*ifp->if_input)(ifp, m);
skip:
desc->sdl0 &= ~htole16(RT_RXDESC_SDL0_DDONE);
bus_dmamap_sync(ring->desc_dma_tag, ring->desc_dma_map,
BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
ring->cur = (ring->cur + 1) % RT_SOFTC_RX_RING_DATA_COUNT;
limit--;
}
if (ring->cur == 0)
RT_WRITE(sc, sc->rx_calc_idx[0],
RT_SOFTC_RX_RING_DATA_COUNT - 1);
else
RT_WRITE(sc, sc->rx_calc_idx[0],
ring->cur - 1);
RT_DPRINTF(sc, RT_DEBUG_RX, "Rx eof: nframes=%d\n", nframes);
sc->rx_packets += nframes;
return (limit == 0);
}
/*
* rt_tx_eof - check for successful transmitted frames and mark their
* descriptor as free.
*/
static void
rt_tx_eof(struct rt_softc *sc, struct rt_softc_tx_ring *ring)
{
struct ifnet *ifp;
struct rt_txdesc *desc;
struct rt_softc_tx_data *data;
uint32_t index;
int ndescs, nframes;
ifp = sc->ifp;
ndescs = 0;
nframes = 0;
for (;;) {
index = RT_READ(sc, sc->tx_dtx_idx[ring->qid]);
if (ring->desc_next == index)
break;
ndescs++;
desc = &ring->desc[ring->desc_next];
bus_dmamap_sync(ring->desc_dma_tag, ring->desc_dma_map,
BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
if (desc->sdl0 & htole16(RT_TXDESC_SDL0_LASTSEG) ||
desc->sdl1 & htole16(RT_TXDESC_SDL1_LASTSEG)) {
nframes++;
data = &ring->data[ring->data_next];
bus_dmamap_sync(ring->data_dma_tag, data->dma_map,
BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(ring->data_dma_tag, data->dma_map);
m_freem(data->m);
data->m = NULL;
if_inc_counter(ifp, IFCOUNTER_OPACKETS, 1);
RT_SOFTC_TX_RING_LOCK(ring);
ring->data_queued--;
ring->data_next = (ring->data_next + 1) %
RT_SOFTC_TX_RING_DATA_COUNT;
RT_SOFTC_TX_RING_UNLOCK(ring);
}
desc->sdl0 &= ~htole16(RT_TXDESC_SDL0_DDONE);
bus_dmamap_sync(ring->desc_dma_tag, ring->desc_dma_map,
BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
RT_SOFTC_TX_RING_LOCK(ring);
ring->desc_queued--;
ring->desc_next = (ring->desc_next + 1) %
RT_SOFTC_TX_RING_DESC_COUNT;
RT_SOFTC_TX_RING_UNLOCK(ring);
}
RT_DPRINTF(sc, RT_DEBUG_TX,
"Tx eof: qid=%d, ndescs=%d, nframes=%d\n", ring->qid, ndescs,
nframes);
}
/*
* rt_update_stats - query statistics counters and update related variables.
*/
static void
rt_update_stats(struct rt_softc *sc)
{
struct ifnet *ifp;
ifp = sc->ifp;
RT_DPRINTF(sc, RT_DEBUG_STATS, "update statistic: \n");
/* XXX do update stats here */
}
/*
* rt_watchdog - reinit device on watchdog event.
*/
static void
rt_watchdog(struct rt_softc *sc)
{
uint32_t tmp;
#ifdef notyet
int ntries;
#endif
if(sc->rt_chipid != RT_CHIPID_RT5350 &&
sc->rt_chipid != RT_CHIPID_MT7620 &&
sc->rt_chipid != RT_CHIPID_MT7621) {
tmp = RT_READ(sc, PSE_BASE + CDMA_OQ_STA);
RT_DPRINTF(sc, RT_DEBUG_WATCHDOG,
"watchdog: PSE_IQ_STA=0x%08x\n", tmp);
}
/* XXX: do not reset */
#ifdef notyet
if (((tmp >> P0_IQ_PCNT_SHIFT) & 0xff) != 0) {
sc->tx_queue_not_empty[0]++;
for (ntries = 0; ntries < 10; ntries++) {
tmp = RT_READ(sc, PSE_BASE + PSE_IQ_STA);
if (((tmp >> P0_IQ_PCNT_SHIFT) & 0xff) == 0)
break;
DELAY(1);
}
}
if (((tmp >> P1_IQ_PCNT_SHIFT) & 0xff) != 0) {
sc->tx_queue_not_empty[1]++;
for (ntries = 0; ntries < 10; ntries++) {
tmp = RT_READ(sc, PSE_BASE + PSE_IQ_STA);
if (((tmp >> P1_IQ_PCNT_SHIFT) & 0xff) == 0)
break;
DELAY(1);
}
}
#endif
}
/*
* rt_update_raw_counters - update counters.
*/
static void
rt_update_raw_counters(struct rt_softc *sc)
{
sc->tx_bytes += RT_READ(sc, CNTR_BASE + GDMA_TX_GBCNT0);
sc->tx_packets += RT_READ(sc, CNTR_BASE + GDMA_TX_GPCNT0);
sc->tx_skip += RT_READ(sc, CNTR_BASE + GDMA_TX_SKIPCNT0);
sc->tx_collision+= RT_READ(sc, CNTR_BASE + GDMA_TX_COLCNT0);
sc->rx_bytes += RT_READ(sc, CNTR_BASE + GDMA_RX_GBCNT0);
sc->rx_packets += RT_READ(sc, CNTR_BASE + GDMA_RX_GPCNT0);
sc->rx_crc_err += RT_READ(sc, CNTR_BASE + GDMA_RX_CSUM_ERCNT0);
sc->rx_short_err+= RT_READ(sc, CNTR_BASE + GDMA_RX_SHORT_ERCNT0);
sc->rx_long_err += RT_READ(sc, CNTR_BASE + GDMA_RX_LONG_ERCNT0);
sc->rx_phy_err += RT_READ(sc, CNTR_BASE + GDMA_RX_FERCNT0);
sc->rx_fifo_overflows+= RT_READ(sc, CNTR_BASE + GDMA_RX_OERCNT0);
}
static void
rt_intr_enable(struct rt_softc *sc, uint32_t intr_mask)
{
uint32_t tmp;
sc->intr_disable_mask &= ~intr_mask;
tmp = sc->intr_enable_mask & ~sc->intr_disable_mask;
RT_WRITE(sc, sc->fe_int_enable, tmp);
}
static void
rt_intr_disable(struct rt_softc *sc, uint32_t intr_mask)
{
uint32_t tmp;
sc->intr_disable_mask |= intr_mask;
tmp = sc->intr_enable_mask & ~sc->intr_disable_mask;
RT_WRITE(sc, sc->fe_int_enable, tmp);
}
/*
* rt_txrx_enable - enable TX/RX DMA
*/
static int
rt_txrx_enable(struct rt_softc *sc)
{
struct ifnet *ifp;
uint32_t tmp;
int ntries;
ifp = sc->ifp;
/* enable Tx/Rx DMA engine */
for (ntries = 0; ntries < 200; ntries++) {
tmp = RT_READ(sc, sc->pdma_glo_cfg);
if (!(tmp & (FE_TX_DMA_BUSY | FE_RX_DMA_BUSY)))
break;
DELAY(1000);
}
if (ntries == 200) {
device_printf(sc->dev, "timeout waiting for DMA engine\n");
return (-1);
}
DELAY(50);
tmp |= FE_TX_WB_DDONE | FE_RX_DMA_EN | FE_TX_DMA_EN;
RT_WRITE(sc, sc->pdma_glo_cfg, tmp);
/* XXX set Rx filter */
return (0);
}
/*
* rt_alloc_rx_ring - allocate RX DMA ring buffer
*/
static int
rt_alloc_rx_ring(struct rt_softc *sc, struct rt_softc_rx_ring *ring, int qid)
{
struct rt_rxdesc *desc;
struct rt_softc_rx_data *data;
bus_dma_segment_t segs[1];
int i, nsegs, error;
error = bus_dma_tag_create(bus_get_dma_tag(sc->dev), PAGE_SIZE, 0,
BUS_SPACE_MAXADDR_32BIT, BUS_SPACE_MAXADDR, NULL, NULL,
RT_SOFTC_RX_RING_DATA_COUNT * sizeof(struct rt_rxdesc), 1,
RT_SOFTC_RX_RING_DATA_COUNT * sizeof(struct rt_rxdesc),
0, NULL, NULL, &ring->desc_dma_tag);
if (error != 0) {
device_printf(sc->dev,
"could not create Rx desc DMA tag\n");
goto fail;
}
error = bus_dmamem_alloc(ring->desc_dma_tag, (void **) &ring->desc,
BUS_DMA_NOWAIT | BUS_DMA_ZERO, &ring->desc_dma_map);
if (error != 0) {
device_printf(sc->dev,
"could not allocate Rx desc DMA memory\n");
goto fail;
}
error = bus_dmamap_load(ring->desc_dma_tag, ring->desc_dma_map,
ring->desc,
RT_SOFTC_RX_RING_DATA_COUNT * sizeof(struct rt_rxdesc),
rt_dma_map_addr, &ring->desc_phys_addr, 0);
if (error != 0) {
device_printf(sc->dev, "could not load Rx desc DMA map\n");
goto fail;
}
error = bus_dma_tag_create(bus_get_dma_tag(sc->dev), PAGE_SIZE, 0,
BUS_SPACE_MAXADDR_32BIT, BUS_SPACE_MAXADDR, NULL, NULL,
MJUMPAGESIZE, 1, MJUMPAGESIZE, 0, NULL, NULL,
&ring->data_dma_tag);
if (error != 0) {
device_printf(sc->dev,
"could not create Rx data DMA tag\n");
goto fail;
}
for (i = 0; i < RT_SOFTC_RX_RING_DATA_COUNT; i++) {
desc = &ring->desc[i];
data = &ring->data[i];
error = bus_dmamap_create(ring->data_dma_tag, 0,
&data->dma_map);
if (error != 0) {
device_printf(sc->dev, "could not create Rx data DMA "
"map\n");
goto fail;
}
data->m = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR,
MJUMPAGESIZE);
if (data->m == NULL) {
device_printf(sc->dev, "could not allocate Rx mbuf\n");
error = ENOMEM;
goto fail;
}
data->m->m_len = data->m->m_pkthdr.len = MJUMPAGESIZE;
error = bus_dmamap_load_mbuf_sg(ring->data_dma_tag,
data->dma_map, data->m, segs, &nsegs, BUS_DMA_NOWAIT);
if (error != 0) {
device_printf(sc->dev,
"could not load Rx mbuf DMA map\n");
goto fail;
}
KASSERT(nsegs == 1, ("%s: too many DMA segments",
device_get_nameunit(sc->dev)));
/* Add 2 for proper align of RX IP header */
desc->sdp0 = htole32(segs[0].ds_addr+2);
desc->sdl0 = htole32(segs[0].ds_len-2);
}
error = bus_dmamap_create(ring->data_dma_tag, 0,
&ring->spare_dma_map);
if (error != 0) {
device_printf(sc->dev,
"could not create Rx spare DMA map\n");
goto fail;
}
bus_dmamap_sync(ring->desc_dma_tag, ring->desc_dma_map,
BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
ring->qid = qid;
return (0);
fail:
rt_free_rx_ring(sc, ring);
return (error);
}
/*
* rt_reset_rx_ring - reset RX ring buffer
*/
static void
rt_reset_rx_ring(struct rt_softc *sc, struct rt_softc_rx_ring *ring)
{
struct rt_rxdesc *desc;
int i;
for (i = 0; i < RT_SOFTC_RX_RING_DATA_COUNT; i++) {
desc = &ring->desc[i];
desc->sdl0 &= ~htole16(RT_RXDESC_SDL0_DDONE);
}
bus_dmamap_sync(ring->desc_dma_tag, ring->desc_dma_map,
BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
ring->cur = 0;
}
/*
* rt_free_rx_ring - free memory used by RX ring buffer
*/
static void
rt_free_rx_ring(struct rt_softc *sc, struct rt_softc_rx_ring *ring)
{
struct rt_softc_rx_data *data;
int i;
if (ring->desc != NULL) {
bus_dmamap_sync(ring->desc_dma_tag, ring->desc_dma_map,
BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(ring->desc_dma_tag, ring->desc_dma_map);
bus_dmamem_free(ring->desc_dma_tag, ring->desc,
ring->desc_dma_map);
}
if (ring->desc_dma_tag != NULL)
bus_dma_tag_destroy(ring->desc_dma_tag);
for (i = 0; i < RT_SOFTC_RX_RING_DATA_COUNT; i++) {
data = &ring->data[i];
if (data->m != NULL) {
bus_dmamap_sync(ring->data_dma_tag, data->dma_map,
BUS_DMASYNC_POSTREAD);
bus_dmamap_unload(ring->data_dma_tag, data->dma_map);
m_freem(data->m);
}
if (data->dma_map != NULL)
bus_dmamap_destroy(ring->data_dma_tag, data->dma_map);
}
if (ring->spare_dma_map != NULL)
bus_dmamap_destroy(ring->data_dma_tag, ring->spare_dma_map);
if (ring->data_dma_tag != NULL)
bus_dma_tag_destroy(ring->data_dma_tag);
}
/*
* rt_alloc_tx_ring - allocate TX ring buffer
*/
static int
rt_alloc_tx_ring(struct rt_softc *sc, struct rt_softc_tx_ring *ring, int qid)
{
struct rt_softc_tx_data *data;
int error, i;
mtx_init(&ring->lock, device_get_nameunit(sc->dev), NULL, MTX_DEF);
error = bus_dma_tag_create(bus_get_dma_tag(sc->dev), PAGE_SIZE, 0,
BUS_SPACE_MAXADDR_32BIT, BUS_SPACE_MAXADDR, NULL, NULL,
RT_SOFTC_TX_RING_DESC_COUNT * sizeof(struct rt_txdesc), 1,
RT_SOFTC_TX_RING_DESC_COUNT * sizeof(struct rt_txdesc),
0, NULL, NULL, &ring->desc_dma_tag);
if (error != 0) {
device_printf(sc->dev,
"could not create Tx desc DMA tag\n");
goto fail;
}
error = bus_dmamem_alloc(ring->desc_dma_tag, (void **) &ring->desc,
BUS_DMA_NOWAIT | BUS_DMA_ZERO, &ring->desc_dma_map);
if (error != 0) {
device_printf(sc->dev,
"could not allocate Tx desc DMA memory\n");
goto fail;
}
error = bus_dmamap_load(ring->desc_dma_tag, ring->desc_dma_map,
ring->desc, (RT_SOFTC_TX_RING_DESC_COUNT *
sizeof(struct rt_txdesc)), rt_dma_map_addr,
&ring->desc_phys_addr, 0);
if (error != 0) {
device_printf(sc->dev, "could not load Tx desc DMA map\n");
goto fail;
}
ring->desc_queued = 0;
ring->desc_cur = 0;
ring->desc_next = 0;
error = bus_dma_tag_create(bus_get_dma_tag(sc->dev), PAGE_SIZE, 0,
BUS_SPACE_MAXADDR_32BIT, BUS_SPACE_MAXADDR, NULL, NULL,
RT_SOFTC_TX_RING_DATA_COUNT * RT_TX_DATA_SEG0_SIZE, 1,
RT_SOFTC_TX_RING_DATA_COUNT * RT_TX_DATA_SEG0_SIZE,
0, NULL, NULL, &ring->seg0_dma_tag);
if (error != 0) {
device_printf(sc->dev,
"could not create Tx seg0 DMA tag\n");
goto fail;
}
error = bus_dmamem_alloc(ring->seg0_dma_tag, (void **) &ring->seg0,
BUS_DMA_NOWAIT | BUS_DMA_ZERO, &ring->seg0_dma_map);
if (error != 0) {
device_printf(sc->dev,
"could not allocate Tx seg0 DMA memory\n");
goto fail;
}
error = bus_dmamap_load(ring->seg0_dma_tag, ring->seg0_dma_map,
ring->seg0,
RT_SOFTC_TX_RING_DATA_COUNT * RT_TX_DATA_SEG0_SIZE,
rt_dma_map_addr, &ring->seg0_phys_addr, 0);
if (error != 0) {
device_printf(sc->dev, "could not load Tx seg0 DMA map\n");
goto fail;
}
error = bus_dma_tag_create(bus_get_dma_tag(sc->dev), PAGE_SIZE, 0,
BUS_SPACE_MAXADDR_32BIT, BUS_SPACE_MAXADDR, NULL, NULL,
MJUMPAGESIZE, RT_SOFTC_MAX_SCATTER, MJUMPAGESIZE, 0, NULL, NULL,
&ring->data_dma_tag);
if (error != 0) {
device_printf(sc->dev,
"could not create Tx data DMA tag\n");
goto fail;
}
for (i = 0; i < RT_SOFTC_TX_RING_DATA_COUNT; i++) {
data = &ring->data[i];
error = bus_dmamap_create(ring->data_dma_tag, 0,
&data->dma_map);
if (error != 0) {
device_printf(sc->dev, "could not create Tx data DMA "
"map\n");
goto fail;
}
}
ring->data_queued = 0;
ring->data_cur = 0;
ring->data_next = 0;
ring->qid = qid;
return (0);
fail:
rt_free_tx_ring(sc, ring);
return (error);
}
/*
* rt_reset_tx_ring - reset TX ring buffer to empty state
*/
static void
rt_reset_tx_ring(struct rt_softc *sc, struct rt_softc_tx_ring *ring)
{
struct rt_softc_tx_data *data;
struct rt_txdesc *desc;
int i;
for (i = 0; i < RT_SOFTC_TX_RING_DESC_COUNT; i++) {
desc = &ring->desc[i];
desc->sdl0 = 0;
desc->sdl1 = 0;
}
ring->desc_queued = 0;
ring->desc_cur = 0;
ring->desc_next = 0;
bus_dmamap_sync(ring->desc_dma_tag, ring->desc_dma_map,
BUS_DMASYNC_PREWRITE);
bus_dmamap_sync(ring->seg0_dma_tag, ring->seg0_dma_map,
BUS_DMASYNC_PREWRITE);
for (i = 0; i < RT_SOFTC_TX_RING_DATA_COUNT; i++) {
data = &ring->data[i];
if (data->m != NULL) {
bus_dmamap_sync(ring->data_dma_tag, data->dma_map,
BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(ring->data_dma_tag, data->dma_map);
m_freem(data->m);
data->m = NULL;
}
}
ring->data_queued = 0;
ring->data_cur = 0;
ring->data_next = 0;
}
/*
* rt_free_tx_ring - free RX ring buffer
*/
static void
rt_free_tx_ring(struct rt_softc *sc, struct rt_softc_tx_ring *ring)
{
struct rt_softc_tx_data *data;
int i;
if (ring->desc != NULL) {
bus_dmamap_sync(ring->desc_dma_tag, ring->desc_dma_map,
BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(ring->desc_dma_tag, ring->desc_dma_map);
bus_dmamem_free(ring->desc_dma_tag, ring->desc,
ring->desc_dma_map);
}
if (ring->desc_dma_tag != NULL)
bus_dma_tag_destroy(ring->desc_dma_tag);
if (ring->seg0 != NULL) {
bus_dmamap_sync(ring->seg0_dma_tag, ring->seg0_dma_map,
BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(ring->seg0_dma_tag, ring->seg0_dma_map);
bus_dmamem_free(ring->seg0_dma_tag, ring->seg0,
ring->seg0_dma_map);
}
if (ring->seg0_dma_tag != NULL)
bus_dma_tag_destroy(ring->seg0_dma_tag);
for (i = 0; i < RT_SOFTC_TX_RING_DATA_COUNT; i++) {
data = &ring->data[i];
if (data->m != NULL) {
bus_dmamap_sync(ring->data_dma_tag, data->dma_map,
BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(ring->data_dma_tag, data->dma_map);
m_freem(data->m);
}
if (data->dma_map != NULL)
bus_dmamap_destroy(ring->data_dma_tag, data->dma_map);
}
if (ring->data_dma_tag != NULL)
bus_dma_tag_destroy(ring->data_dma_tag);
mtx_destroy(&ring->lock);
}
/*
* rt_dma_map_addr - get address of busdma segment
*/
static void
rt_dma_map_addr(void *arg, bus_dma_segment_t *segs, int nseg, int error)
{
if (error != 0)
return;
KASSERT(nseg == 1, ("too many DMA segments, %d should be 1", nseg));
*(bus_addr_t *) arg = segs[0].ds_addr;
}
/*
* rt_sysctl_attach - attach sysctl nodes for NIC counters.
*/
static void
rt_sysctl_attach(struct rt_softc *sc)
{
struct sysctl_ctx_list *ctx;
struct sysctl_oid *tree;
struct sysctl_oid *stats;
ctx = device_get_sysctl_ctx(sc->dev);
tree = device_get_sysctl_tree(sc->dev);
/* statistic counters */
stats = SYSCTL_ADD_NODE(ctx, SYSCTL_CHILDREN(tree), OID_AUTO,
"stats", CTLFLAG_RD, 0, "statistic");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"interrupts", CTLFLAG_RD, &sc->interrupts,
"all interrupts");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"tx_coherent_interrupts", CTLFLAG_RD, &sc->tx_coherent_interrupts,
"Tx coherent interrupts");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_coherent_interrupts", CTLFLAG_RD, &sc->rx_coherent_interrupts,
"Rx coherent interrupts");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_interrupts", CTLFLAG_RD, &sc->rx_interrupts[0],
"Rx interrupts");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_delay_interrupts", CTLFLAG_RD, &sc->rx_delay_interrupts,
"Rx delay interrupts");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ3_interrupts", CTLFLAG_RD, &sc->tx_interrupts[3],
"Tx AC3 interrupts");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ2_interrupts", CTLFLAG_RD, &sc->tx_interrupts[2],
"Tx AC2 interrupts");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ1_interrupts", CTLFLAG_RD, &sc->tx_interrupts[1],
"Tx AC1 interrupts");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ0_interrupts", CTLFLAG_RD, &sc->tx_interrupts[0],
"Tx AC0 interrupts");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"tx_delay_interrupts", CTLFLAG_RD, &sc->tx_delay_interrupts,
"Tx delay interrupts");
SYSCTL_ADD_INT(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ3_desc_queued", CTLFLAG_RD, &sc->tx_ring[3].desc_queued,
0, "Tx AC3 descriptors queued");
SYSCTL_ADD_INT(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ3_data_queued", CTLFLAG_RD, &sc->tx_ring[3].data_queued,
0, "Tx AC3 data queued");
SYSCTL_ADD_INT(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ2_desc_queued", CTLFLAG_RD, &sc->tx_ring[2].desc_queued,
0, "Tx AC2 descriptors queued");
SYSCTL_ADD_INT(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ2_data_queued", CTLFLAG_RD, &sc->tx_ring[2].data_queued,
0, "Tx AC2 data queued");
SYSCTL_ADD_INT(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ1_desc_queued", CTLFLAG_RD, &sc->tx_ring[1].desc_queued,
0, "Tx AC1 descriptors queued");
SYSCTL_ADD_INT(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ1_data_queued", CTLFLAG_RD, &sc->tx_ring[1].data_queued,
0, "Tx AC1 data queued");
SYSCTL_ADD_INT(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ0_desc_queued", CTLFLAG_RD, &sc->tx_ring[0].desc_queued,
0, "Tx AC0 descriptors queued");
SYSCTL_ADD_INT(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ0_data_queued", CTLFLAG_RD, &sc->tx_ring[0].data_queued,
0, "Tx AC0 data queued");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ3_data_queue_full", CTLFLAG_RD, &sc->tx_data_queue_full[3],
"Tx AC3 data queue full");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ2_data_queue_full", CTLFLAG_RD, &sc->tx_data_queue_full[2],
"Tx AC2 data queue full");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ1_data_queue_full", CTLFLAG_RD, &sc->tx_data_queue_full[1],
"Tx AC1 data queue full");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"TXQ0_data_queue_full", CTLFLAG_RD, &sc->tx_data_queue_full[0],
"Tx AC0 data queue full");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"tx_watchdog_timeouts", CTLFLAG_RD, &sc->tx_watchdog_timeouts,
"Tx watchdog timeouts");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"tx_defrag_packets", CTLFLAG_RD, &sc->tx_defrag_packets,
"Tx defragmented packets");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"no_tx_desc_avail", CTLFLAG_RD, &sc->no_tx_desc_avail,
"no Tx descriptors available");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_mbuf_alloc_errors", CTLFLAG_RD, &sc->rx_mbuf_alloc_errors,
"Rx mbuf allocation errors");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_mbuf_dmamap_errors", CTLFLAG_RD, &sc->rx_mbuf_dmamap_errors,
"Rx mbuf DMA mapping errors");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"tx_queue_0_not_empty", CTLFLAG_RD, &sc->tx_queue_not_empty[0],
"Tx queue 0 not empty");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"tx_queue_1_not_empty", CTLFLAG_RD, &sc->tx_queue_not_empty[1],
"Tx queue 1 not empty");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_packets", CTLFLAG_RD, &sc->rx_packets,
"Rx packets");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_crc_errors", CTLFLAG_RD, &sc->rx_crc_err,
"Rx CRC errors");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_phy_errors", CTLFLAG_RD, &sc->rx_phy_err,
"Rx PHY errors");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_dup_packets", CTLFLAG_RD, &sc->rx_dup_packets,
"Rx duplicate packets");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_fifo_overflows", CTLFLAG_RD, &sc->rx_fifo_overflows,
"Rx FIFO overflows");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_bytes", CTLFLAG_RD, &sc->rx_bytes,
"Rx bytes");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_long_err", CTLFLAG_RD, &sc->rx_long_err,
"Rx too long frame errors");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"rx_short_err", CTLFLAG_RD, &sc->rx_short_err,
"Rx too short frame errors");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"tx_bytes", CTLFLAG_RD, &sc->tx_bytes,
"Tx bytes");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"tx_packets", CTLFLAG_RD, &sc->tx_packets,
"Tx packets");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"tx_skip", CTLFLAG_RD, &sc->tx_skip,
"Tx skip count for GDMA ports");
SYSCTL_ADD_ULONG(ctx, SYSCTL_CHILDREN(stats), OID_AUTO,
"tx_collision", CTLFLAG_RD, &sc->tx_collision,
"Tx collision count for GDMA ports");
}
#if defined(IF_RT_PHY_SUPPORT) || defined(RT_MDIO)
/* This code is only work RT2880 and same chip. */
/* TODO: make RT3052 and later support code. But nobody need it? */
static int
rt_miibus_readreg(device_t dev, int phy, int reg)
{
struct rt_softc *sc = device_get_softc(dev);
int dat;
/*
* PSEUDO_PHYAD is a special value for indicate switch attached.
* No one PHY use PSEUDO_PHYAD (0x1e) address.
*/
#ifndef RT_MDIO
if (phy == 31) {
/* Fake PHY ID for bfeswitch attach */
switch (reg) {
case MII_BMSR:
return (BMSR_EXTSTAT|BMSR_MEDIAMASK);
case MII_PHYIDR1:
return (0x40); /* As result of faking */
case MII_PHYIDR2: /* PHY will detect as */
return (0x6250); /* bfeswitch */
}
}
#endif
/* Wait prev command done if any */
while (RT_READ(sc, MDIO_ACCESS) & MDIO_CMD_ONGO);
dat = ((phy << MDIO_PHY_ADDR_SHIFT) & MDIO_PHY_ADDR_MASK) |
((reg << MDIO_PHYREG_ADDR_SHIFT) & MDIO_PHYREG_ADDR_MASK);
RT_WRITE(sc, MDIO_ACCESS, dat);
RT_WRITE(sc, MDIO_ACCESS, dat | MDIO_CMD_ONGO);
while (RT_READ(sc, MDIO_ACCESS) & MDIO_CMD_ONGO);
return (RT_READ(sc, MDIO_ACCESS) & MDIO_PHY_DATA_MASK);
}
static int
rt_miibus_writereg(device_t dev, int phy, int reg, int val)
{
struct rt_softc *sc = device_get_softc(dev);
int dat;
/* Wait prev command done if any */
while (RT_READ(sc, MDIO_ACCESS) & MDIO_CMD_ONGO);
dat = MDIO_CMD_WR |
((phy << MDIO_PHY_ADDR_SHIFT) & MDIO_PHY_ADDR_MASK) |
((reg << MDIO_PHYREG_ADDR_SHIFT) & MDIO_PHYREG_ADDR_MASK) |
(val & MDIO_PHY_DATA_MASK);
RT_WRITE(sc, MDIO_ACCESS, dat);
RT_WRITE(sc, MDIO_ACCESS, dat | MDIO_CMD_ONGO);
while (RT_READ(sc, MDIO_ACCESS) & MDIO_CMD_ONGO);
return (0);
}
#endif
#ifdef IF_RT_PHY_SUPPORT
void
rt_miibus_statchg(device_t dev)
{
struct rt_softc *sc = device_get_softc(dev);
struct mii_data *mii;
mii = device_get_softc(sc->rt_miibus);
if ((mii->mii_media_status & (IFM_ACTIVE | IFM_AVALID)) ==
(IFM_ACTIVE | IFM_AVALID)) {
switch (IFM_SUBTYPE(mii->mii_media_active)) {
case IFM_10_T:
case IFM_100_TX:
/* XXX check link here */
sc->flags |= 1;
break;
default:
break;
}
}
}
#endif /* IF_RT_PHY_SUPPORT */
static device_method_t rt_dev_methods[] =
{
DEVMETHOD(device_probe, rt_probe),
DEVMETHOD(device_attach, rt_attach),
DEVMETHOD(device_detach, rt_detach),
DEVMETHOD(device_shutdown, rt_shutdown),
DEVMETHOD(device_suspend, rt_suspend),
DEVMETHOD(device_resume, rt_resume),
#ifdef IF_RT_PHY_SUPPORT
/* MII interface */
DEVMETHOD(miibus_readreg, rt_miibus_readreg),
DEVMETHOD(miibus_writereg, rt_miibus_writereg),
DEVMETHOD(miibus_statchg, rt_miibus_statchg),
#endif
DEVMETHOD_END
};
static driver_t rt_driver =
{
"rt",
rt_dev_methods,
sizeof(struct rt_softc)
};
static devclass_t rt_dev_class;
DRIVER_MODULE(rt, nexus, rt_driver, rt_dev_class, 0, 0);
#ifdef FDT
DRIVER_MODULE(rt, simplebus, rt_driver, rt_dev_class, 0, 0);
#endif
MODULE_DEPEND(rt, ether, 1, 1, 1);
MODULE_DEPEND(rt, miibus, 1, 1, 1);
#ifdef RT_MDIO
MODULE_DEPEND(rt, mdio, 1, 1, 1);
static int rtmdio_probe(device_t);
static int rtmdio_attach(device_t);
static int rtmdio_detach(device_t);
static struct mtx miibus_mtx;
MTX_SYSINIT(miibus_mtx, &miibus_mtx, "rt mii lock", MTX_DEF);
/*
* Declare an additional, separate driver for accessing the MDIO bus.
*/
static device_method_t rtmdio_methods[] = {
/* Device interface */
DEVMETHOD(device_probe, rtmdio_probe),
DEVMETHOD(device_attach, rtmdio_attach),
DEVMETHOD(device_detach, rtmdio_detach),
/* bus interface */
DEVMETHOD(bus_add_child, device_add_child_ordered),
/* MDIO access */
DEVMETHOD(mdio_readreg, rt_miibus_readreg),
DEVMETHOD(mdio_writereg, rt_miibus_writereg),
};
DEFINE_CLASS_0(rtmdio, rtmdio_driver, rtmdio_methods,
sizeof(struct rt_softc));
static devclass_t rtmdio_devclass;
DRIVER_MODULE(miiproxy, rt, miiproxy_driver, miiproxy_devclass, 0, 0);
DRIVER_MODULE(rtmdio, simplebus, rtmdio_driver, rtmdio_devclass, 0, 0);
DRIVER_MODULE(mdio, rtmdio, mdio_driver, mdio_devclass, 0, 0);
static int
rtmdio_probe(device_t dev)
{
if (!ofw_bus_status_okay(dev))
return (ENXIO);
if (!ofw_bus_is_compatible(dev, "ralink,rt2880-mdio"))
return (ENXIO);
device_set_desc(dev, "FV built-in ethernet interface, MDIO controller");
return(0);
}
static int
rtmdio_attach(device_t dev)
{
struct rt_softc *sc;
int error;
sc = device_get_softc(dev);
sc->dev = dev;
sc->mem_rid = 0;
sc->mem = bus_alloc_resource_any(dev, SYS_RES_MEMORY,
&sc->mem_rid, RF_ACTIVE | RF_SHAREABLE);
if (sc->mem == NULL) {
device_printf(dev, "couldn't map memory\n");
error = ENXIO;
goto fail;
}
sc->bst = rman_get_bustag(sc->mem);
sc->bsh = rman_get_bushandle(sc->mem);
bus_generic_probe(dev);
bus_enumerate_hinted_children(dev);
error = bus_generic_attach(dev);
fail:
return(error);
}
static int
rtmdio_detach(device_t dev)
{
return(0);
}
#endif