c0e3e9d4d2
get it out of a stuck condition that can be caused by GEM_MAC_RX_OVERFLOW. - In gem_reset_rxdma() call gem_setladrf() in order to reprogram the RX filter and restore the previous content of GEM_MAC_RX_CONFIG. While at it consistently use the newly introduced sc_mac_rxcfg throughout the driver instead of reading the its old content. - Increment if_iqdrops instead of if_ierrors in case of RX buffer allocation failure. - According to the GEM datasheet the RX MAC should also be disabled in gem_setladrf() before changing its configuration. - Add error messages to gem_disable_{r,t}x() and take advantage of these throughout the driver instead of duplicating their functionality all over the place. In joint forces with: yongari
283 lines
9.6 KiB
C
283 lines
9.6 KiB
C
/*-
|
|
* Copyright (C) 2001 Eduardo Horvath.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* from: NetBSD: gemvar.h,v 1.8 2002/05/15 02:36:12 matt Exp
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
#ifndef _IF_GEMVAR_H
|
|
#define _IF_GEMVAR_H
|
|
|
|
#include <sys/queue.h>
|
|
#include <sys/callout.h>
|
|
|
|
/*
|
|
* Transmit descriptor ring size - this is arbitrary, but allocate
|
|
* enough descriptors for 64 pending transmissions and 16 segments
|
|
* per packet. This limit is not actually enforced (packets with
|
|
* more segments can be sent, depending on the busdma backend); it
|
|
* is however used as an estimate for the TX window size.
|
|
*/
|
|
#define GEM_NTXSEGS 16
|
|
|
|
#define GEM_TXQUEUELEN 64
|
|
#define GEM_NTXDESC (GEM_TXQUEUELEN * GEM_NTXSEGS)
|
|
#define GEM_MAXTXFREE (GEM_NTXDESC - 1)
|
|
#define GEM_NTXDESC_MASK (GEM_NTXDESC - 1)
|
|
#define GEM_NEXTTX(x) ((x + 1) & GEM_NTXDESC_MASK)
|
|
|
|
/*
|
|
* Receive descriptor ring size - we have one RX buffer per incoming
|
|
* packet, so this logic is a little simpler.
|
|
*/
|
|
#define GEM_NRXDESC 256
|
|
#define GEM_NRXDESC_MASK (GEM_NRXDESC - 1)
|
|
#define GEM_NEXTRX(x) ((x + 1) & GEM_NRXDESC_MASK)
|
|
|
|
/*
|
|
* How many ticks to wait until to retry on a RX descriptor that is
|
|
* still owned by the hardware.
|
|
*/
|
|
#define GEM_RXOWN_TICKS (hz / 50)
|
|
|
|
/*
|
|
* Control structures are DMA'd to the chip. We allocate them
|
|
* in a single clump that maps to a single DMA segment to make
|
|
* several things easier.
|
|
*/
|
|
struct gem_control_data {
|
|
struct gem_desc gcd_txdescs[GEM_NTXDESC]; /* TX descriptors */
|
|
struct gem_desc gcd_rxdescs[GEM_NRXDESC]; /* RX descriptors */
|
|
};
|
|
|
|
#define GEM_CDOFF(x) offsetof(struct gem_control_data, x)
|
|
#define GEM_CDTXOFF(x) GEM_CDOFF(gcd_txdescs[(x)])
|
|
#define GEM_CDRXOFF(x) GEM_CDOFF(gcd_rxdescs[(x)])
|
|
|
|
/*
|
|
* software state for transmit job mbufs (may be elements of mbuf chains)
|
|
*/
|
|
struct gem_txsoft {
|
|
struct mbuf *txs_mbuf; /* head of our mbuf chain */
|
|
bus_dmamap_t txs_dmamap; /* our DMA map */
|
|
u_int txs_firstdesc; /* first descriptor in packet */
|
|
u_int txs_lastdesc; /* last descriptor in packet */
|
|
u_int txs_ndescs; /* number of descriptors */
|
|
STAILQ_ENTRY(gem_txsoft) txs_q;
|
|
};
|
|
|
|
STAILQ_HEAD(gem_txsq, gem_txsoft);
|
|
|
|
/*
|
|
* software state for receive jobs
|
|
*/
|
|
struct gem_rxsoft {
|
|
struct mbuf *rxs_mbuf; /* head of our mbuf chain */
|
|
bus_dmamap_t rxs_dmamap; /* our DMA map */
|
|
bus_addr_t rxs_paddr; /* physical address of the segment */
|
|
};
|
|
|
|
/*
|
|
* software state per device
|
|
*/
|
|
struct gem_softc {
|
|
struct ifnet *sc_ifp;
|
|
struct mtx sc_mtx;
|
|
device_t sc_miibus;
|
|
struct mii_data *sc_mii; /* MII media control */
|
|
device_t sc_dev; /* generic device information */
|
|
u_char sc_enaddr[ETHER_ADDR_LEN];
|
|
struct callout sc_tick_ch; /* tick callout */
|
|
struct callout sc_rx_ch; /* delayed RX callout */
|
|
u_int sc_wdog_timer; /* watchdog timer */
|
|
|
|
void *sc_ih;
|
|
struct resource *sc_res[3];
|
|
#define GEM_RES_INTR 0
|
|
#define GEM_RES_BANK1 1
|
|
#define GEM_RES_BANK2 2
|
|
|
|
bus_dma_tag_t sc_pdmatag; /* parent bus DMA tag */
|
|
bus_dma_tag_t sc_rdmatag; /* RX bus DMA tag */
|
|
bus_dma_tag_t sc_tdmatag; /* TX bus DMA tag */
|
|
bus_dma_tag_t sc_cdmatag; /* control data bus DMA tag */
|
|
bus_dmamap_t sc_dmamap; /* bus DMA handle */
|
|
|
|
u_int sc_variant;
|
|
#define GEM_UNKNOWN 0 /* don't know */
|
|
#define GEM_SUN_GEM 1 /* Sun GEM */
|
|
#define GEM_SUN_ERI 2 /* Sun ERI */
|
|
#define GEM_APPLE_GMAC 3 /* Apple GMAC */
|
|
#define GEM_APPLE_K2_GMAC 4 /* Apple K2 GMAC */
|
|
|
|
#define GEM_IS_APPLE(sc) \
|
|
((sc)->sc_variant == GEM_APPLE_GMAC || \
|
|
(sc)->sc_variant == GEM_APPLE_K2_GMAC)
|
|
|
|
u_int sc_flags;
|
|
#define GEM_INITED (1 << 0) /* reset persistent regs init'ed */
|
|
#define GEM_LINK (1 << 1) /* link is up */
|
|
#define GEM_PCI (1 << 2) /* PCI busses are little-endian */
|
|
#define GEM_PCI66 (1 << 3) /* PCI bus runs at 66MHz */
|
|
#define GEM_SERDES (1 << 4) /* use the SERDES */
|
|
|
|
/*
|
|
* ring buffer DMA stuff
|
|
*/
|
|
bus_dmamap_t sc_cddmamap; /* control data DMA map */
|
|
bus_addr_t sc_cddma;
|
|
|
|
/*
|
|
* software state for transmit and receive descriptors
|
|
*/
|
|
struct gem_txsoft sc_txsoft[GEM_TXQUEUELEN];
|
|
struct gem_rxsoft sc_rxsoft[GEM_NRXDESC];
|
|
|
|
/*
|
|
* control data structures
|
|
*/
|
|
struct gem_control_data *sc_control_data;
|
|
#define sc_txdescs sc_control_data->gcd_txdescs
|
|
#define sc_rxdescs sc_control_data->gcd_rxdescs
|
|
|
|
u_int sc_txfree; /* number of free TX descriptors */
|
|
u_int sc_txnext; /* next ready TX descriptor */
|
|
u_int sc_txwin; /* TX desc. since last TX intr. */
|
|
|
|
struct gem_txsq sc_txfreeq; /* free TX descsofts */
|
|
struct gem_txsq sc_txdirtyq; /* dirty TX descsofts */
|
|
|
|
u_int sc_rxptr; /* next ready RX descriptor/state */
|
|
u_int sc_rxfifosize; /* RX FIFO size (bytes) */
|
|
|
|
uint32_t sc_mac_rxcfg; /* RX MAC conf. % GEM_MAC_RX_ENABLE */
|
|
|
|
int sc_ifflags;
|
|
u_long sc_csum_features;
|
|
};
|
|
|
|
#define GEM_BANKN_BARRIER(n, sc, offs, len, flags) \
|
|
bus_barrier((sc)->sc_res[(n)], (offs), (len), (flags))
|
|
#define GEM_BANK1_BARRIER(sc, offs, len, flags) \
|
|
GEM_BANKN_BARRIER(GEM_RES_BANK1, (sc), (offs), (len), (flags))
|
|
#define GEM_BANK2_BARRIER(sc, offs, len, flags) \
|
|
GEM_BANKN_BARRIER(GEM_RES_BANK2, (sc), (offs), (len), (flags))
|
|
|
|
#define GEM_BANKN_READ_M(n, m, sc, offs) \
|
|
bus_read_ ## m((sc)->sc_res[(n)], (offs))
|
|
#define GEM_BANK1_READ_1(sc, offs) \
|
|
GEM_BANKN_READ_M(GEM_RES_BANK1, 1, (sc), (offs))
|
|
#define GEM_BANK1_READ_2(sc, offs) \
|
|
GEM_BANKN_READ_M(GEM_RES_BANK1, 2, (sc), (offs))
|
|
#define GEM_BANK1_READ_4(sc, offs) \
|
|
GEM_BANKN_READ_M(GEM_RES_BANK1, 4, (sc), (offs))
|
|
#define GEM_BANK2_READ_1(sc, offs) \
|
|
GEM_BANKN_READ_M(GEM_RES_BANK2, 1, (sc), (offs))
|
|
#define GEM_BANK2_READ_2(sc, offs) \
|
|
GEM_BANKN_READ_M(GEM_RES_BANK2, 2, (sc), (offs))
|
|
#define GEM_BANK2_READ_4(sc, offs) \
|
|
GEM_BANKN_READ_M(GEM_RES_BANK2, 4, (sc), (offs))
|
|
|
|
#define GEM_BANKN_WRITE_M(n, m, sc, offs, v) \
|
|
bus_write_ ## m((sc)->sc_res[n], (offs), (v))
|
|
#define GEM_BANK1_WRITE_1(sc, offs, v) \
|
|
GEM_BANKN_WRITE_M(GEM_RES_BANK1, 1, (sc), (offs), (v))
|
|
#define GEM_BANK1_WRITE_2(sc, offs, v) \
|
|
GEM_BANKN_WRITE_M(GEM_RES_BANK1, 2, (sc), (offs), (v))
|
|
#define GEM_BANK1_WRITE_4(sc, offs, v) \
|
|
GEM_BANKN_WRITE_M(GEM_RES_BANK1, 4, (sc), (offs), (v))
|
|
#define GEM_BANK2_WRITE_1(sc, offs, v) \
|
|
GEM_BANKN_WRITE_M(GEM_RES_BANK2, 1, (sc), (offs), (v))
|
|
#define GEM_BANK2_WRITE_2(sc, offs, v) \
|
|
GEM_BANKN_WRITE_M(GEM_RES_BANK2, 2, (sc), (offs), (v))
|
|
#define GEM_BANK2_WRITE_4(sc, offs, v) \
|
|
GEM_BANKN_WRITE_M(GEM_RES_BANK2, 4, (sc), (offs), (v))
|
|
|
|
/* XXX this should be handled by bus_dma(9). */
|
|
#define GEM_DMA_READ(sc, v) \
|
|
((((sc)->sc_flags & GEM_PCI) != 0) ? le64toh(v) : be64toh(v))
|
|
#define GEM_DMA_WRITE(sc, v) \
|
|
((((sc)->sc_flags & GEM_PCI) != 0) ? htole64(v) : htobe64(v))
|
|
|
|
#define GEM_CDTXADDR(sc, x) ((sc)->sc_cddma + GEM_CDTXOFF((x)))
|
|
#define GEM_CDRXADDR(sc, x) ((sc)->sc_cddma + GEM_CDRXOFF((x)))
|
|
|
|
#define GEM_CDSYNC(sc, ops) \
|
|
bus_dmamap_sync((sc)->sc_cdmatag, (sc)->sc_cddmamap, (ops));
|
|
|
|
#define GEM_INIT_RXDESC(sc, x) \
|
|
do { \
|
|
struct gem_rxsoft *__rxs = &sc->sc_rxsoft[(x)]; \
|
|
struct gem_desc *__rxd = &sc->sc_rxdescs[(x)]; \
|
|
struct mbuf *__m = __rxs->rxs_mbuf; \
|
|
\
|
|
__m->m_data = __m->m_ext.ext_buf; \
|
|
__rxd->gd_addr = \
|
|
GEM_DMA_WRITE((sc), __rxs->rxs_paddr); \
|
|
__rxd->gd_flags = GEM_DMA_WRITE((sc), \
|
|
(((__m->m_ext.ext_size) << GEM_RD_BUFSHIFT) & \
|
|
GEM_RD_BUFSIZE) | GEM_RD_OWN); \
|
|
} while (0)
|
|
|
|
#define GEM_UPDATE_RXDESC(sc, x) \
|
|
do { \
|
|
struct gem_rxsoft *__rxs = &sc->sc_rxsoft[(x)]; \
|
|
struct gem_desc *__rxd = &sc->sc_rxdescs[(x)]; \
|
|
struct mbuf *__m = __rxs->rxs_mbuf; \
|
|
\
|
|
__rxd->gd_flags = GEM_DMA_WRITE((sc), \
|
|
(((__m->m_ext.ext_size) << GEM_RD_BUFSHIFT) & \
|
|
GEM_RD_BUFSIZE) | GEM_RD_OWN); \
|
|
} while (0)
|
|
|
|
#define GEM_LOCK_INIT(_sc, _name) \
|
|
mtx_init(&(_sc)->sc_mtx, _name, MTX_NETWORK_LOCK, MTX_DEF)
|
|
#define GEM_LOCK(_sc) mtx_lock(&(_sc)->sc_mtx)
|
|
#define GEM_UNLOCK(_sc) mtx_unlock(&(_sc)->sc_mtx)
|
|
#define GEM_LOCK_ASSERT(_sc, _what) mtx_assert(&(_sc)->sc_mtx, (_what))
|
|
#define GEM_LOCK_DESTROY(_sc) mtx_destroy(&(_sc)->sc_mtx)
|
|
|
|
#ifdef _KERNEL
|
|
extern devclass_t gem_devclass;
|
|
|
|
int gem_attach(struct gem_softc *sc);
|
|
void gem_detach(struct gem_softc *sc);
|
|
void gem_intr(void *v);
|
|
void gem_resume(struct gem_softc *sc);
|
|
void gem_suspend(struct gem_softc *sc);
|
|
|
|
int gem_mediachange(struct ifnet *ifp);
|
|
void gem_mediastatus(struct ifnet *ifp, struct ifmediareq *ifmr);
|
|
|
|
/* MII methods & callbacks */
|
|
int gem_mii_readreg(device_t dev, int phy, int reg);
|
|
void gem_mii_statchg(device_t dev);
|
|
int gem_mii_writereg(device_t dev, int phy, int reg, int val);
|
|
|
|
#endif /* _KERNEL */
|
|
|
|
#endif
|