freebsd-skq/sys/dev/gem/if_gemvar.h
marius 6aa8875fe3 Revert the part of r194763 which added a dying flag and instead
call ether_ifdetach(9) before stopping the controller and the
callouts. The consensus is that the latter is now safe to do and
should also solve the problem of active BPF listeners clearing
promiscuous mode can result in the tick callout being restarted
which in turn will trigger a panic once it's actually gone.
2009-06-24 19:04:08 +00:00

283 lines
9.6 KiB
C

/*-
* Copyright (C) 2001 Eduardo Horvath.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*
* from: NetBSD: gemvar.h,v 1.8 2002/05/15 02:36:12 matt Exp
*
* $FreeBSD$
*/
#ifndef _IF_GEMVAR_H
#define _IF_GEMVAR_H
#include <sys/queue.h>
#include <sys/callout.h>
/*
* Transmit descriptor ring size - this is arbitrary, but allocate
* enough descriptors for 64 pending transmissions and 16 segments
* per packet. This limit is not actually enforced (packets with
* more segments can be sent, depending on the busdma backend); it
* is however used as an estimate for the TX window size.
*/
#define GEM_NTXSEGS 16
#define GEM_TXQUEUELEN 64
#define GEM_NTXDESC (GEM_TXQUEUELEN * GEM_NTXSEGS)
#define GEM_MAXTXFREE (GEM_NTXDESC - 1)
#define GEM_NTXDESC_MASK (GEM_NTXDESC - 1)
#define GEM_NEXTTX(x) ((x + 1) & GEM_NTXDESC_MASK)
/*
* Receive descriptor ring size - we have one RX buffer per incoming
* packet, so this logic is a little simpler.
*/
#define GEM_NRXDESC 256
#define GEM_NRXDESC_MASK (GEM_NRXDESC - 1)
#define GEM_NEXTRX(x) ((x + 1) & GEM_NRXDESC_MASK)
/*
* How many ticks to wait until to retry on a RX descriptor that is
* still owned by the hardware.
*/
#define GEM_RXOWN_TICKS (hz / 50)
/*
* Control structures are DMA'd to the chip. We allocate them
* in a single clump that maps to a single DMA segment to make
* several things easier.
*/
struct gem_control_data {
struct gem_desc gcd_txdescs[GEM_NTXDESC]; /* TX descriptors */
struct gem_desc gcd_rxdescs[GEM_NRXDESC]; /* RX descriptors */
};
#define GEM_CDOFF(x) offsetof(struct gem_control_data, x)
#define GEM_CDTXOFF(x) GEM_CDOFF(gcd_txdescs[(x)])
#define GEM_CDRXOFF(x) GEM_CDOFF(gcd_rxdescs[(x)])
/*
* software state for transmit job mbufs (may be elements of mbuf chains)
*/
struct gem_txsoft {
struct mbuf *txs_mbuf; /* head of our mbuf chain */
bus_dmamap_t txs_dmamap; /* our DMA map */
u_int txs_firstdesc; /* first descriptor in packet */
u_int txs_lastdesc; /* last descriptor in packet */
u_int txs_ndescs; /* number of descriptors */
STAILQ_ENTRY(gem_txsoft) txs_q;
};
STAILQ_HEAD(gem_txsq, gem_txsoft);
/*
* software state for receive jobs
*/
struct gem_rxsoft {
struct mbuf *rxs_mbuf; /* head of our mbuf chain */
bus_dmamap_t rxs_dmamap; /* our DMA map */
bus_addr_t rxs_paddr; /* physical address of the segment */
};
/*
* software state per device
*/
struct gem_softc {
struct ifnet *sc_ifp;
struct mtx sc_mtx;
device_t sc_miibus;
struct mii_data *sc_mii; /* MII media control */
device_t sc_dev; /* generic device information */
u_char sc_enaddr[ETHER_ADDR_LEN];
struct callout sc_tick_ch; /* tick callout */
struct callout sc_rx_ch; /* delayed RX callout */
u_int sc_wdog_timer; /* watchdog timer */
void *sc_ih;
struct resource *sc_res[3];
#define GEM_RES_INTR 0
#define GEM_RES_BANK1 1
#define GEM_RES_BANK2 2
bus_dma_tag_t sc_pdmatag; /* parent bus DMA tag */
bus_dma_tag_t sc_rdmatag; /* RX bus DMA tag */
bus_dma_tag_t sc_tdmatag; /* TX bus DMA tag */
bus_dma_tag_t sc_cdmatag; /* control data bus DMA tag */
bus_dmamap_t sc_dmamap; /* bus DMA handle */
int sc_phyad; /* PHY to use or -1 for any */
u_int sc_variant;
#define GEM_UNKNOWN 0 /* don't know */
#define GEM_SUN_GEM 1 /* Sun GEM */
#define GEM_SUN_ERI 2 /* Sun ERI */
#define GEM_APPLE_GMAC 3 /* Apple GMAC */
#define GEM_APPLE_K2_GMAC 4 /* Apple K2 GMAC */
#define GEM_IS_APPLE(sc) \
((sc)->sc_variant == GEM_APPLE_GMAC || \
(sc)->sc_variant == GEM_APPLE_K2_GMAC)
u_int sc_flags;
#define GEM_INITED (1 << 0) /* reset persistent regs init'ed */
#define GEM_LINK (1 << 1) /* link is up */
#define GEM_PCI (1 << 2) /* PCI busses are little-endian */
#define GEM_PCI66 (1 << 3) /* PCI bus runs at 66MHz */
#define GEM_SERDES (1 << 4) /* use the SERDES */
/*
* ring buffer DMA stuff
*/
bus_dmamap_t sc_cddmamap; /* control data DMA map */
bus_addr_t sc_cddma;
/*
* software state for transmit and receive descriptors
*/
struct gem_txsoft sc_txsoft[GEM_TXQUEUELEN];
struct gem_rxsoft sc_rxsoft[GEM_NRXDESC];
/*
* control data structures
*/
struct gem_control_data *sc_control_data;
#define sc_txdescs sc_control_data->gcd_txdescs
#define sc_rxdescs sc_control_data->gcd_rxdescs
u_int sc_txfree; /* number of free TX descriptors */
u_int sc_txnext; /* next ready TX descriptor */
u_int sc_txwin; /* TX desc. since last TX intr. */
struct gem_txsq sc_txfreeq; /* free TX descsofts */
struct gem_txsq sc_txdirtyq; /* dirty TX descsofts */
u_int sc_rxptr; /* next ready RX descriptor/state */
u_int sc_rxfifosize; /* RX FIFO size (bytes) */
int sc_ifflags;
u_long sc_csum_features;
};
#define GEM_BANKN_BARRIER(n, sc, offs, len, flags) \
bus_barrier((sc)->sc_res[(n)], (offs), (len), (flags))
#define GEM_BANK1_BARRIER(sc, offs, len, flags) \
GEM_BANKN_BARRIER(GEM_RES_BANK1, (sc), (offs), (len), (flags))
#define GEM_BANK2_BARRIER(sc, offs, len, flags) \
GEM_BANKN_BARRIER(GEM_RES_BANK2, (sc), (offs), (len), (flags))
#define GEM_BANKN_READ_M(n, m, sc, offs) \
bus_read_ ## m((sc)->sc_res[(n)], (offs))
#define GEM_BANK1_READ_1(sc, offs) \
GEM_BANKN_READ_M(GEM_RES_BANK1, 1, (sc), (offs))
#define GEM_BANK1_READ_2(sc, offs) \
GEM_BANKN_READ_M(GEM_RES_BANK1, 2, (sc), (offs))
#define GEM_BANK1_READ_4(sc, offs) \
GEM_BANKN_READ_M(GEM_RES_BANK1, 4, (sc), (offs))
#define GEM_BANK2_READ_1(sc, offs) \
GEM_BANKN_READ_M(GEM_RES_BANK2, 1, (sc), (offs))
#define GEM_BANK2_READ_2(sc, offs) \
GEM_BANKN_READ_M(GEM_RES_BANK2, 2, (sc), (offs))
#define GEM_BANK2_READ_4(sc, offs) \
GEM_BANKN_READ_M(GEM_RES_BANK2, 4, (sc), (offs))
#define GEM_BANKN_WRITE_M(n, m, sc, offs, v) \
bus_write_ ## m((sc)->sc_res[n], (offs), (v))
#define GEM_BANK1_WRITE_1(sc, offs, v) \
GEM_BANKN_WRITE_M(GEM_RES_BANK1, 1, (sc), (offs), (v))
#define GEM_BANK1_WRITE_2(sc, offs, v) \
GEM_BANKN_WRITE_M(GEM_RES_BANK1, 2, (sc), (offs), (v))
#define GEM_BANK1_WRITE_4(sc, offs, v) \
GEM_BANKN_WRITE_M(GEM_RES_BANK1, 4, (sc), (offs), (v))
#define GEM_BANK2_WRITE_1(sc, offs, v) \
GEM_BANKN_WRITE_M(GEM_RES_BANK2, 1, (sc), (offs), (v))
#define GEM_BANK2_WRITE_2(sc, offs, v) \
GEM_BANKN_WRITE_M(GEM_RES_BANK2, 2, (sc), (offs), (v))
#define GEM_BANK2_WRITE_4(sc, offs, v) \
GEM_BANKN_WRITE_M(GEM_RES_BANK2, 4, (sc), (offs), (v))
/* XXX this should be handled by bus_dma(9). */
#define GEM_DMA_READ(sc, v) \
((((sc)->sc_flags & GEM_PCI) != 0) ? le64toh(v) : be64toh(v))
#define GEM_DMA_WRITE(sc, v) \
((((sc)->sc_flags & GEM_PCI) != 0) ? htole64(v) : htobe64(v))
#define GEM_CDTXADDR(sc, x) ((sc)->sc_cddma + GEM_CDTXOFF((x)))
#define GEM_CDRXADDR(sc, x) ((sc)->sc_cddma + GEM_CDRXOFF((x)))
#define GEM_CDSYNC(sc, ops) \
bus_dmamap_sync((sc)->sc_cdmatag, (sc)->sc_cddmamap, (ops));
#define GEM_INIT_RXDESC(sc, x) \
do { \
struct gem_rxsoft *__rxs = &sc->sc_rxsoft[(x)]; \
struct gem_desc *__rxd = &sc->sc_rxdescs[(x)]; \
struct mbuf *__m = __rxs->rxs_mbuf; \
\
__m->m_data = __m->m_ext.ext_buf; \
__rxd->gd_addr = \
GEM_DMA_WRITE((sc), __rxs->rxs_paddr); \
__rxd->gd_flags = GEM_DMA_WRITE((sc), \
(((__m->m_ext.ext_size) << GEM_RD_BUFSHIFT) & \
GEM_RD_BUFSIZE) | GEM_RD_OWN); \
} while (0)
#define GEM_UPDATE_RXDESC(sc, x) \
do { \
struct gem_rxsoft *__rxs = &sc->sc_rxsoft[(x)]; \
struct gem_desc *__rxd = &sc->sc_rxdescs[(x)]; \
struct mbuf *__m = __rxs->rxs_mbuf; \
\
__rxd->gd_flags = GEM_DMA_WRITE((sc), \
(((__m->m_ext.ext_size) << GEM_RD_BUFSHIFT) & \
GEM_RD_BUFSIZE) | GEM_RD_OWN); \
} while (0)
#define GEM_LOCK_INIT(_sc, _name) \
mtx_init(&(_sc)->sc_mtx, _name, MTX_NETWORK_LOCK, MTX_DEF)
#define GEM_LOCK(_sc) mtx_lock(&(_sc)->sc_mtx)
#define GEM_UNLOCK(_sc) mtx_unlock(&(_sc)->sc_mtx)
#define GEM_LOCK_ASSERT(_sc, _what) mtx_assert(&(_sc)->sc_mtx, (_what))
#define GEM_LOCK_DESTROY(_sc) mtx_destroy(&(_sc)->sc_mtx)
#ifdef _KERNEL
extern devclass_t gem_devclass;
int gem_attach(struct gem_softc *sc);
void gem_detach(struct gem_softc *sc);
void gem_intr(void *v);
void gem_resume(struct gem_softc *sc);
void gem_suspend(struct gem_softc *sc);
int gem_mediachange(struct ifnet *ifp);
void gem_mediastatus(struct ifnet *ifp, struct ifmediareq *ifmr);
/* MII methods & callbacks */
int gem_mii_readreg(device_t dev, int phy, int reg);
void gem_mii_statchg(device_t dev);
int gem_mii_writereg(device_t dev, int phy, int reg, int val);
#endif /* _KERNEL */
#endif