37e1f78fb7
o Disable HMEDEBUG by default. o Add CTASSERTs ensuring that HME_NRXDESC and HME_NTXDESC are set to legal values. o Use appropriate maxsize, nsegments and maxsegsize parameters when creating DMA tags and correct some comments related to them. o The FreeBSD bus_dmamap_sync(9) supports ored together flags for quite some time now so collapse calls accordingly. o Add missing BUS_DMASYNC_PREREAD when syncing the control DMA maps in hme_rint() and hme_start_locked(). o Keep state of the link state and use it to enable or disable the MAC in hme_mii_statchg() accordingly as well as to return early from hme_start_locked() in case the link is down. o Introduce a sc_flags and use it to replace individual members like sc_pci. o Add bus_barrier(9) calls to hme_mac_bitflip(), hme_mii_readreg(), hme_mii_writereg() and hme_stop() to ensure the respective bit has been written before we starting polling on it and for the right bits to change. o Rather just returning in case hme_mac_bitflip() fails and leaving us in an undefined state report the problem and move on; chances are the requested configuration will become active shortly after. o Don't call hme_start_locked() in hme_init_locked() unconditionally but only after calls to hme_init_locked() when it's appropriate, i.e. in hme_watchdog(). o Add a KASSERT which asserts nsegs is valid also to hme_load_txmbuf(). o In hme_load_txmbuf(): - use a maximum of the newly introduced HME_NTXSEGS segments instead of the incorrect HME_NTXQ, which reflects the maximum TX queue length, for loading the mbufs and put the DMA segments back onto the stack instead of the softc as 16 should be ok there. - use the common errno(2) return values instead of homegrown ones, - given that hme_load_txmbuf() is allowed to fail resulting in a packet drop for quite some time now implement the functionality of hme_txcksum() by means of m_pullup(9), which de-obfuscates the code and allows to always retrieve the correct length of the IP header, [1] - also add a KASSERT which asserts nsegs is valid, - take advantage of m_collapse(9) instead of m_defrag(9) for performance reasons. o Don't bother to check whether the interface is running or whether its queue is empty before calling hme_start_locked() in hme_tint(), the former will check these anyway. o In hme_intr() call hme_rint() before hme_tint() as gem_tint() may take quite a while to return when it calls hme_start_locked(). o Get rid of sc_debug and just check if_flags for IFF_DEBUG directly. o Add a shadow sc_ifflags so we don't reset the chip when unnecessary. o Handle IFF_ALLMULTI correctly. [2] o Use PCIR_BAR instead of a homegrown macro. o Replace sc_enaddr[6] with sc_enaddr[ETHER_ADDR_LEN]. o Use the maximum of 256 TX descriptors for better performance as using all of them has no additional static cost rather than using just half of them. Reported by: rwatson [2] Suggested by: yongari [1] Reviewed by: yongari
166 lines
5.8 KiB
C
166 lines
5.8 KiB
C
/*-
|
|
* Copyright (c) 1999 The NetBSD Foundation, Inc.
|
|
* All rights reserved.
|
|
*
|
|
* This code is derived from software contributed to The NetBSD Foundation
|
|
* by Paul Kranenburg.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed by the NetBSD
|
|
* Foundation, Inc. and its contributors.
|
|
* 4. Neither the name of The NetBSD Foundation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
|
|
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
|
|
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
|
|
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
* from: NetBSD: hmevar.h,v 1.5 2000/06/25 01:10:04 eeh Exp
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
#include <sys/callout.h>
|
|
|
|
/*
|
|
* Number of receive and transmit descriptors. For each receive descriptor,
|
|
* an mbuf cluster is allocated and set up to receive a packet, and a dma map
|
|
* is created. Therefore, this number should not be too high to not waste
|
|
* memory.
|
|
* TX descriptors have no static cost, except for the memory directly allocated
|
|
* for them. TX queue elements (the number of which is fixed by HME_NTXQ) hold
|
|
* the software state for a transmit job; each has a dmamap allocated for it.
|
|
* There may be multiple descriptors allocated to a single queue element.
|
|
* HME_NTXQ and HME_NTXSEGS are completely arbitrary.
|
|
*/
|
|
#define HME_NRXDESC 128
|
|
#define HME_NTXDESC 256
|
|
#define HME_NTXQ 64
|
|
#define HME_NTXSEGS 16
|
|
|
|
/* Maximum size of a mapped RX buffer. */
|
|
#define HME_BUFSZ 1600
|
|
|
|
/*
|
|
* RX DMA descriptor. The descriptors are preallocated; the dma map is
|
|
* reused.
|
|
*/
|
|
struct hme_rxdesc {
|
|
struct mbuf *hrx_m;
|
|
bus_dmamap_t hrx_dmamap;
|
|
};
|
|
|
|
/* Lazily leave at least one burst size grace space. */
|
|
#define HME_DESC_RXLEN(sc, d) \
|
|
ulmin(HME_BUFSZ, (d)->hrx_m->m_len - (sc)->sc_burst)
|
|
|
|
struct hme_txdesc {
|
|
struct mbuf *htx_m;
|
|
bus_dmamap_t htx_dmamap;
|
|
int htx_lastdesc;
|
|
STAILQ_ENTRY(hme_txdesc) htx_q;
|
|
};
|
|
|
|
STAILQ_HEAD(hme_txdq, hme_txdesc);
|
|
|
|
struct hme_ring {
|
|
/* Ring Descriptors */
|
|
caddr_t rb_membase; /* Packet buffer: CPU address */
|
|
bus_addr_t rb_dmabase; /* Packet buffer: DMA address */
|
|
caddr_t rb_txd; /* Transmit descriptors */
|
|
bus_addr_t rb_txddma; /* DMA address of same */
|
|
caddr_t rb_rxd; /* Receive descriptors */
|
|
bus_addr_t rb_rxddma; /* DMA address of same */
|
|
|
|
/* Ring Descriptor state */
|
|
int rb_tdhead, rb_tdtail;
|
|
int rb_rdtail;
|
|
int rb_td_nbusy;
|
|
|
|
/* Descriptors */
|
|
struct hme_rxdesc rb_rxdesc[HME_NRXDESC];
|
|
struct hme_txdesc rb_txdesc[HME_NTXQ];
|
|
|
|
struct hme_txdq rb_txfreeq;
|
|
struct hme_txdq rb_txbusyq;
|
|
|
|
bus_dmamap_t rb_spare_dmamap;
|
|
};
|
|
|
|
struct hme_softc {
|
|
struct ifnet *sc_ifp;
|
|
struct ifmedia sc_ifmedia;
|
|
device_t sc_dev;
|
|
device_t sc_miibus;
|
|
struct mii_data *sc_mii; /* MII media control */
|
|
u_char sc_enaddr[ETHER_ADDR_LEN];
|
|
struct callout sc_tick_ch; /* tick callout */
|
|
int sc_wdog_timer; /* watchdog timer */
|
|
|
|
/* The following bus handles are to be provided by the bus front-end */
|
|
bus_dma_tag_t sc_pdmatag; /* bus dma parent tag */
|
|
bus_dma_tag_t sc_cdmatag; /* control bus dma tag */
|
|
bus_dmamap_t sc_cdmamap; /* control bus dma handle */
|
|
bus_dma_tag_t sc_rdmatag; /* RX bus dma tag */
|
|
bus_dma_tag_t sc_tdmatag; /* RX bus dma tag */
|
|
bus_space_handle_t sc_sebh; /* HME Global registers */
|
|
bus_space_handle_t sc_erxh; /* HME ERX registers */
|
|
bus_space_handle_t sc_etxh; /* HME ETX registers */
|
|
bus_space_handle_t sc_mach; /* HME MAC registers */
|
|
bus_space_handle_t sc_mifh; /* HME MIF registers */
|
|
bus_space_tag_t sc_sebt; /* HME Global registers */
|
|
bus_space_tag_t sc_erxt; /* HME ERX registers */
|
|
bus_space_tag_t sc_etxt; /* HME ETX registers */
|
|
bus_space_tag_t sc_mact; /* HME MAC registers */
|
|
bus_space_tag_t sc_mift; /* HME MIF registers */
|
|
int sc_burst; /* DVMA burst size in effect */
|
|
int sc_phys[2]; /* MII instance -> PHY map */
|
|
|
|
u_int sc_flags;
|
|
#define HME_LINK (1 << 0) /* link is up */
|
|
#define HME_PCI (1 << 1) /* PCI busses are little-endian */
|
|
|
|
int sc_ifflags;
|
|
int sc_csum_features;
|
|
|
|
/* Ring descriptor */
|
|
struct hme_ring sc_rb;
|
|
|
|
struct mtx sc_lock;
|
|
};
|
|
|
|
#define HME_LOCK(_sc) mtx_lock(&(_sc)->sc_lock)
|
|
#define HME_UNLOCK(_sc) mtx_unlock(&(_sc)->sc_lock)
|
|
#define HME_LOCK_ASSERT(_sc, _what) mtx_assert(&(_sc)->sc_lock, (_what))
|
|
|
|
extern devclass_t hme_devclass;
|
|
|
|
int hme_config(struct hme_softc *);
|
|
void hme_detach(struct hme_softc *);
|
|
void hme_suspend(struct hme_softc *);
|
|
void hme_resume(struct hme_softc *);
|
|
void hme_intr(void *);
|
|
|
|
/* MII methods & callbacks */
|
|
int hme_mii_readreg(device_t, int, int);
|
|
int hme_mii_writereg(device_t, int, int, int);
|
|
void hme_mii_statchg(device_t);
|