freebsd-dev/sys/pci/if_ax.c
Peter Wemm 96b3554e5c Use COMPAT_PCI_DRIVER() for registration if it exists. This shouldn't
hurt the driver portability to 3.x too much for where drivers are shared.
1999-04-24 20:17:05 +00:00

2246 lines
52 KiB
C

/*
* Copyright (c) 1997, 1998, 1999
* Bill Paul <wpaul@ctr.columbia.edu>. All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. All advertising materials mentioning features or use of this software
* must display the following acknowledgement:
* This product includes software developed by Bill Paul.
* 4. Neither the name of the author nor the names of any co-contributors
* may be used to endorse or promote products derived from this software
* without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY Bill Paul AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL Bill Paul OR THE VOICES IN HIS HEAD
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
* THE POSSIBILITY OF SUCH DAMAGE.
*
* $Id: if_ax.c,v 1.7 1999/04/08 17:42:47 wpaul Exp $
*/
/*
* ASIX AX88140A and AX88141 fast ethernet PCI NIC driver.
*
* Written by Bill Paul <wpaul@ctr.columbia.edu>
* Electrical Engineering Department
* Columbia University, New York City
*/
/*
* The ASIX Electronics AX88140A is still another DEC 21x4x clone. It's
* a reasonably close copy of the tulip, except for the receiver filter
* programming. Where the DEC chip has a special setup frame that
* needs to be downloaded into the transmit DMA engine, the ASIX chip
* has a less complicated setup frame which is written into one of
* the registers.
*/
#include "bpfilter.h"
#include <sys/param.h>
#include <sys/systm.h>
#include <sys/sockio.h>
#include <sys/mbuf.h>
#include <sys/malloc.h>
#include <sys/kernel.h>
#include <sys/socket.h>
#include <net/if.h>
#include <net/if_arp.h>
#include <net/ethernet.h>
#include <net/if_dl.h>
#include <net/if_media.h>
#if NBPFILTER > 0
#include <net/bpf.h>
#endif
#include <vm/vm.h> /* for vtophys */
#include <vm/pmap.h> /* for vtophys */
#include <machine/clock.h> /* for DELAY */
#include <machine/bus_pio.h>
#include <machine/bus_memio.h>
#include <machine/bus.h>
#include <pci/pcireg.h>
#include <pci/pcivar.h>
#define AX_USEIOSPACE
/* #define AX_BACKGROUND_AUTONEG */
#include <pci/if_axreg.h>
#ifndef lint
static const char rcsid[] =
"$Id: if_ax.c,v 1.7 1999/04/08 17:42:47 wpaul Exp $";
#endif
/*
* Various supported device vendors/types and their names.
*/
static struct ax_type ax_devs[] = {
{ AX_VENDORID, AX_DEVICEID_AX88140A,
"ASIX AX88140A 10/100BaseTX" },
{ AX_VENDORID, AX_DEVICEID_AX88140A,
"ASIX AX88141 10/100BaseTX" },
{ 0, 0, NULL }
};
/*
* Various supported PHY vendors/types and their names. Note that
* this driver will work with pretty much any MII-compliant PHY,
* so failure to positively identify the chip is not a fatal error.
*/
static struct ax_type ax_phys[] = {
{ TI_PHY_VENDORID, TI_PHY_10BT, "<TI ThunderLAN 10BT (internal)>" },
{ TI_PHY_VENDORID, TI_PHY_100VGPMI, "<TI TNETE211 100VG Any-LAN>" },
{ NS_PHY_VENDORID, NS_PHY_83840A, "<National Semiconductor DP83840A>"},
{ LEVEL1_PHY_VENDORID, LEVEL1_PHY_LXT970, "<Level 1 LXT970>" },
{ INTEL_PHY_VENDORID, INTEL_PHY_82555, "<Intel 82555>" },
{ SEEQ_PHY_VENDORID, SEEQ_PHY_80220, "<SEEQ 80220>" },
{ 0, 0, "<MII-compliant physical interface>" }
};
static unsigned long ax_count = 0;
static const char *ax_probe __P((pcici_t, pcidi_t));
static void ax_attach __P((pcici_t, int));
static int ax_newbuf __P((struct ax_softc *,
struct ax_chain_onefrag *));
static int ax_encap __P((struct ax_softc *, struct ax_chain *,
struct mbuf *));
static void ax_rxeof __P((struct ax_softc *));
static void ax_rxeoc __P((struct ax_softc *));
static void ax_txeof __P((struct ax_softc *));
static void ax_txeoc __P((struct ax_softc *));
static void ax_intr __P((void *));
static void ax_start __P((struct ifnet *));
static int ax_ioctl __P((struct ifnet *, u_long, caddr_t));
static void ax_init __P((void *));
static void ax_stop __P((struct ax_softc *));
static void ax_watchdog __P((struct ifnet *));
static void ax_shutdown __P((int, void *));
static int ax_ifmedia_upd __P((struct ifnet *));
static void ax_ifmedia_sts __P((struct ifnet *, struct ifmediareq *));
static void ax_delay __P((struct ax_softc *));
static void ax_eeprom_idle __P((struct ax_softc *));
static void ax_eeprom_putbyte __P((struct ax_softc *, int));
static void ax_eeprom_getword __P((struct ax_softc *, int, u_int16_t *));
static void ax_read_eeprom __P((struct ax_softc *, caddr_t, int,
int, int));
static void ax_mii_writebit __P((struct ax_softc *, int));
static int ax_mii_readbit __P((struct ax_softc *));
static void ax_mii_sync __P((struct ax_softc *));
static void ax_mii_send __P((struct ax_softc *, u_int32_t, int));
static int ax_mii_readreg __P((struct ax_softc *, struct ax_mii_frame *));
static int ax_mii_writereg __P((struct ax_softc *, struct ax_mii_frame *));
static u_int16_t ax_phy_readreg __P((struct ax_softc *, int));
static void ax_phy_writereg __P((struct ax_softc *, int, int));
static void ax_autoneg_xmit __P((struct ax_softc *));
static void ax_autoneg_mii __P((struct ax_softc *, int, int));
static void ax_setmode_mii __P((struct ax_softc *, int));
static void ax_setmode __P((struct ax_softc *, int, int));
static void ax_getmode_mii __P((struct ax_softc *));
static void ax_setcfg __P((struct ax_softc *, int));
static u_int32_t ax_calchash __P((caddr_t));
static void ax_setmulti __P((struct ax_softc *));
static void ax_reset __P((struct ax_softc *));
static int ax_list_rx_init __P((struct ax_softc *));
static int ax_list_tx_init __P((struct ax_softc *));
#define AX_SETBIT(sc, reg, x) \
CSR_WRITE_4(sc, reg, \
CSR_READ_4(sc, reg) | x)
#define AX_CLRBIT(sc, reg, x) \
CSR_WRITE_4(sc, reg, \
CSR_READ_4(sc, reg) & ~x)
#define SIO_SET(x) \
CSR_WRITE_4(sc, AX_SIO, \
CSR_READ_4(sc, AX_SIO) | x)
#define SIO_CLR(x) \
CSR_WRITE_4(sc, AX_SIO, \
CSR_READ_4(sc, AX_SIO) & ~x)
static void ax_delay(sc)
struct ax_softc *sc;
{
int idx;
for (idx = (300 / 33) + 1; idx > 0; idx--)
CSR_READ_4(sc, AX_BUSCTL);
}
static void ax_eeprom_idle(sc)
struct ax_softc *sc;
{
register int i;
CSR_WRITE_4(sc, AX_SIO, AX_SIO_EESEL);
ax_delay(sc);
AX_SETBIT(sc, AX_SIO, AX_SIO_ROMCTL_READ);
ax_delay(sc);
AX_SETBIT(sc, AX_SIO, AX_SIO_EE_CS);
ax_delay(sc);
AX_SETBIT(sc, AX_SIO, AX_SIO_EE_CLK);
ax_delay(sc);
for (i = 0; i < 25; i++) {
AX_CLRBIT(sc, AX_SIO, AX_SIO_EE_CLK);
ax_delay(sc);
AX_SETBIT(sc, AX_SIO, AX_SIO_EE_CLK);
ax_delay(sc);
}
AX_CLRBIT(sc, AX_SIO, AX_SIO_EE_CLK);
ax_delay(sc);
AX_CLRBIT(sc, AX_SIO, AX_SIO_EE_CS);
ax_delay(sc);
CSR_WRITE_4(sc, AX_SIO, 0x00000000);
return;
}
/*
* Send a read command and address to the EEPROM, check for ACK.
*/
static void ax_eeprom_putbyte(sc, addr)
struct ax_softc *sc;
int addr;
{
register int d, i;
d = addr | AX_EECMD_READ;
/*
* Feed in each bit and stobe the clock.
*/
for (i = 0x400; i; i >>= 1) {
if (d & i) {
SIO_SET(AX_SIO_EE_DATAIN);
} else {
SIO_CLR(AX_SIO_EE_DATAIN);
}
ax_delay(sc);
SIO_SET(AX_SIO_EE_CLK);
ax_delay(sc);
SIO_CLR(AX_SIO_EE_CLK);
ax_delay(sc);
}
return;
}
/*
* Read a word of data stored in the EEPROM at address 'addr.'
*/
static void ax_eeprom_getword(sc, addr, dest)
struct ax_softc *sc;
int addr;
u_int16_t *dest;
{
register int i;
u_int16_t word = 0;
/* Force EEPROM to idle state. */
ax_eeprom_idle(sc);
/* Enter EEPROM access mode. */
CSR_WRITE_4(sc, AX_SIO, AX_SIO_EESEL);
ax_delay(sc);
AX_SETBIT(sc, AX_SIO, AX_SIO_ROMCTL_READ);
ax_delay(sc);
AX_SETBIT(sc, AX_SIO, AX_SIO_EE_CS);
ax_delay(sc);
AX_SETBIT(sc, AX_SIO, AX_SIO_EE_CLK);
ax_delay(sc);
/*
* Send address of word we want to read.
*/
ax_eeprom_putbyte(sc, addr);
/*
* Start reading bits from EEPROM.
*/
for (i = 0x8000; i; i >>= 1) {
SIO_SET(AX_SIO_EE_CLK);
ax_delay(sc);
if (CSR_READ_4(sc, AX_SIO) & AX_SIO_EE_DATAOUT)
word |= i;
ax_delay(sc);
SIO_CLR(AX_SIO_EE_CLK);
ax_delay(sc);
}
/* Turn off EEPROM access mode. */
ax_eeprom_idle(sc);
*dest = word;
return;
}
/*
* Read a sequence of words from the EEPROM.
*/
static void ax_read_eeprom(sc, dest, off, cnt, swap)
struct ax_softc *sc;
caddr_t dest;
int off;
int cnt;
int swap;
{
int i;
u_int16_t word = 0, *ptr;
for (i = 0; i < cnt; i++) {
ax_eeprom_getword(sc, off + i, &word);
ptr = (u_int16_t *)(dest + (i * 2));
if (swap)
*ptr = ntohs(word);
else
*ptr = word;
}
return;
}
/*
* Write a bit to the MII bus.
*/
static void ax_mii_writebit(sc, bit)
struct ax_softc *sc;
int bit;
{
if (bit)
CSR_WRITE_4(sc, AX_SIO, AX_SIO_ROMCTL_WRITE|AX_SIO_MII_DATAOUT);
else
CSR_WRITE_4(sc, AX_SIO, AX_SIO_ROMCTL_WRITE);
AX_SETBIT(sc, AX_SIO, AX_SIO_MII_CLK);
AX_CLRBIT(sc, AX_SIO, AX_SIO_MII_CLK);
return;
}
/*
* Read a bit from the MII bus.
*/
static int ax_mii_readbit(sc)
struct ax_softc *sc;
{
CSR_WRITE_4(sc, AX_SIO, AX_SIO_ROMCTL_READ|AX_SIO_MII_DIR);
CSR_READ_4(sc, AX_SIO);
AX_SETBIT(sc, AX_SIO, AX_SIO_MII_CLK);
AX_CLRBIT(sc, AX_SIO, AX_SIO_MII_CLK);
if (CSR_READ_4(sc, AX_SIO) & AX_SIO_MII_DATAIN)
return(1);
return(0);
}
/*
* Sync the PHYs by setting data bit and strobing the clock 32 times.
*/
static void ax_mii_sync(sc)
struct ax_softc *sc;
{
register int i;
CSR_WRITE_4(sc, AX_SIO, AX_SIO_ROMCTL_WRITE);
for (i = 0; i < 32; i++)
ax_mii_writebit(sc, 1);
return;
}
/*
* Clock a series of bits through the MII.
*/
static void ax_mii_send(sc, bits, cnt)
struct ax_softc *sc;
u_int32_t bits;
int cnt;
{
int i;
for (i = (0x1 << (cnt - 1)); i; i >>= 1)
ax_mii_writebit(sc, bits & i);
}
/*
* Read an PHY register through the MII.
*/
static int ax_mii_readreg(sc, frame)
struct ax_softc *sc;
struct ax_mii_frame *frame;
{
int i, ack, s;
s = splimp();
/*
* Set up frame for RX.
*/
frame->mii_stdelim = AX_MII_STARTDELIM;
frame->mii_opcode = AX_MII_READOP;
frame->mii_turnaround = 0;
frame->mii_data = 0;
/*
* Sync the PHYs.
*/
ax_mii_sync(sc);
/*
* Send command/address info.
*/
ax_mii_send(sc, frame->mii_stdelim, 2);
ax_mii_send(sc, frame->mii_opcode, 2);
ax_mii_send(sc, frame->mii_phyaddr, 5);
ax_mii_send(sc, frame->mii_regaddr, 5);
#ifdef notdef
/* Idle bit */
ax_mii_writebit(sc, 1);
ax_mii_writebit(sc, 0);
#endif
/* Check for ack */
ack = ax_mii_readbit(sc);
/*
* Now try reading data bits. If the ack failed, we still
* need to clock through 16 cycles to keep the PHY(s) in sync.
*/
if (ack) {
for(i = 0; i < 16; i++) {
ax_mii_readbit(sc);
}
goto fail;
}
for (i = 0x8000; i; i >>= 1) {
if (!ack) {
if (ax_mii_readbit(sc))
frame->mii_data |= i;
}
}
fail:
ax_mii_writebit(sc, 0);
ax_mii_writebit(sc, 0);
splx(s);
if (ack)
return(1);
return(0);
}
/*
* Write to a PHY register through the MII.
*/
static int ax_mii_writereg(sc, frame)
struct ax_softc *sc;
struct ax_mii_frame *frame;
{
int s;
s = splimp();
/*
* Set up frame for TX.
*/
frame->mii_stdelim = AX_MII_STARTDELIM;
frame->mii_opcode = AX_MII_WRITEOP;
frame->mii_turnaround = AX_MII_TURNAROUND;
/*
* Sync the PHYs.
*/
ax_mii_sync(sc);
ax_mii_send(sc, frame->mii_stdelim, 2);
ax_mii_send(sc, frame->mii_opcode, 2);
ax_mii_send(sc, frame->mii_phyaddr, 5);
ax_mii_send(sc, frame->mii_regaddr, 5);
ax_mii_send(sc, frame->mii_turnaround, 2);
ax_mii_send(sc, frame->mii_data, 16);
/* Idle bit. */
ax_mii_writebit(sc, 0);
ax_mii_writebit(sc, 0);
splx(s);
return(0);
}
static u_int16_t ax_phy_readreg(sc, reg)
struct ax_softc *sc;
int reg;
{
struct ax_mii_frame frame;
bzero((char *)&frame, sizeof(frame));
frame.mii_phyaddr = sc->ax_phy_addr;
frame.mii_regaddr = reg;
ax_mii_readreg(sc, &frame);
return(frame.mii_data);
}
static void ax_phy_writereg(sc, reg, data)
struct ax_softc *sc;
int reg;
int data;
{
struct ax_mii_frame frame;
bzero((char *)&frame, sizeof(frame));
frame.mii_phyaddr = sc->ax_phy_addr;
frame.mii_regaddr = reg;
frame.mii_data = data;
ax_mii_writereg(sc, &frame);
return;
}
/*
* Calculate CRC of a multicast group address, return the lower 6 bits.
*/
static u_int32_t ax_calchash(addr)
caddr_t addr;
{
u_int32_t crc, carry;
int i, j;
u_int8_t c;
/* Compute CRC for the address value. */
crc = 0xFFFFFFFF; /* initial value */
for (i = 0; i < 6; i++) {
c = *(addr + i);
for (j = 0; j < 8; j++) {
carry = ((crc & 0x80000000) ? 1 : 0) ^ (c & 0x01);
crc <<= 1;
c >>= 1;
if (carry)
crc = (crc ^ 0x04c11db6) | carry;
}
}
/* return the filter bit position */
return((crc >> 26) & 0x0000003F);
}
static void ax_setmulti(sc)
struct ax_softc *sc;
{
struct ifnet *ifp;
int h = 0;
u_int32_t hashes[2] = { 0, 0 };
struct ifmultiaddr *ifma;
u_int32_t rxfilt;
ifp = &sc->arpcom.ac_if;
rxfilt = CSR_READ_4(sc, AX_NETCFG);
if (ifp->if_flags & IFF_ALLMULTI || ifp->if_flags & IFF_PROMISC) {
rxfilt |= AX_NETCFG_RX_ALLMULTI;
CSR_WRITE_4(sc, AX_NETCFG, rxfilt);
return;
} else
rxfilt &= ~AX_NETCFG_RX_ALLMULTI;
/* first, zot all the existing hash bits */
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_MAR0);
CSR_WRITE_4(sc, AX_FILTDATA, 0);
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_MAR1);
CSR_WRITE_4(sc, AX_FILTDATA, 0);
/* now program new ones */
for (ifma = ifp->if_multiaddrs.lh_first; ifma != NULL;
ifma = ifma->ifma_link.le_next) {
if (ifma->ifma_addr->sa_family != AF_LINK)
continue;
h = ax_calchash(LLADDR((struct sockaddr_dl *)ifma->ifma_addr));
if (h < 32)
hashes[0] |= (1 << h);
else
hashes[1] |= (1 << (h - 32));
}
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_MAR0);
CSR_WRITE_4(sc, AX_FILTDATA, hashes[0]);
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_MAR1);
CSR_WRITE_4(sc, AX_FILTDATA, hashes[1]);
CSR_WRITE_4(sc, AX_NETCFG, rxfilt);
return;
}
/*
* Initiate an autonegotiation session.
*/
static void ax_autoneg_xmit(sc)
struct ax_softc *sc;
{
u_int16_t phy_sts;
ax_phy_writereg(sc, PHY_BMCR, PHY_BMCR_RESET);
DELAY(500);
while(ax_phy_readreg(sc, PHY_BMCR)
& PHY_BMCR_RESET);
phy_sts = ax_phy_readreg(sc, PHY_BMCR);
phy_sts |= PHY_BMCR_AUTONEGENBL|PHY_BMCR_AUTONEGRSTR;
ax_phy_writereg(sc, PHY_BMCR, phy_sts);
return;
}
/*
* Invoke autonegotiation on a PHY.
*/
static void ax_autoneg_mii(sc, flag, verbose)
struct ax_softc *sc;
int flag;
int verbose;
{
u_int16_t phy_sts = 0, media, advert, ability;
struct ifnet *ifp;
struct ifmedia *ifm;
ifm = &sc->ifmedia;
ifp = &sc->arpcom.ac_if;
ifm->ifm_media = IFM_ETHER | IFM_AUTO;
/*
* The 100baseT4 PHY on the 3c905-T4 has the 'autoneg supported'
* bit cleared in the status register, but has the 'autoneg enabled'
* bit set in the control register. This is a contradiction, and
* I'm not sure how to handle it. If you want to force an attempt
* to autoneg for 100baseT4 PHYs, #define FORCE_AUTONEG_TFOUR
* and see what happens.
*/
#ifndef FORCE_AUTONEG_TFOUR
/*
* First, see if autoneg is supported. If not, there's
* no point in continuing.
*/
phy_sts = ax_phy_readreg(sc, PHY_BMSR);
if (!(phy_sts & PHY_BMSR_CANAUTONEG)) {
if (verbose)
printf("ax%d: autonegotiation not supported\n",
sc->ax_unit);
ifm->ifm_media = IFM_ETHER|IFM_10_T|IFM_HDX;
return;
}
#endif
switch (flag) {
case AX_FLAG_FORCEDELAY:
/*
* XXX Never use this option anywhere but in the probe
* routine: making the kernel stop dead in its tracks
* for three whole seconds after we've gone multi-user
* is really bad manners.
*/
ax_autoneg_xmit(sc);
DELAY(5000000);
break;
case AX_FLAG_SCHEDDELAY:
/*
* Wait for the transmitter to go idle before starting
* an autoneg session, otherwise ax_start() may clobber
* our timeout, and we don't want to allow transmission
* during an autoneg session since that can screw it up.
*/
if (sc->ax_cdata.ax_tx_head != NULL) {
sc->ax_want_auto = 1;
return;
}
ax_autoneg_xmit(sc);
ifp->if_timer = 5;
sc->ax_autoneg = 1;
sc->ax_want_auto = 0;
return;
break;
case AX_FLAG_DELAYTIMEO:
ifp->if_timer = 0;
sc->ax_autoneg = 0;
break;
default:
printf("ax%d: invalid autoneg flag: %d\n", sc->ax_unit, flag);
return;
}
if (ax_phy_readreg(sc, PHY_BMSR) & PHY_BMSR_AUTONEGCOMP) {
if (verbose)
printf("ax%d: autoneg complete, ", sc->ax_unit);
phy_sts = ax_phy_readreg(sc, PHY_BMSR);
} else {
if (verbose)
printf("ax%d: autoneg not complete, ", sc->ax_unit);
}
media = ax_phy_readreg(sc, PHY_BMCR);
/* Link is good. Report modes and set duplex mode. */
if (ax_phy_readreg(sc, PHY_BMSR) & PHY_BMSR_LINKSTAT) {
if (verbose)
printf("link status good ");
advert = ax_phy_readreg(sc, PHY_ANAR);
ability = ax_phy_readreg(sc, PHY_LPAR);
if (advert & PHY_ANAR_100BT4 && ability & PHY_ANAR_100BT4) {
ifm->ifm_media = IFM_ETHER|IFM_100_T4;
media |= PHY_BMCR_SPEEDSEL;
media &= ~PHY_BMCR_DUPLEX;
printf("(100baseT4)\n");
} else if (advert & PHY_ANAR_100BTXFULL &&
ability & PHY_ANAR_100BTXFULL) {
ifm->ifm_media = IFM_ETHER|IFM_100_TX|IFM_FDX;
media |= PHY_BMCR_SPEEDSEL;
media |= PHY_BMCR_DUPLEX;
printf("(full-duplex, 100Mbps)\n");
} else if (advert & PHY_ANAR_100BTXHALF &&
ability & PHY_ANAR_100BTXHALF) {
ifm->ifm_media = IFM_ETHER|IFM_100_TX|IFM_HDX;
media |= PHY_BMCR_SPEEDSEL;
media &= ~PHY_BMCR_DUPLEX;
printf("(half-duplex, 100Mbps)\n");
} else if (advert & PHY_ANAR_10BTFULL &&
ability & PHY_ANAR_10BTFULL) {
ifm->ifm_media = IFM_ETHER|IFM_10_T|IFM_FDX;
media &= ~PHY_BMCR_SPEEDSEL;
media |= PHY_BMCR_DUPLEX;
printf("(full-duplex, 10Mbps)\n");
} else if (advert & PHY_ANAR_10BTHALF &&
ability & PHY_ANAR_10BTHALF) {
ifm->ifm_media = IFM_ETHER|IFM_10_T|IFM_HDX;
media &= ~PHY_BMCR_SPEEDSEL;
media &= ~PHY_BMCR_DUPLEX;
printf("(half-duplex, 10Mbps)\n");
}
media &= ~PHY_BMCR_AUTONEGENBL;
/* Set ASIC's duplex mode to match the PHY. */
ax_setcfg(sc, media);
ax_phy_writereg(sc, PHY_BMCR, media);
} else {
if (verbose)
printf("no carrier\n");
}
ax_init(sc);
if (sc->ax_tx_pend) {
sc->ax_autoneg = 0;
sc->ax_tx_pend = 0;
ax_start(ifp);
}
return;
}
static void ax_getmode_mii(sc)
struct ax_softc *sc;
{
u_int16_t bmsr;
struct ifnet *ifp;
ifp = &sc->arpcom.ac_if;
bmsr = ax_phy_readreg(sc, PHY_BMSR);
if (bootverbose)
printf("ax%d: PHY status word: %x\n", sc->ax_unit, bmsr);
/* fallback */
sc->ifmedia.ifm_media = IFM_ETHER|IFM_10_T|IFM_HDX;
if (bmsr & PHY_BMSR_10BTHALF) {
if (bootverbose)
printf("ax%d: 10Mbps half-duplex mode supported\n",
sc->ax_unit);
ifmedia_add(&sc->ifmedia,
IFM_ETHER|IFM_10_T|IFM_HDX, 0, NULL);
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_10_T, 0, NULL);
}
if (bmsr & PHY_BMSR_10BTFULL) {
if (bootverbose)
printf("ax%d: 10Mbps full-duplex mode supported\n",
sc->ax_unit);
ifmedia_add(&sc->ifmedia,
IFM_ETHER|IFM_10_T|IFM_FDX, 0, NULL);
sc->ifmedia.ifm_media = IFM_ETHER|IFM_10_T|IFM_FDX;
}
if (bmsr & PHY_BMSR_100BTXHALF) {
if (bootverbose)
printf("ax%d: 100Mbps half-duplex mode supported\n",
sc->ax_unit);
ifp->if_baudrate = 100000000;
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_100_TX, 0, NULL);
ifmedia_add(&sc->ifmedia,
IFM_ETHER|IFM_100_TX|IFM_HDX, 0, NULL);
sc->ifmedia.ifm_media = IFM_ETHER|IFM_100_TX|IFM_HDX;
}
if (bmsr & PHY_BMSR_100BTXFULL) {
if (bootverbose)
printf("ax%d: 100Mbps full-duplex mode supported\n",
sc->ax_unit);
ifp->if_baudrate = 100000000;
ifmedia_add(&sc->ifmedia,
IFM_ETHER|IFM_100_TX|IFM_FDX, 0, NULL);
sc->ifmedia.ifm_media = IFM_ETHER|IFM_100_TX|IFM_FDX;
}
/* Some also support 100BaseT4. */
if (bmsr & PHY_BMSR_100BT4) {
if (bootverbose)
printf("ax%d: 100baseT4 mode supported\n", sc->ax_unit);
ifp->if_baudrate = 100000000;
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_100_T4, 0, NULL);
sc->ifmedia.ifm_media = IFM_ETHER|IFM_100_T4;
#ifdef FORCE_AUTONEG_TFOUR
if (bootverbose)
printf("ax%d: forcing on autoneg support for BT4\n",
sc->ax_unit);
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_AUTO, 0 NULL):
sc->ifmedia.ifm_media = IFM_ETHER|IFM_AUTO;
#endif
}
if (bmsr & PHY_BMSR_CANAUTONEG) {
if (bootverbose)
printf("ax%d: autoneg supported\n", sc->ax_unit);
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_AUTO, 0, NULL);
sc->ifmedia.ifm_media = IFM_ETHER|IFM_AUTO;
}
return;
}
/*
* Set speed and duplex mode.
*/
static void ax_setmode_mii(sc, media)
struct ax_softc *sc;
int media;
{
u_int16_t bmcr;
struct ifnet *ifp;
ifp = &sc->arpcom.ac_if;
/*
* If an autoneg session is in progress, stop it.
*/
if (sc->ax_autoneg) {
printf("ax%d: canceling autoneg session\n", sc->ax_unit);
ifp->if_timer = sc->ax_autoneg = sc->ax_want_auto = 0;
bmcr = ax_phy_readreg(sc, PHY_BMCR);
bmcr &= ~PHY_BMCR_AUTONEGENBL;
ax_phy_writereg(sc, PHY_BMCR, bmcr);
}
printf("ax%d: selecting MII, ", sc->ax_unit);
bmcr = ax_phy_readreg(sc, PHY_BMCR);
bmcr &= ~(PHY_BMCR_AUTONEGENBL|PHY_BMCR_SPEEDSEL|
PHY_BMCR_DUPLEX|PHY_BMCR_LOOPBK);
if (IFM_SUBTYPE(media) == IFM_100_T4) {
printf("100Mbps/T4, half-duplex\n");
bmcr |= PHY_BMCR_SPEEDSEL;
bmcr &= ~PHY_BMCR_DUPLEX;
}
if (IFM_SUBTYPE(media) == IFM_100_TX) {
printf("100Mbps, ");
bmcr |= PHY_BMCR_SPEEDSEL;
}
if (IFM_SUBTYPE(media) == IFM_10_T) {
printf("10Mbps, ");
bmcr &= ~PHY_BMCR_SPEEDSEL;
}
if ((media & IFM_GMASK) == IFM_FDX) {
printf("full duplex\n");
bmcr |= PHY_BMCR_DUPLEX;
} else {
printf("half duplex\n");
bmcr &= ~PHY_BMCR_DUPLEX;
}
ax_setcfg(sc, bmcr);
ax_phy_writereg(sc, PHY_BMCR, bmcr);
return;
}
/*
* Set speed and duplex mode on internal transceiver.
*/
static void ax_setmode(sc, media, verbose)
struct ax_softc *sc;
int media;
int verbose;
{
struct ifnet *ifp;
u_int32_t mode;
ifp = &sc->arpcom.ac_if;
if (verbose)
printf("ax%d: selecting internal xcvr, ", sc->ax_unit);
mode = CSR_READ_4(sc, AX_NETCFG);
mode &= ~(AX_NETCFG_FULLDUPLEX|AX_NETCFG_PORTSEL|
AX_NETCFG_PCS|AX_NETCFG_SCRAMBLER|AX_NETCFG_SPEEDSEL);
if (IFM_SUBTYPE(media) == IFM_100_T4) {
if (verbose)
printf("100Mbps/T4, half-duplex\n");
mode |= AX_NETCFG_PORTSEL|AX_NETCFG_PCS|AX_NETCFG_SCRAMBLER;
}
if (IFM_SUBTYPE(media) == IFM_100_TX) {
if (verbose)
printf("100Mbps, ");
mode |= AX_NETCFG_PORTSEL|AX_NETCFG_PCS|AX_NETCFG_SCRAMBLER;
}
if (IFM_SUBTYPE(media) == IFM_10_T) {
if (verbose)
printf("10Mbps, ");
mode &= ~AX_NETCFG_PORTSEL;
mode |= AX_NETCFG_SPEEDSEL;
}
if ((media & IFM_GMASK) == IFM_FDX) {
if (verbose)
printf("full duplex\n");
mode |= AX_NETCFG_FULLDUPLEX;
} else {
if (verbose)
printf("half duplex\n");
mode &= ~AX_NETCFG_FULLDUPLEX;
}
CSR_WRITE_4(sc, AX_NETCFG, mode);
return;
}
/*
* In order to fiddle with the
* 'full-duplex' and '100Mbps' bits in the netconfig register, we
* first have to put the transmit and/or receive logic in the idle state.
*/
static void ax_setcfg(sc, bmcr)
struct ax_softc *sc;
int bmcr;
{
int i, restart = 0;
if (CSR_READ_4(sc, AX_NETCFG) & (AX_NETCFG_TX_ON|AX_NETCFG_RX_ON)) {
restart = 1;
AX_CLRBIT(sc, AX_NETCFG, (AX_NETCFG_TX_ON|AX_NETCFG_RX_ON));
for (i = 0; i < AX_TIMEOUT; i++) {
DELAY(10);
if (CSR_READ_4(sc, AX_ISR) & AX_ISR_TX_IDLE)
break;
}
if (i == AX_TIMEOUT)
printf("ax%d: failed to force tx and "
"rx to idle state\n", sc->ax_unit);
}
if (bmcr & PHY_BMCR_SPEEDSEL)
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_SPEEDSEL);
else
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_SPEEDSEL);
if (bmcr & PHY_BMCR_DUPLEX)
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_FULLDUPLEX);
else
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_FULLDUPLEX);
if (restart)
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_TX_ON|AX_NETCFG_RX_ON);
return;
}
static void ax_reset(sc)
struct ax_softc *sc;
{
register int i;
AX_SETBIT(sc, AX_BUSCTL, AX_BUSCTL_RESET);
for (i = 0; i < AX_TIMEOUT; i++) {
DELAY(10);
if (!(CSR_READ_4(sc, AX_BUSCTL) & AX_BUSCTL_RESET))
break;
}
#ifdef notdef
if (i == AX_TIMEOUT)
printf("ax%d: reset never completed!\n", sc->ax_unit);
#endif
CSR_WRITE_4(sc, AX_BUSCTL, AX_BUSCTL_CONFIG);
/* Wait a little while for the chip to get its brains in order. */
DELAY(1000);
return;
}
/*
* Probe for an ASIX chip. Check the PCI vendor and device
* IDs against our list and return a device name if we find a match.
*/
static const char *
ax_probe(config_id, device_id)
pcici_t config_id;
pcidi_t device_id;
{
struct ax_type *t;
u_int32_t rev;
t = ax_devs;
while(t->ax_name != NULL) {
if ((device_id & 0xFFFF) == t->ax_vid &&
((device_id >> 16) & 0xFFFF) == t->ax_did) {
/* Check the PCI revision */
rev = pci_conf_read(config_id, AX_PCI_REVID) & 0xFF;
if (rev >= AX_REVISION_88141)
t++;
return(t->ax_name);
}
t++;
}
return(NULL);
}
/*
* Attach the interface. Allocate softc structures, do ifmedia
* setup and ethernet/BPF attach.
*/
static void
ax_attach(config_id, unit)
pcici_t config_id;
int unit;
{
int s, i;
#ifndef AX_USEIOSPACE
vm_offset_t pbase, vbase;
#endif
u_char eaddr[ETHER_ADDR_LEN];
u_int32_t command;
struct ax_softc *sc;
struct ifnet *ifp;
int media = IFM_ETHER|IFM_100_TX|IFM_FDX;
unsigned int round;
caddr_t roundptr;
struct ax_type *p;
u_int16_t phy_vid, phy_did, phy_sts;
s = splimp();
sc = malloc(sizeof(struct ax_softc), M_DEVBUF, M_NOWAIT);
if (sc == NULL) {
printf("ax%d: no memory for softc struct!\n", unit);
goto fail;
}
bzero(sc, sizeof(struct ax_softc));
/*
* Handle power management nonsense.
*/
command = pci_conf_read(config_id, AX_PCI_CAPID) & 0x000000FF;
if (command == 0x01) {
command = pci_conf_read(config_id, AX_PCI_PWRMGMTCTRL);
if (command & AX_PSTATE_MASK) {
u_int32_t iobase, membase, irq;
/* Save important PCI config data. */
iobase = pci_conf_read(config_id, AX_PCI_LOIO);
membase = pci_conf_read(config_id, AX_PCI_LOMEM);
irq = pci_conf_read(config_id, AX_PCI_INTLINE);
/* Reset the power state. */
printf("ax%d: chip is in D%d power mode "
"-- setting to D0\n", unit, command & AX_PSTATE_MASK);
command &= 0xFFFFFFFC;
pci_conf_write(config_id, AX_PCI_PWRMGMTCTRL, command);
/* Restore PCI config data. */
pci_conf_write(config_id, AX_PCI_LOIO, iobase);
pci_conf_write(config_id, AX_PCI_LOMEM, membase);
pci_conf_write(config_id, AX_PCI_INTLINE, irq);
}
}
/*
* Map control/status registers.
*/
command = pci_conf_read(config_id, PCI_COMMAND_STATUS_REG);
command |= (PCIM_CMD_PORTEN|PCIM_CMD_MEMEN|PCIM_CMD_BUSMASTEREN);
pci_conf_write(config_id, PCI_COMMAND_STATUS_REG, command);
command = pci_conf_read(config_id, PCI_COMMAND_STATUS_REG);
#ifdef AX_USEIOSPACE
if (!(command & PCIM_CMD_PORTEN)) {
printf("ax%d: failed to enable I/O ports!\n", unit);
free(sc, M_DEVBUF);
goto fail;
}
if (!pci_map_port(config_id, AX_PCI_LOIO,
(u_short *)&(sc->ax_bhandle))) {
printf ("ax%d: couldn't map ports\n", unit);
goto fail;
}
#ifdef __i386__
sc->ax_btag = I386_BUS_SPACE_IO;
#endif
#ifdef __alpha__
sc->ax_btag = ALPHA_BUS_SPACE_IO;
#endif
#else
if (!(command & PCIM_CMD_MEMEN)) {
printf("ax%d: failed to enable memory mapping!\n", unit);
goto fail;
}
if (!pci_map_mem(config_id, AX_PCI_LOMEM, &vbase, &pbase)) {
printf ("ax%d: couldn't map memory\n", unit);
goto fail;
}
#ifdef __i386__
sc->ax_btag = I386_BUS_SPACE_MEM;
#endif
#ifdef __alpha__
sc->ax_btag = ALPHA_BUS_SPACE_MEM;
#endif
sc->ax_bhandle = vbase;
#endif
/* Allocate interrupt */
if (!pci_map_int(config_id, ax_intr, sc, &net_imask)) {
printf("ax%d: couldn't map interrupt\n", unit);
goto fail;
}
/* Reset the adapter. */
ax_reset(sc);
/*
* Get station address from the EEPROM.
*/
ax_read_eeprom(sc, (caddr_t)&eaddr, AX_EE_NODEADDR, 3, 0);
/*
* An ASIX chip was detected. Inform the world.
*/
printf("ax%d: Ethernet address: %6D\n", unit, eaddr, ":");
sc->ax_unit = unit;
bcopy(eaddr, (char *)&sc->arpcom.ac_enaddr, ETHER_ADDR_LEN);
sc->ax_ldata_ptr = malloc(sizeof(struct ax_list_data) + 8,
M_DEVBUF, M_NOWAIT);
if (sc->ax_ldata_ptr == NULL) {
free(sc, M_DEVBUF);
printf("ax%d: no memory for list buffers!\n", unit);
goto fail;
}
sc->ax_ldata = (struct ax_list_data *)sc->ax_ldata_ptr;
round = (unsigned int)sc->ax_ldata_ptr & 0xF;
roundptr = sc->ax_ldata_ptr;
for (i = 0; i < 8; i++) {
if (round % 8) {
round++;
roundptr++;
} else
break;
}
sc->ax_ldata = (struct ax_list_data *)roundptr;
bzero(sc->ax_ldata, sizeof(struct ax_list_data));
ifp = &sc->arpcom.ac_if;
ifp->if_softc = sc;
ifp->if_unit = unit;
ifp->if_name = "ax";
ifp->if_mtu = ETHERMTU;
ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
ifp->if_ioctl = ax_ioctl;
ifp->if_output = ether_output;
ifp->if_start = ax_start;
ifp->if_watchdog = ax_watchdog;
ifp->if_init = ax_init;
ifp->if_baudrate = 10000000;
ifp->if_snd.ifq_maxlen = AX_TX_LIST_CNT - 1;
if (bootverbose)
printf("ax%d: probing for a PHY\n", sc->ax_unit);
for (i = AX_PHYADDR_MIN; i < AX_PHYADDR_MAX + 1; i++) {
if (bootverbose)
printf("ax%d: checking address: %d\n",
sc->ax_unit, i);
sc->ax_phy_addr = i;
ax_phy_writereg(sc, PHY_BMCR, PHY_BMCR_RESET);
DELAY(500);
while(ax_phy_readreg(sc, PHY_BMCR)
& PHY_BMCR_RESET);
if ((phy_sts = ax_phy_readreg(sc, PHY_BMSR)))
break;
}
if (phy_sts) {
phy_vid = ax_phy_readreg(sc, PHY_VENID);
phy_did = ax_phy_readreg(sc, PHY_DEVID);
if (bootverbose)
printf("ax%d: found PHY at address %d, ",
sc->ax_unit, sc->ax_phy_addr);
if (bootverbose)
printf("vendor id: %x device id: %x\n",
phy_vid, phy_did);
p = ax_phys;
while(p->ax_vid) {
if (phy_vid == p->ax_vid &&
(phy_did | 0x000F) == p->ax_did) {
sc->ax_pinfo = p;
break;
}
p++;
}
if (sc->ax_pinfo == NULL)
sc->ax_pinfo = &ax_phys[PHY_UNKNOWN];
if (bootverbose)
printf("ax%d: PHY type: %s\n",
sc->ax_unit, sc->ax_pinfo->ax_name);
} else {
#ifdef DIAGNOSTIC
printf("ax%d: MII without any phy!\n", sc->ax_unit);
#endif
}
/*
* Do ifmedia setup.
*/
ifmedia_init(&sc->ifmedia, 0, ax_ifmedia_upd, ax_ifmedia_sts);
if (sc->ax_pinfo != NULL) {
ax_getmode_mii(sc);
ax_autoneg_mii(sc, AX_FLAG_FORCEDELAY, 1);
} else {
ifmedia_add(&sc->ifmedia,
IFM_ETHER|IFM_10_T|IFM_HDX, 0, NULL);
ifmedia_add(&sc->ifmedia,
IFM_ETHER|IFM_10_T|IFM_FDX, 0, NULL);
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_10_T, 0, NULL);
ifmedia_add(&sc->ifmedia,
IFM_ETHER|IFM_100_TX|IFM_HDX, 0, NULL);
ifmedia_add(&sc->ifmedia,
IFM_ETHER|IFM_100_TX|IFM_FDX, 0, NULL);
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_100_TX, 0, NULL);
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_AUTO, 0, NULL);
}
media = sc->ifmedia.ifm_media;
ax_stop(sc);
ifmedia_set(&sc->ifmedia, media);
/*
* Call MI attach routines.
*/
if_attach(ifp);
ether_ifattach(ifp);
#if NBPFILTER > 0
bpfattach(ifp, DLT_EN10MB, sizeof(struct ether_header));
#endif
at_shutdown(ax_shutdown, sc, SHUTDOWN_POST_SYNC);
fail:
splx(s);
return;
}
/*
* Initialize the transmit descriptors.
*/
static int ax_list_tx_init(sc)
struct ax_softc *sc;
{
struct ax_chain_data *cd;
struct ax_list_data *ld;
int i;
cd = &sc->ax_cdata;
ld = sc->ax_ldata;
for (i = 0; i < AX_TX_LIST_CNT; i++) {
cd->ax_tx_chain[i].ax_ptr = &ld->ax_tx_list[i];
if (i == (AX_TX_LIST_CNT - 1))
cd->ax_tx_chain[i].ax_nextdesc =
&cd->ax_tx_chain[0];
else
cd->ax_tx_chain[i].ax_nextdesc =
&cd->ax_tx_chain[i + 1];
}
cd->ax_tx_free = &cd->ax_tx_chain[0];
cd->ax_tx_tail = cd->ax_tx_head = NULL;
return(0);
}
/*
* Initialize the RX descriptors and allocate mbufs for them. Note that
* we arrange the descriptors in a closed ring, so that the last descriptor
* points back to the first.
*/
static int ax_list_rx_init(sc)
struct ax_softc *sc;
{
struct ax_chain_data *cd;
struct ax_list_data *ld;
int i;
cd = &sc->ax_cdata;
ld = sc->ax_ldata;
for (i = 0; i < AX_RX_LIST_CNT; i++) {
cd->ax_rx_chain[i].ax_ptr =
(volatile struct ax_desc *)&ld->ax_rx_list[i];
if (ax_newbuf(sc, &cd->ax_rx_chain[i]) == ENOBUFS)
return(ENOBUFS);
if (i == (AX_RX_LIST_CNT - 1)) {
cd->ax_rx_chain[i].ax_nextdesc =
&cd->ax_rx_chain[0];
ld->ax_rx_list[i].ax_next =
vtophys(&ld->ax_rx_list[0]);
} else {
cd->ax_rx_chain[i].ax_nextdesc =
&cd->ax_rx_chain[i + 1];
ld->ax_rx_list[i].ax_next =
vtophys(&ld->ax_rx_list[i + 1]);
}
}
cd->ax_rx_head = &cd->ax_rx_chain[0];
return(0);
}
/*
* Initialize an RX descriptor and attach an MBUF cluster.
* Note: the length fields are only 11 bits wide, which means the
* largest size we can specify is 2047. This is important because
* MCLBYTES is 2048, so we have to subtract one otherwise we'll
* overflow the field and make a mess.
*/
static int ax_newbuf(sc, c)
struct ax_softc *sc;
struct ax_chain_onefrag *c;
{
struct mbuf *m_new = NULL;
MGETHDR(m_new, M_DONTWAIT, MT_DATA);
if (m_new == NULL) {
printf("ax%d: no memory for rx list -- packet dropped!\n",
sc->ax_unit);
return(ENOBUFS);
}
MCLGET(m_new, M_DONTWAIT);
if (!(m_new->m_flags & M_EXT)) {
printf("ax%d: no memory for rx list -- packet dropped!\n",
sc->ax_unit);
m_freem(m_new);
return(ENOBUFS);
}
c->ax_mbuf = m_new;
c->ax_ptr->ax_status = AX_RXSTAT;
c->ax_ptr->ax_data = vtophys(mtod(m_new, caddr_t));
c->ax_ptr->ax_ctl = MCLBYTES - 1;
return(0);
}
/*
* A frame has been uploaded: pass the resulting mbuf chain up to
* the higher level protocols.
*/
static void ax_rxeof(sc)
struct ax_softc *sc;
{
struct ether_header *eh;
struct mbuf *m;
struct ifnet *ifp;
struct ax_chain_onefrag *cur_rx;
int total_len = 0;
u_int32_t rxstat;
ifp = &sc->arpcom.ac_if;
while(!((rxstat = sc->ax_cdata.ax_rx_head->ax_ptr->ax_status) &
AX_RXSTAT_OWN)) {
#ifdef __alpha__
struct mbuf *m0 = NULL;
#endif
cur_rx = sc->ax_cdata.ax_rx_head;
sc->ax_cdata.ax_rx_head = cur_rx->ax_nextdesc;
/*
* If an error occurs, update stats, clear the
* status word and leave the mbuf cluster in place:
* it should simply get re-used next time this descriptor
* comes up in the ring.
*/
if (rxstat & AX_RXSTAT_RXERR) {
ifp->if_ierrors++;
if (rxstat & AX_RXSTAT_COLLSEEN)
ifp->if_collisions++;
cur_rx->ax_ptr->ax_status = AX_RXSTAT;
cur_rx->ax_ptr->ax_ctl = (MCLBYTES - 1);
continue;
}
/* No errors; receive the packet. */
m = cur_rx->ax_mbuf;
total_len = AX_RXBYTES(cur_rx->ax_ptr->ax_status);
total_len -= ETHER_CRC_LEN;
#ifdef __alpha__
/*
* Try to conjure up a new mbuf cluster. If that
* fails, it means we have an out of memory condition and
* should leave the buffer in place and continue. This will
* result in a lost packet, but there's little else we
* can do in this situation.
*/
if (ax_newbuf(sc, cur_rx) == ENOBUFS) {
ifp->if_ierrors++;
cur_rx->ax_ptr->ax_status = AX_RXSTAT;
cur_rx->ax_ptr->ax_ctl = (MCLBYTES - 1);
continue;
}
/*
* Sadly, the ASIX chip doesn't decode the last few
* bits of the RX DMA buffer address, so we have to
* cheat in order to obtain proper payload alignment
* on the alpha.
*/
MGETHDR(m0, M_DONTWAIT, MT_DATA);
if (m0 == NULL) {
ifp->if_ierrors++;
cur_rx->ax_ptr->ax_status = AX_RXSTAT;
cur_rx->ax_ptr->ax_ctl = (MCLBYTES - 1);
continue;
}
m0->m_data += 2;
if (total_len <= (MHLEN - 2)) {
bcopy(mtod(m, caddr_t), mtod(m0, caddr_t), total_len); m_freem(m);
m = m0;
m->m_pkthdr.len = m->m_len = total_len;
} else {
bcopy(mtod(m, caddr_t), mtod(m0, caddr_t), (MHLEN - 2));
m->m_len = total_len - (MHLEN - 2);
m->m_data += (MHLEN - 2);
m0->m_next = m;
m0->m_len = (MHLEN - 2);
m = m0;
m->m_pkthdr.len = total_len;
}
m->m_pkthdr.rcvif = ifp;
#else
if (total_len < MINCLSIZE) {
m = m_devget(mtod(cur_rx->ax_mbuf, char *),
total_len, 0, ifp, NULL);
cur_rx->ax_ptr->ax_status = AX_RXSTAT;
cur_rx->ax_ptr->ax_ctl = (MCLBYTES - 1);
if (m == NULL) {
ifp->if_ierrors++;
continue;
}
} else {
m = cur_rx->ax_mbuf;
/*
* Try to conjure up a new mbuf cluster. If that
* fails, it means we have an out of memory condition and
* should leave the buffer in place and continue. This will
* result in a lost packet, but there's little else we
* can do in this situation.
*/
if (ax_newbuf(sc, cur_rx) == ENOBUFS) {
ifp->if_ierrors++;
cur_rx->ax_ptr->ax_status = AX_RXSTAT;
cur_rx->ax_ptr->ax_ctl = (MCLBYTES - 1);
continue;
}
m->m_pkthdr.rcvif = ifp;
m->m_pkthdr.len = m->m_len = total_len;
}
#endif
ifp->if_ipackets++;
eh = mtod(m, struct ether_header *);
#if NBPFILTER > 0
/*
* Handle BPF listeners. Let the BPF user see the packet, but
* don't pass it up to the ether_input() layer unless it's
* a broadcast packet, multicast packet, matches our ethernet
* address or the interface is in promiscuous mode.
*/
if (ifp->if_bpf) {
bpf_mtap(ifp, m);
if (ifp->if_flags & IFF_PROMISC &&
(bcmp(eh->ether_dhost, sc->arpcom.ac_enaddr,
ETHER_ADDR_LEN) &&
(eh->ether_dhost[0] & 1) == 0)) {
m_freem(m);
continue;
}
}
#endif
/* Remove header from mbuf and pass it on. */
m_adj(m, sizeof(struct ether_header));
ether_input(ifp, eh, m);
}
return;
}
void ax_rxeoc(sc)
struct ax_softc *sc;
{
ax_rxeof(sc);
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_RX_ON);
CSR_WRITE_4(sc, AX_RXADDR, vtophys(sc->ax_cdata.ax_rx_head->ax_ptr));
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_RX_ON);
CSR_WRITE_4(sc, AX_RXSTART, 0xFFFFFFFF);
return;
}
/*
* A frame was downloaded to the chip. It's safe for us to clean up
* the list buffers.
*/
static void ax_txeof(sc)
struct ax_softc *sc;
{
struct ax_chain *cur_tx;
struct ifnet *ifp;
ifp = &sc->arpcom.ac_if;
/* Clear the timeout timer. */
ifp->if_timer = 0;
if (sc->ax_cdata.ax_tx_head == NULL)
return;
/*
* Go through our tx list and free mbufs for those
* frames that have been transmitted.
*/
while(sc->ax_cdata.ax_tx_head->ax_mbuf != NULL) {
u_int32_t txstat;
cur_tx = sc->ax_cdata.ax_tx_head;
txstat = AX_TXSTATUS(cur_tx);
if (txstat & AX_TXSTAT_OWN)
break;
if (txstat & AX_TXSTAT_ERRSUM) {
ifp->if_oerrors++;
if (txstat & AX_TXSTAT_EXCESSCOLL)
ifp->if_collisions++;
if (txstat & AX_TXSTAT_LATECOLL)
ifp->if_collisions++;
}
ifp->if_collisions += (txstat & AX_TXSTAT_COLLCNT) >> 3;
ifp->if_opackets++;
m_freem(cur_tx->ax_mbuf);
cur_tx->ax_mbuf = NULL;
if (sc->ax_cdata.ax_tx_head == sc->ax_cdata.ax_tx_tail) {
sc->ax_cdata.ax_tx_head = NULL;
sc->ax_cdata.ax_tx_tail = NULL;
break;
}
sc->ax_cdata.ax_tx_head = cur_tx->ax_nextdesc;
}
return;
}
/*
* TX 'end of channel' interrupt handler.
*/
static void ax_txeoc(sc)
struct ax_softc *sc;
{
struct ifnet *ifp;
ifp = &sc->arpcom.ac_if;
ifp->if_timer = 0;
if (sc->ax_cdata.ax_tx_head == NULL) {
ifp->if_flags &= ~IFF_OACTIVE;
sc->ax_cdata.ax_tx_tail = NULL;
if (sc->ax_want_auto)
ax_autoneg_mii(sc, AX_FLAG_DELAYTIMEO, 1);
}
return;
}
static void ax_intr(arg)
void *arg;
{
struct ax_softc *sc;
struct ifnet *ifp;
u_int32_t status;
sc = arg;
ifp = &sc->arpcom.ac_if;
/* Supress unwanted interrupts */
if (!(ifp->if_flags & IFF_UP)) {
ax_stop(sc);
return;
}
/* Disable interrupts. */
CSR_WRITE_4(sc, AX_IMR, 0x00000000);
for (;;) {
status = CSR_READ_4(sc, AX_ISR);
if (status)
CSR_WRITE_4(sc, AX_ISR, status);
if ((status & AX_INTRS) == 0)
break;
if ((status & AX_ISR_TX_OK) || (status & AX_ISR_TX_EARLY))
ax_txeof(sc);
if (status & AX_ISR_TX_NOBUF)
ax_txeoc(sc);
if (status & AX_ISR_TX_IDLE) {
ax_txeof(sc);
if (sc->ax_cdata.ax_tx_head != NULL) {
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_TX_ON);
CSR_WRITE_4(sc, AX_TXSTART, 0xFFFFFFFF);
}
}
if (status & AX_ISR_TX_UNDERRUN) {
u_int32_t cfg;
cfg = CSR_READ_4(sc, AX_NETCFG);
if ((cfg & AX_NETCFG_TX_THRESH) == AX_TXTHRESH_160BYTES)
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_STORENFWD);
else
CSR_WRITE_4(sc, AX_NETCFG, cfg + 0x4000);
}
if (status & AX_ISR_RX_OK)
ax_rxeof(sc);
if ((status & AX_ISR_RX_WATDOGTIMEO)
|| (status & AX_ISR_RX_NOBUF))
ax_rxeoc(sc);
if (status & AX_ISR_BUS_ERR) {
ax_reset(sc);
ax_init(sc);
}
}
/* Re-enable interrupts. */
CSR_WRITE_4(sc, AX_IMR, AX_INTRS);
if (ifp->if_snd.ifq_head != NULL) {
ax_start(ifp);
}
return;
}
/*
* Encapsulate an mbuf chain in a descriptor by coupling the mbuf data
* pointers to the fragment pointers.
*/
static int ax_encap(sc, c, m_head)
struct ax_softc *sc;
struct ax_chain *c;
struct mbuf *m_head;
{
int frag = 0;
volatile struct ax_desc *f = NULL;
int total_len;
struct mbuf *m;
/*
* Start packing the mbufs in this chain into
* the fragment pointers. Stop when we run out
* of fragments or hit the end of the mbuf chain.
*/
m = m_head;
total_len = 0;
for (m = m_head, frag = 0; m != NULL; m = m->m_next) {
if (m->m_len != 0) {
if (frag == AX_MAXFRAGS)
break;
total_len += m->m_len;
f = &c->ax_ptr->ax_frag[frag];
f->ax_ctl = m->m_len;
if (frag == 0) {
f->ax_status = 0;
f->ax_ctl |= AX_TXCTL_FIRSTFRAG;
} else
f->ax_status = AX_TXSTAT_OWN;
f->ax_next = vtophys(&c->ax_ptr->ax_frag[frag + 1]);
f->ax_data = vtophys(mtod(m, vm_offset_t));
frag++;
}
}
/*
* Handle special case: we ran out of fragments,
* but we have more mbufs left in the chain. Copy the
* data into an mbuf cluster. Note that we don't
* bother clearing the values in the other fragment
* pointers/counters; it wouldn't gain us anything,
* and would waste cycles.
*/
if (m != NULL) {
struct mbuf *m_new = NULL;
MGETHDR(m_new, M_DONTWAIT, MT_DATA);
if (m_new == NULL) {
printf("ax%d: no memory for tx list", sc->ax_unit);
return(1);
}
if (m_head->m_pkthdr.len > MHLEN) {
MCLGET(m_new, M_DONTWAIT);
if (!(m_new->m_flags & M_EXT)) {
m_freem(m_new);
printf("ax%d: no memory for tx list",
sc->ax_unit);
return(1);
}
}
m_copydata(m_head, 0, m_head->m_pkthdr.len,
mtod(m_new, caddr_t));
m_new->m_pkthdr.len = m_new->m_len = m_head->m_pkthdr.len;
m_freem(m_head);
m_head = m_new;
f = &c->ax_ptr->ax_frag[0];
f->ax_status = 0;
f->ax_data = vtophys(mtod(m_new, caddr_t));
f->ax_ctl = total_len = m_new->m_len;
f->ax_ctl |= AX_TXCTL_FIRSTFRAG;
frag = 1;
}
c->ax_mbuf = m_head;
c->ax_lastdesc = frag - 1;
AX_TXCTL(c) |= AX_TXCTL_LASTFRAG|AX_TXCTL_FINT;
c->ax_ptr->ax_frag[0].ax_ctl |= AX_TXCTL_FINT;
AX_TXNEXT(c) = vtophys(&c->ax_nextdesc->ax_ptr->ax_frag[0]);
return(0);
}
/*
* Main transmit routine. To avoid having to do mbuf copies, we put pointers
* to the mbuf data regions directly in the transmit lists. We also save a
* copy of the pointers since the transmit list fragment pointers are
* physical addresses.
*/
static void ax_start(ifp)
struct ifnet *ifp;
{
struct ax_softc *sc;
struct mbuf *m_head = NULL;
struct ax_chain *cur_tx = NULL, *start_tx;
sc = ifp->if_softc;
if (sc->ax_autoneg) {
sc->ax_tx_pend = 1;
return;
}
/*
* Check for an available queue slot. If there are none,
* punt.
*/
if (sc->ax_cdata.ax_tx_free->ax_mbuf != NULL) {
ifp->if_flags |= IFF_OACTIVE;
return;
}
start_tx = sc->ax_cdata.ax_tx_free;
while(sc->ax_cdata.ax_tx_free->ax_mbuf == NULL) {
IF_DEQUEUE(&ifp->if_snd, m_head);
if (m_head == NULL)
break;
/* Pick a descriptor off the free list. */
cur_tx = sc->ax_cdata.ax_tx_free;
sc->ax_cdata.ax_tx_free = cur_tx->ax_nextdesc;
/* Pack the data into the descriptor. */
ax_encap(sc, cur_tx, m_head);
if (cur_tx != start_tx)
AX_TXOWN(cur_tx) = AX_TXSTAT_OWN;
#if NBPFILTER > 0
/*
* If there's a BPF listener, bounce a copy of this frame
* to him.
*/
if (ifp->if_bpf)
bpf_mtap(ifp, cur_tx->ax_mbuf);
#endif
AX_TXOWN(cur_tx) = AX_TXSTAT_OWN;
CSR_WRITE_4(sc, AX_TXSTART, 0xFFFFFFFF);
}
sc->ax_cdata.ax_tx_tail = cur_tx;
if (sc->ax_cdata.ax_tx_head == NULL)
sc->ax_cdata.ax_tx_head = start_tx;
/*
* Set a timeout in case the chip goes out to lunch.
*/
ifp->if_timer = 5;
return;
}
static void ax_init(xsc)
void *xsc;
{
struct ax_softc *sc = xsc;
struct ifnet *ifp = &sc->arpcom.ac_if;
u_int16_t phy_bmcr = 0;
int s;
if (sc->ax_autoneg)
return;
s = splimp();
if (sc->ax_pinfo != NULL)
phy_bmcr = ax_phy_readreg(sc, PHY_BMCR);
/*
* Cancel pending I/O and free all RX/TX buffers.
*/
ax_stop(sc);
ax_reset(sc);
/*
* Set cache alignment and burst length.
*/
CSR_WRITE_4(sc, AX_BUSCTL, AX_BUSCTL_CONFIG);
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_HEARTBEAT);
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_STORENFWD);
if (sc->ax_pinfo != NULL) {
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_PORTSEL);
ax_setcfg(sc, ax_phy_readreg(sc, PHY_BMCR));
} else
ax_setmode(sc, sc->ifmedia.ifm_media, 0);
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_TX_THRESH);
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_SPEEDSEL);
if (IFM_SUBTYPE(sc->ifmedia.ifm_media) == IFM_10_T)
AX_SETBIT(sc, AX_NETCFG, AX_TXTHRESH_160BYTES);
else
AX_SETBIT(sc, AX_NETCFG, AX_TXTHRESH_72BYTES);
/* Init our MAC address */
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_PAR0);
CSR_WRITE_4(sc, AX_FILTDATA, *(u_int32_t *)(&sc->arpcom.ac_enaddr[0]));
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_PAR1);
CSR_WRITE_4(sc, AX_FILTDATA, *(u_int32_t *)(&sc->arpcom.ac_enaddr[4]));
/* Init circular RX list. */
if (ax_list_rx_init(sc) == ENOBUFS) {
printf("ax%d: initialization failed: no "
"memory for rx buffers\n", sc->ax_unit);
ax_stop(sc);
(void)splx(s);
return;
}
/*
* Init tx descriptors.
*/
ax_list_tx_init(sc);
/* If we want promiscuous mode, set the allframes bit. */
if (ifp->if_flags & IFF_PROMISC) {
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_RX_PROMISC);
} else {
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_RX_PROMISC);
}
/*
* Set the capture broadcast bit to capture broadcast frames.
*/
if (ifp->if_flags & IFF_BROADCAST) {
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_RX_BROAD);
} else {
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_RX_BROAD);
}
/*
* Load the multicast filter.
*/
ax_setmulti(sc);
/*
* Load the address of the RX list.
*/
CSR_WRITE_4(sc, AX_RXADDR, vtophys(sc->ax_cdata.ax_rx_head->ax_ptr));
CSR_WRITE_4(sc, AX_TXADDR, vtophys(&sc->ax_ldata->ax_tx_list[0]));
/*
* Enable interrupts.
*/
CSR_WRITE_4(sc, AX_IMR, AX_INTRS);
CSR_WRITE_4(sc, AX_ISR, 0xFFFFFFFF);
/* Enable receiver and transmitter. */
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_TX_ON|AX_NETCFG_RX_ON);
CSR_WRITE_4(sc, AX_RXSTART, 0xFFFFFFFF);
/* Restore state of BMCR */
if (sc->ax_pinfo != NULL)
ax_phy_writereg(sc, PHY_BMCR, phy_bmcr);
ifp->if_flags |= IFF_RUNNING;
ifp->if_flags &= ~IFF_OACTIVE;
(void)splx(s);
return;
}
/*
* Set media options.
*/
static int ax_ifmedia_upd(ifp)
struct ifnet *ifp;
{
struct ax_softc *sc;
struct ifmedia *ifm;
sc = ifp->if_softc;
ifm = &sc->ifmedia;
if (IFM_TYPE(ifm->ifm_media) != IFM_ETHER)
return(EINVAL);
if (IFM_SUBTYPE(ifm->ifm_media) == IFM_AUTO)
ax_autoneg_mii(sc, AX_FLAG_SCHEDDELAY, 1);
else {
if (sc->ax_pinfo == NULL)
ax_setmode(sc, ifm->ifm_media, 1);
else
ax_setmode_mii(sc, ifm->ifm_media);
}
return(0);
}
/*
* Report current media status.
*/
static void ax_ifmedia_sts(ifp, ifmr)
struct ifnet *ifp;
struct ifmediareq *ifmr;
{
struct ax_softc *sc;
u_int16_t advert = 0, ability = 0;
u_int32_t media = 0;
sc = ifp->if_softc;
ifmr->ifm_active = IFM_ETHER;
if (sc->ax_pinfo == NULL) {
media = CSR_READ_4(sc, AX_NETCFG);
if (media & AX_NETCFG_PORTSEL)
ifmr->ifm_active = IFM_ETHER|IFM_100_TX;
else
ifmr->ifm_active = IFM_ETHER|IFM_10_T;
if (media & AX_NETCFG_FULLDUPLEX)
ifmr->ifm_active |= IFM_FDX;
else
ifmr->ifm_active |= IFM_HDX;
return;
}
if (!(ax_phy_readreg(sc, PHY_BMCR) & PHY_BMCR_AUTONEGENBL)) {
if (ax_phy_readreg(sc, PHY_BMCR) & PHY_BMCR_SPEEDSEL)
ifmr->ifm_active = IFM_ETHER|IFM_100_TX;
else
ifmr->ifm_active = IFM_ETHER|IFM_10_T;
if (ax_phy_readreg(sc, PHY_BMCR) & PHY_BMCR_DUPLEX)
ifmr->ifm_active |= IFM_FDX;
else
ifmr->ifm_active |= IFM_HDX;
return;
}
ability = ax_phy_readreg(sc, PHY_LPAR);
advert = ax_phy_readreg(sc, PHY_ANAR);
if (advert & PHY_ANAR_100BT4 &&
ability & PHY_ANAR_100BT4) {
ifmr->ifm_active = IFM_ETHER|IFM_100_T4;
} else if (advert & PHY_ANAR_100BTXFULL &&
ability & PHY_ANAR_100BTXFULL) {
ifmr->ifm_active = IFM_ETHER|IFM_100_TX|IFM_FDX;
} else if (advert & PHY_ANAR_100BTXHALF &&
ability & PHY_ANAR_100BTXHALF) {
ifmr->ifm_active = IFM_ETHER|IFM_100_TX|IFM_HDX;
} else if (advert & PHY_ANAR_10BTFULL &&
ability & PHY_ANAR_10BTFULL) {
ifmr->ifm_active = IFM_ETHER|IFM_10_T|IFM_FDX;
} else if (advert & PHY_ANAR_10BTHALF &&
ability & PHY_ANAR_10BTHALF) {
ifmr->ifm_active = IFM_ETHER|IFM_10_T|IFM_HDX;
}
return;
}
static int ax_ioctl(ifp, command, data)
struct ifnet *ifp;
u_long command;
caddr_t data;
{
struct ax_softc *sc = ifp->if_softc;
struct ifreq *ifr = (struct ifreq *) data;
int s, error = 0;
s = splimp();
switch(command) {
case SIOCSIFADDR:
case SIOCGIFADDR:
case SIOCSIFMTU:
error = ether_ioctl(ifp, command, data);
break;
case SIOCSIFFLAGS:
if (ifp->if_flags & IFF_UP) {
ax_init(sc);
} else {
if (ifp->if_flags & IFF_RUNNING)
ax_stop(sc);
}
error = 0;
break;
case SIOCADDMULTI:
case SIOCDELMULTI:
ax_setmulti(sc);
error = 0;
break;
case SIOCGIFMEDIA:
case SIOCSIFMEDIA:
error = ifmedia_ioctl(ifp, ifr, &sc->ifmedia, command);
break;
default:
error = EINVAL;
break;
}
(void)splx(s);
return(error);
}
static void ax_watchdog(ifp)
struct ifnet *ifp;
{
struct ax_softc *sc;
sc = ifp->if_softc;
if (sc->ax_autoneg) {
ax_autoneg_mii(sc, AX_FLAG_DELAYTIMEO, 1);
return;
}
ifp->if_oerrors++;
printf("ax%d: watchdog timeout\n", sc->ax_unit);
if (sc->ax_pinfo != NULL) {
if (!(ax_phy_readreg(sc, PHY_BMSR) & PHY_BMSR_LINKSTAT))
printf("ax%d: no carrier - transceiver "
"cable problem?\n", sc->ax_unit);
}
ax_stop(sc);
ax_reset(sc);
ax_init(sc);
if (ifp->if_snd.ifq_head != NULL)
ax_start(ifp);
return;
}
/*
* Stop the adapter and free any mbufs allocated to the
* RX and TX lists.
*/
static void ax_stop(sc)
struct ax_softc *sc;
{
register int i;
struct ifnet *ifp;
ifp = &sc->arpcom.ac_if;
ifp->if_timer = 0;
AX_CLRBIT(sc, AX_NETCFG, (AX_NETCFG_RX_ON|AX_NETCFG_TX_ON));
CSR_WRITE_4(sc, AX_IMR, 0x00000000);
CSR_WRITE_4(sc, AX_TXADDR, 0x00000000);
CSR_WRITE_4(sc, AX_RXADDR, 0x00000000);
/*
* Free data in the RX lists.
*/
for (i = 0; i < AX_RX_LIST_CNT; i++) {
if (sc->ax_cdata.ax_rx_chain[i].ax_mbuf != NULL) {
m_freem(sc->ax_cdata.ax_rx_chain[i].ax_mbuf);
sc->ax_cdata.ax_rx_chain[i].ax_mbuf = NULL;
}
}
bzero((char *)&sc->ax_ldata->ax_rx_list,
sizeof(sc->ax_ldata->ax_rx_list));
/*
* Free the TX list buffers.
*/
for (i = 0; i < AX_TX_LIST_CNT; i++) {
if (sc->ax_cdata.ax_tx_chain[i].ax_mbuf != NULL) {
m_freem(sc->ax_cdata.ax_tx_chain[i].ax_mbuf);
sc->ax_cdata.ax_tx_chain[i].ax_mbuf = NULL;
}
}
bzero((char *)&sc->ax_ldata->ax_tx_list,
sizeof(sc->ax_ldata->ax_tx_list));
ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
return;
}
/*
* Stop all chip I/O so that the kernel's probe routines don't
* get confused by errant DMAs when rebooting.
*/
static void ax_shutdown(howto, arg)
int howto;
void *arg;
{
struct ax_softc *sc = (struct ax_softc *)arg;
ax_stop(sc);
return;
}
static struct pci_device ax_device = {
"ax",
ax_probe,
ax_attach,
&ax_count,
NULL
};
#ifdef COMPAT_PCI_DRIVER
COMPAT_PCI_DRIVER(ax, ax_device);
#else
DATA_SET(pcidevice_set, ax_device);
#endif /* COMPAT_PCI_DRIVER */