975fba8a24
kernel compile
2176 lines
50 KiB
C
2176 lines
50 KiB
C
/*
|
|
* Copyright (c) 1997, 1998, 1999
|
|
* Bill Paul <wpaul@ctr.columbia.edu>. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed by Bill Paul.
|
|
* 4. Neither the name of the author nor the names of any co-contributors
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY Bill Paul AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL Bill Paul OR THE VOICES IN HIS HEAD
|
|
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
|
|
* THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
* $Id: if_ax.c,v 1.3 1999/01/16 20:40:52 wpaul Exp $
|
|
*/
|
|
|
|
/*
|
|
* ASIX AX88140A fast ethernet PCI NIC driver.
|
|
*
|
|
* Written by Bill Paul <wpaul@ctr.columbia.edu>
|
|
* Electrical Engineering Department
|
|
* Columbia University, New York City
|
|
*/
|
|
|
|
/*
|
|
* The ASIX Electronics AX88140A is still another DEC 21x4x clone. It's
|
|
* a reasonably close copy of the tulip, except for the receiver filter
|
|
* programming. Where the DEC chip has a special setup frame that
|
|
* needs to be downloaded into the transmit DMA engine, the ASIX chip
|
|
* has a less complicated setup frame which is written into one of
|
|
* the registers.
|
|
*/
|
|
|
|
#include "bpfilter.h"
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/sockio.h>
|
|
#include <sys/mbuf.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/socket.h>
|
|
|
|
#include <net/if.h>
|
|
#include <net/if_arp.h>
|
|
#include <net/ethernet.h>
|
|
#include <net/if_dl.h>
|
|
#include <net/if_media.h>
|
|
|
|
#if NBPFILTER > 0
|
|
#include <net/bpf.h>
|
|
#endif
|
|
|
|
#include <vm/vm.h> /* for vtophys */
|
|
#include <vm/pmap.h> /* for vtophys */
|
|
#include <machine/clock.h> /* for DELAY */
|
|
#include <machine/bus_pio.h>
|
|
#include <machine/bus_memio.h>
|
|
#include <machine/bus.h>
|
|
|
|
#include <pci/pcireg.h>
|
|
#include <pci/pcivar.h>
|
|
|
|
#define AX_USEIOSPACE
|
|
|
|
/* #define AX_BACKGROUND_AUTONEG */
|
|
|
|
#include <pci/if_axreg.h>
|
|
|
|
#ifndef lint
|
|
static const char rcsid[] =
|
|
"$Id: if_ax.c,v 1.3 1999/01/16 20:40:52 wpaul Exp $";
|
|
#endif
|
|
|
|
/*
|
|
* Various supported device vendors/types and their names.
|
|
*/
|
|
static struct ax_type ax_devs[] = {
|
|
{ AX_VENDORID, AX_DEVICEID_AX88140A,
|
|
"ASIX AX88140A 10/100BaseTX" },
|
|
{ 0, 0, NULL }
|
|
};
|
|
|
|
/*
|
|
* Various supported PHY vendors/types and their names. Note that
|
|
* this driver will work with pretty much any MII-compliant PHY,
|
|
* so failure to positively identify the chip is not a fatal error.
|
|
*/
|
|
|
|
static struct ax_type ax_phys[] = {
|
|
{ TI_PHY_VENDORID, TI_PHY_10BT, "<TI ThunderLAN 10BT (internal)>" },
|
|
{ TI_PHY_VENDORID, TI_PHY_100VGPMI, "<TI TNETE211 100VG Any-LAN>" },
|
|
{ NS_PHY_VENDORID, NS_PHY_83840A, "<National Semiconductor DP83840A>"},
|
|
{ LEVEL1_PHY_VENDORID, LEVEL1_PHY_LXT970, "<Level 1 LXT970>" },
|
|
{ INTEL_PHY_VENDORID, INTEL_PHY_82555, "<Intel 82555>" },
|
|
{ SEEQ_PHY_VENDORID, SEEQ_PHY_80220, "<SEEQ 80220>" },
|
|
{ 0, 0, "<MII-compliant physical interface>" }
|
|
};
|
|
|
|
static unsigned long ax_count = 0;
|
|
static const char *ax_probe __P((pcici_t, pcidi_t));
|
|
static void ax_attach __P((pcici_t, int));
|
|
|
|
static int ax_newbuf __P((struct ax_softc *,
|
|
struct ax_chain_onefrag *));
|
|
static int ax_encap __P((struct ax_softc *, struct ax_chain *,
|
|
struct mbuf *));
|
|
|
|
static void ax_rxeof __P((struct ax_softc *));
|
|
static void ax_rxeoc __P((struct ax_softc *));
|
|
static void ax_txeof __P((struct ax_softc *));
|
|
static void ax_txeoc __P((struct ax_softc *));
|
|
static void ax_intr __P((void *));
|
|
static void ax_start __P((struct ifnet *));
|
|
static int ax_ioctl __P((struct ifnet *, u_long, caddr_t));
|
|
static void ax_init __P((void *));
|
|
static void ax_stop __P((struct ax_softc *));
|
|
static void ax_watchdog __P((struct ifnet *));
|
|
static void ax_shutdown __P((int, void *));
|
|
static int ax_ifmedia_upd __P((struct ifnet *));
|
|
static void ax_ifmedia_sts __P((struct ifnet *, struct ifmediareq *));
|
|
|
|
static void ax_delay __P((struct ax_softc *));
|
|
static void ax_eeprom_idle __P((struct ax_softc *));
|
|
static void ax_eeprom_putbyte __P((struct ax_softc *, int));
|
|
static void ax_eeprom_getword __P((struct ax_softc *, int, u_int16_t *));
|
|
static void ax_read_eeprom __P((struct ax_softc *, caddr_t, int,
|
|
int, int));
|
|
|
|
static void ax_mii_writebit __P((struct ax_softc *, int));
|
|
static int ax_mii_readbit __P((struct ax_softc *));
|
|
static void ax_mii_sync __P((struct ax_softc *));
|
|
static void ax_mii_send __P((struct ax_softc *, u_int32_t, int));
|
|
static int ax_mii_readreg __P((struct ax_softc *, struct ax_mii_frame *));
|
|
static int ax_mii_writereg __P((struct ax_softc *, struct ax_mii_frame *));
|
|
static u_int16_t ax_phy_readreg __P((struct ax_softc *, int));
|
|
static void ax_phy_writereg __P((struct ax_softc *, int, int));
|
|
|
|
static void ax_autoneg_xmit __P((struct ax_softc *));
|
|
static void ax_autoneg_mii __P((struct ax_softc *, int, int));
|
|
static void ax_setmode_mii __P((struct ax_softc *, int));
|
|
static void ax_setmode __P((struct ax_softc *, int, int));
|
|
static void ax_getmode_mii __P((struct ax_softc *));
|
|
static void ax_setcfg __P((struct ax_softc *, int));
|
|
static u_int32_t ax_calchash __P((caddr_t));
|
|
static void ax_setmulti __P((struct ax_softc *));
|
|
static void ax_reset __P((struct ax_softc *));
|
|
static int ax_list_rx_init __P((struct ax_softc *));
|
|
static int ax_list_tx_init __P((struct ax_softc *));
|
|
|
|
#define AX_SETBIT(sc, reg, x) \
|
|
CSR_WRITE_4(sc, reg, \
|
|
CSR_READ_4(sc, reg) | x)
|
|
|
|
#define AX_CLRBIT(sc, reg, x) \
|
|
CSR_WRITE_4(sc, reg, \
|
|
CSR_READ_4(sc, reg) & ~x)
|
|
|
|
#define SIO_SET(x) \
|
|
CSR_WRITE_4(sc, AX_SIO, \
|
|
CSR_READ_4(sc, AX_SIO) | x)
|
|
|
|
#define SIO_CLR(x) \
|
|
CSR_WRITE_4(sc, AX_SIO, \
|
|
CSR_READ_4(sc, AX_SIO) & ~x)
|
|
|
|
static void ax_delay(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
int idx;
|
|
|
|
for (idx = (300 / 33) + 1; idx > 0; idx--)
|
|
CSR_READ_4(sc, AX_BUSCTL);
|
|
}
|
|
|
|
static void ax_eeprom_idle(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
register int i;
|
|
|
|
CSR_WRITE_4(sc, AX_SIO, AX_SIO_EESEL);
|
|
ax_delay(sc);
|
|
AX_SETBIT(sc, AX_SIO, AX_SIO_ROMCTL_READ);
|
|
ax_delay(sc);
|
|
AX_SETBIT(sc, AX_SIO, AX_SIO_EE_CS);
|
|
ax_delay(sc);
|
|
AX_SETBIT(sc, AX_SIO, AX_SIO_EE_CLK);
|
|
ax_delay(sc);
|
|
|
|
for (i = 0; i < 25; i++) {
|
|
AX_CLRBIT(sc, AX_SIO, AX_SIO_EE_CLK);
|
|
ax_delay(sc);
|
|
AX_SETBIT(sc, AX_SIO, AX_SIO_EE_CLK);
|
|
ax_delay(sc);
|
|
}
|
|
|
|
AX_CLRBIT(sc, AX_SIO, AX_SIO_EE_CLK);
|
|
ax_delay(sc);
|
|
AX_CLRBIT(sc, AX_SIO, AX_SIO_EE_CS);
|
|
ax_delay(sc);
|
|
CSR_WRITE_4(sc, AX_SIO, 0x00000000);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Send a read command and address to the EEPROM, check for ACK.
|
|
*/
|
|
static void ax_eeprom_putbyte(sc, addr)
|
|
struct ax_softc *sc;
|
|
int addr;
|
|
{
|
|
register int d, i;
|
|
|
|
d = addr | AX_EECMD_READ;
|
|
|
|
/*
|
|
* Feed in each bit and stobe the clock.
|
|
*/
|
|
for (i = 0x400; i; i >>= 1) {
|
|
if (d & i) {
|
|
SIO_SET(AX_SIO_EE_DATAIN);
|
|
} else {
|
|
SIO_CLR(AX_SIO_EE_DATAIN);
|
|
}
|
|
ax_delay(sc);
|
|
SIO_SET(AX_SIO_EE_CLK);
|
|
ax_delay(sc);
|
|
SIO_CLR(AX_SIO_EE_CLK);
|
|
ax_delay(sc);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Read a word of data stored in the EEPROM at address 'addr.'
|
|
*/
|
|
static void ax_eeprom_getword(sc, addr, dest)
|
|
struct ax_softc *sc;
|
|
int addr;
|
|
u_int16_t *dest;
|
|
{
|
|
register int i;
|
|
u_int16_t word = 0;
|
|
|
|
/* Force EEPROM to idle state. */
|
|
ax_eeprom_idle(sc);
|
|
|
|
/* Enter EEPROM access mode. */
|
|
CSR_WRITE_4(sc, AX_SIO, AX_SIO_EESEL);
|
|
ax_delay(sc);
|
|
AX_SETBIT(sc, AX_SIO, AX_SIO_ROMCTL_READ);
|
|
ax_delay(sc);
|
|
AX_SETBIT(sc, AX_SIO, AX_SIO_EE_CS);
|
|
ax_delay(sc);
|
|
AX_SETBIT(sc, AX_SIO, AX_SIO_EE_CLK);
|
|
ax_delay(sc);
|
|
|
|
/*
|
|
* Send address of word we want to read.
|
|
*/
|
|
ax_eeprom_putbyte(sc, addr);
|
|
|
|
/*
|
|
* Start reading bits from EEPROM.
|
|
*/
|
|
for (i = 0x8000; i; i >>= 1) {
|
|
SIO_SET(AX_SIO_EE_CLK);
|
|
ax_delay(sc);
|
|
if (CSR_READ_4(sc, AX_SIO) & AX_SIO_EE_DATAOUT)
|
|
word |= i;
|
|
ax_delay(sc);
|
|
SIO_CLR(AX_SIO_EE_CLK);
|
|
ax_delay(sc);
|
|
}
|
|
|
|
/* Turn off EEPROM access mode. */
|
|
ax_eeprom_idle(sc);
|
|
|
|
*dest = word;
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Read a sequence of words from the EEPROM.
|
|
*/
|
|
static void ax_read_eeprom(sc, dest, off, cnt, swap)
|
|
struct ax_softc *sc;
|
|
caddr_t dest;
|
|
int off;
|
|
int cnt;
|
|
int swap;
|
|
{
|
|
int i;
|
|
u_int16_t word = 0, *ptr;
|
|
|
|
for (i = 0; i < cnt; i++) {
|
|
ax_eeprom_getword(sc, off + i, &word);
|
|
ptr = (u_int16_t *)(dest + (i * 2));
|
|
if (swap)
|
|
*ptr = ntohs(word);
|
|
else
|
|
*ptr = word;
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Write a bit to the MII bus.
|
|
*/
|
|
static void ax_mii_writebit(sc, bit)
|
|
struct ax_softc *sc;
|
|
int bit;
|
|
{
|
|
if (bit)
|
|
CSR_WRITE_4(sc, AX_SIO, AX_SIO_ROMCTL_WRITE|AX_SIO_MII_DATAOUT);
|
|
else
|
|
CSR_WRITE_4(sc, AX_SIO, AX_SIO_ROMCTL_WRITE);
|
|
|
|
AX_SETBIT(sc, AX_SIO, AX_SIO_MII_CLK);
|
|
AX_CLRBIT(sc, AX_SIO, AX_SIO_MII_CLK);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Read a bit from the MII bus.
|
|
*/
|
|
static int ax_mii_readbit(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
CSR_WRITE_4(sc, AX_SIO, AX_SIO_ROMCTL_READ|AX_SIO_MII_DIR);
|
|
CSR_READ_4(sc, AX_SIO);
|
|
AX_SETBIT(sc, AX_SIO, AX_SIO_MII_CLK);
|
|
AX_CLRBIT(sc, AX_SIO, AX_SIO_MII_CLK);
|
|
if (CSR_READ_4(sc, AX_SIO) & AX_SIO_MII_DATAIN)
|
|
return(1);
|
|
|
|
return(0);
|
|
}
|
|
|
|
/*
|
|
* Sync the PHYs by setting data bit and strobing the clock 32 times.
|
|
*/
|
|
static void ax_mii_sync(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
register int i;
|
|
|
|
CSR_WRITE_4(sc, AX_SIO, AX_SIO_ROMCTL_WRITE);
|
|
|
|
for (i = 0; i < 32; i++)
|
|
ax_mii_writebit(sc, 1);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Clock a series of bits through the MII.
|
|
*/
|
|
static void ax_mii_send(sc, bits, cnt)
|
|
struct ax_softc *sc;
|
|
u_int32_t bits;
|
|
int cnt;
|
|
{
|
|
int i;
|
|
|
|
for (i = (0x1 << (cnt - 1)); i; i >>= 1)
|
|
ax_mii_writebit(sc, bits & i);
|
|
}
|
|
|
|
/*
|
|
* Read an PHY register through the MII.
|
|
*/
|
|
static int ax_mii_readreg(sc, frame)
|
|
struct ax_softc *sc;
|
|
struct ax_mii_frame *frame;
|
|
|
|
{
|
|
int i, ack, s;
|
|
|
|
s = splimp();
|
|
|
|
/*
|
|
* Set up frame for RX.
|
|
*/
|
|
frame->mii_stdelim = AX_MII_STARTDELIM;
|
|
frame->mii_opcode = AX_MII_READOP;
|
|
frame->mii_turnaround = 0;
|
|
frame->mii_data = 0;
|
|
|
|
/*
|
|
* Sync the PHYs.
|
|
*/
|
|
ax_mii_sync(sc);
|
|
|
|
/*
|
|
* Send command/address info.
|
|
*/
|
|
ax_mii_send(sc, frame->mii_stdelim, 2);
|
|
ax_mii_send(sc, frame->mii_opcode, 2);
|
|
ax_mii_send(sc, frame->mii_phyaddr, 5);
|
|
ax_mii_send(sc, frame->mii_regaddr, 5);
|
|
|
|
#ifdef notdef
|
|
/* Idle bit */
|
|
ax_mii_writebit(sc, 1);
|
|
ax_mii_writebit(sc, 0);
|
|
#endif
|
|
|
|
/* Check for ack */
|
|
ack = ax_mii_readbit(sc);
|
|
|
|
/*
|
|
* Now try reading data bits. If the ack failed, we still
|
|
* need to clock through 16 cycles to keep the PHY(s) in sync.
|
|
*/
|
|
if (ack) {
|
|
for(i = 0; i < 16; i++) {
|
|
ax_mii_readbit(sc);
|
|
}
|
|
goto fail;
|
|
}
|
|
|
|
for (i = 0x8000; i; i >>= 1) {
|
|
if (!ack) {
|
|
if (ax_mii_readbit(sc))
|
|
frame->mii_data |= i;
|
|
}
|
|
}
|
|
|
|
fail:
|
|
|
|
ax_mii_writebit(sc, 0);
|
|
ax_mii_writebit(sc, 0);
|
|
|
|
splx(s);
|
|
|
|
if (ack)
|
|
return(1);
|
|
return(0);
|
|
}
|
|
|
|
/*
|
|
* Write to a PHY register through the MII.
|
|
*/
|
|
static int ax_mii_writereg(sc, frame)
|
|
struct ax_softc *sc;
|
|
struct ax_mii_frame *frame;
|
|
|
|
{
|
|
int s;
|
|
|
|
s = splimp();
|
|
/*
|
|
* Set up frame for TX.
|
|
*/
|
|
|
|
frame->mii_stdelim = AX_MII_STARTDELIM;
|
|
frame->mii_opcode = AX_MII_WRITEOP;
|
|
frame->mii_turnaround = AX_MII_TURNAROUND;
|
|
|
|
/*
|
|
* Sync the PHYs.
|
|
*/
|
|
ax_mii_sync(sc);
|
|
|
|
ax_mii_send(sc, frame->mii_stdelim, 2);
|
|
ax_mii_send(sc, frame->mii_opcode, 2);
|
|
ax_mii_send(sc, frame->mii_phyaddr, 5);
|
|
ax_mii_send(sc, frame->mii_regaddr, 5);
|
|
ax_mii_send(sc, frame->mii_turnaround, 2);
|
|
ax_mii_send(sc, frame->mii_data, 16);
|
|
|
|
/* Idle bit. */
|
|
ax_mii_writebit(sc, 0);
|
|
ax_mii_writebit(sc, 0);
|
|
|
|
splx(s);
|
|
|
|
return(0);
|
|
}
|
|
|
|
static u_int16_t ax_phy_readreg(sc, reg)
|
|
struct ax_softc *sc;
|
|
int reg;
|
|
{
|
|
struct ax_mii_frame frame;
|
|
|
|
bzero((char *)&frame, sizeof(frame));
|
|
|
|
frame.mii_phyaddr = sc->ax_phy_addr;
|
|
frame.mii_regaddr = reg;
|
|
ax_mii_readreg(sc, &frame);
|
|
|
|
return(frame.mii_data);
|
|
}
|
|
|
|
static void ax_phy_writereg(sc, reg, data)
|
|
struct ax_softc *sc;
|
|
int reg;
|
|
int data;
|
|
{
|
|
struct ax_mii_frame frame;
|
|
|
|
bzero((char *)&frame, sizeof(frame));
|
|
|
|
frame.mii_phyaddr = sc->ax_phy_addr;
|
|
frame.mii_regaddr = reg;
|
|
frame.mii_data = data;
|
|
|
|
ax_mii_writereg(sc, &frame);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Calculate CRC of a multicast group address, return the lower 6 bits.
|
|
*/
|
|
static u_int32_t ax_calchash(addr)
|
|
caddr_t addr;
|
|
{
|
|
u_int32_t crc, carry;
|
|
int i, j;
|
|
u_int8_t c;
|
|
|
|
/* Compute CRC for the address value. */
|
|
crc = 0xFFFFFFFF; /* initial value */
|
|
|
|
for (i = 0; i < 6; i++) {
|
|
c = *(addr + i);
|
|
for (j = 0; j < 8; j++) {
|
|
carry = ((crc & 0x80000000) ? 1 : 0) ^ (c & 0x01);
|
|
crc <<= 1;
|
|
c >>= 1;
|
|
if (carry)
|
|
crc = (crc ^ 0x04c11db6) | carry;
|
|
}
|
|
}
|
|
|
|
/* return the filter bit position */
|
|
return((crc >> 26) & 0x0000003F);
|
|
}
|
|
|
|
static void ax_setmulti(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
struct ifnet *ifp;
|
|
int h = 0;
|
|
u_int32_t hashes[2] = { 0, 0 };
|
|
struct ifmultiaddr *ifma;
|
|
u_int32_t rxfilt;
|
|
|
|
ifp = &sc->arpcom.ac_if;
|
|
|
|
rxfilt = CSR_READ_4(sc, AX_NETCFG);
|
|
|
|
if (ifp->if_flags & IFF_ALLMULTI || ifp->if_flags & IFF_PROMISC) {
|
|
rxfilt |= AX_NETCFG_RX_ALLMULTI;
|
|
CSR_WRITE_4(sc, AX_NETCFG, rxfilt);
|
|
return;
|
|
} else
|
|
rxfilt &= ~AX_NETCFG_RX_ALLMULTI;
|
|
|
|
/* first, zot all the existing hash bits */
|
|
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_MAR0);
|
|
CSR_WRITE_4(sc, AX_FILTDATA, 0);
|
|
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_MAR1);
|
|
CSR_WRITE_4(sc, AX_FILTDATA, 0);
|
|
|
|
/* now program new ones */
|
|
for (ifma = ifp->if_multiaddrs.lh_first; ifma != NULL;
|
|
ifma = ifma->ifma_link.le_next) {
|
|
if (ifma->ifma_addr->sa_family != AF_LINK)
|
|
continue;
|
|
h = ax_calchash(LLADDR((struct sockaddr_dl *)ifma->ifma_addr));
|
|
if (h < 32)
|
|
hashes[0] |= (1 << h);
|
|
else
|
|
hashes[1] |= (1 << (h - 32));
|
|
}
|
|
|
|
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_MAR0);
|
|
CSR_WRITE_4(sc, AX_FILTDATA, hashes[0]);
|
|
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_MAR1);
|
|
CSR_WRITE_4(sc, AX_FILTDATA, hashes[1]);
|
|
CSR_WRITE_4(sc, AX_NETCFG, rxfilt);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Initiate an autonegotiation session.
|
|
*/
|
|
static void ax_autoneg_xmit(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
u_int16_t phy_sts;
|
|
|
|
ax_phy_writereg(sc, PHY_BMCR, PHY_BMCR_RESET);
|
|
DELAY(500);
|
|
while(ax_phy_readreg(sc, PHY_BMCR)
|
|
& PHY_BMCR_RESET);
|
|
|
|
phy_sts = ax_phy_readreg(sc, PHY_BMCR);
|
|
phy_sts |= PHY_BMCR_AUTONEGENBL|PHY_BMCR_AUTONEGRSTR;
|
|
ax_phy_writereg(sc, PHY_BMCR, phy_sts);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Invoke autonegotiation on a PHY.
|
|
*/
|
|
static void ax_autoneg_mii(sc, flag, verbose)
|
|
struct ax_softc *sc;
|
|
int flag;
|
|
int verbose;
|
|
{
|
|
u_int16_t phy_sts = 0, media, advert, ability;
|
|
struct ifnet *ifp;
|
|
struct ifmedia *ifm;
|
|
|
|
ifm = &sc->ifmedia;
|
|
ifp = &sc->arpcom.ac_if;
|
|
|
|
ifm->ifm_media = IFM_ETHER | IFM_AUTO;
|
|
|
|
/*
|
|
* The 100baseT4 PHY on the 3c905-T4 has the 'autoneg supported'
|
|
* bit cleared in the status register, but has the 'autoneg enabled'
|
|
* bit set in the control register. This is a contradiction, and
|
|
* I'm not sure how to handle it. If you want to force an attempt
|
|
* to autoneg for 100baseT4 PHYs, #define FORCE_AUTONEG_TFOUR
|
|
* and see what happens.
|
|
*/
|
|
#ifndef FORCE_AUTONEG_TFOUR
|
|
/*
|
|
* First, see if autoneg is supported. If not, there's
|
|
* no point in continuing.
|
|
*/
|
|
phy_sts = ax_phy_readreg(sc, PHY_BMSR);
|
|
if (!(phy_sts & PHY_BMSR_CANAUTONEG)) {
|
|
if (verbose)
|
|
printf("ax%d: autonegotiation not supported\n",
|
|
sc->ax_unit);
|
|
ifm->ifm_media = IFM_ETHER|IFM_10_T|IFM_HDX;
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
switch (flag) {
|
|
case AX_FLAG_FORCEDELAY:
|
|
/*
|
|
* XXX Never use this option anywhere but in the probe
|
|
* routine: making the kernel stop dead in its tracks
|
|
* for three whole seconds after we've gone multi-user
|
|
* is really bad manners.
|
|
*/
|
|
ax_autoneg_xmit(sc);
|
|
DELAY(5000000);
|
|
break;
|
|
case AX_FLAG_SCHEDDELAY:
|
|
/*
|
|
* Wait for the transmitter to go idle before starting
|
|
* an autoneg session, otherwise ax_start() may clobber
|
|
* our timeout, and we don't want to allow transmission
|
|
* during an autoneg session since that can screw it up.
|
|
*/
|
|
if (sc->ax_cdata.ax_tx_head != NULL) {
|
|
sc->ax_want_auto = 1;
|
|
return;
|
|
}
|
|
ax_autoneg_xmit(sc);
|
|
ifp->if_timer = 5;
|
|
sc->ax_autoneg = 1;
|
|
sc->ax_want_auto = 0;
|
|
return;
|
|
break;
|
|
case AX_FLAG_DELAYTIMEO:
|
|
ifp->if_timer = 0;
|
|
sc->ax_autoneg = 0;
|
|
break;
|
|
default:
|
|
printf("ax%d: invalid autoneg flag: %d\n", sc->ax_unit, flag);
|
|
return;
|
|
}
|
|
|
|
if (ax_phy_readreg(sc, PHY_BMSR) & PHY_BMSR_AUTONEGCOMP) {
|
|
if (verbose)
|
|
printf("ax%d: autoneg complete, ", sc->ax_unit);
|
|
phy_sts = ax_phy_readreg(sc, PHY_BMSR);
|
|
} else {
|
|
if (verbose)
|
|
printf("ax%d: autoneg not complete, ", sc->ax_unit);
|
|
}
|
|
|
|
media = ax_phy_readreg(sc, PHY_BMCR);
|
|
|
|
/* Link is good. Report modes and set duplex mode. */
|
|
if (ax_phy_readreg(sc, PHY_BMSR) & PHY_BMSR_LINKSTAT) {
|
|
if (verbose)
|
|
printf("link status good ");
|
|
advert = ax_phy_readreg(sc, PHY_ANAR);
|
|
ability = ax_phy_readreg(sc, PHY_LPAR);
|
|
|
|
if (advert & PHY_ANAR_100BT4 && ability & PHY_ANAR_100BT4) {
|
|
ifm->ifm_media = IFM_ETHER|IFM_100_T4;
|
|
media |= PHY_BMCR_SPEEDSEL;
|
|
media &= ~PHY_BMCR_DUPLEX;
|
|
printf("(100baseT4)\n");
|
|
} else if (advert & PHY_ANAR_100BTXFULL &&
|
|
ability & PHY_ANAR_100BTXFULL) {
|
|
ifm->ifm_media = IFM_ETHER|IFM_100_TX|IFM_FDX;
|
|
media |= PHY_BMCR_SPEEDSEL;
|
|
media |= PHY_BMCR_DUPLEX;
|
|
printf("(full-duplex, 100Mbps)\n");
|
|
} else if (advert & PHY_ANAR_100BTXHALF &&
|
|
ability & PHY_ANAR_100BTXHALF) {
|
|
ifm->ifm_media = IFM_ETHER|IFM_100_TX|IFM_HDX;
|
|
media |= PHY_BMCR_SPEEDSEL;
|
|
media &= ~PHY_BMCR_DUPLEX;
|
|
printf("(half-duplex, 100Mbps)\n");
|
|
} else if (advert & PHY_ANAR_10BTFULL &&
|
|
ability & PHY_ANAR_10BTFULL) {
|
|
ifm->ifm_media = IFM_ETHER|IFM_10_T|IFM_FDX;
|
|
media &= ~PHY_BMCR_SPEEDSEL;
|
|
media |= PHY_BMCR_DUPLEX;
|
|
printf("(full-duplex, 10Mbps)\n");
|
|
} else if (advert & PHY_ANAR_10BTHALF &&
|
|
ability & PHY_ANAR_10BTHALF) {
|
|
ifm->ifm_media = IFM_ETHER|IFM_10_T|IFM_HDX;
|
|
media &= ~PHY_BMCR_SPEEDSEL;
|
|
media &= ~PHY_BMCR_DUPLEX;
|
|
printf("(half-duplex, 10Mbps)\n");
|
|
}
|
|
|
|
media &= ~PHY_BMCR_AUTONEGENBL;
|
|
|
|
/* Set ASIC's duplex mode to match the PHY. */
|
|
ax_setcfg(sc, media);
|
|
ax_phy_writereg(sc, PHY_BMCR, media);
|
|
} else {
|
|
if (verbose)
|
|
printf("no carrier\n");
|
|
}
|
|
|
|
ax_init(sc);
|
|
|
|
if (sc->ax_tx_pend) {
|
|
sc->ax_autoneg = 0;
|
|
sc->ax_tx_pend = 0;
|
|
ax_start(ifp);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
static void ax_getmode_mii(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
u_int16_t bmsr;
|
|
struct ifnet *ifp;
|
|
|
|
ifp = &sc->arpcom.ac_if;
|
|
|
|
bmsr = ax_phy_readreg(sc, PHY_BMSR);
|
|
if (bootverbose)
|
|
printf("ax%d: PHY status word: %x\n", sc->ax_unit, bmsr);
|
|
|
|
/* fallback */
|
|
sc->ifmedia.ifm_media = IFM_ETHER|IFM_10_T|IFM_HDX;
|
|
|
|
if (bmsr & PHY_BMSR_10BTHALF) {
|
|
if (bootverbose)
|
|
printf("ax%d: 10Mbps half-duplex mode supported\n",
|
|
sc->ax_unit);
|
|
ifmedia_add(&sc->ifmedia,
|
|
IFM_ETHER|IFM_10_T|IFM_HDX, 0, NULL);
|
|
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_10_T, 0, NULL);
|
|
}
|
|
|
|
if (bmsr & PHY_BMSR_10BTFULL) {
|
|
if (bootverbose)
|
|
printf("ax%d: 10Mbps full-duplex mode supported\n",
|
|
sc->ax_unit);
|
|
ifmedia_add(&sc->ifmedia,
|
|
IFM_ETHER|IFM_10_T|IFM_FDX, 0, NULL);
|
|
sc->ifmedia.ifm_media = IFM_ETHER|IFM_10_T|IFM_FDX;
|
|
}
|
|
|
|
if (bmsr & PHY_BMSR_100BTXHALF) {
|
|
if (bootverbose)
|
|
printf("ax%d: 100Mbps half-duplex mode supported\n",
|
|
sc->ax_unit);
|
|
ifp->if_baudrate = 100000000;
|
|
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_100_TX, 0, NULL);
|
|
ifmedia_add(&sc->ifmedia,
|
|
IFM_ETHER|IFM_100_TX|IFM_HDX, 0, NULL);
|
|
sc->ifmedia.ifm_media = IFM_ETHER|IFM_100_TX|IFM_HDX;
|
|
}
|
|
|
|
if (bmsr & PHY_BMSR_100BTXFULL) {
|
|
if (bootverbose)
|
|
printf("ax%d: 100Mbps full-duplex mode supported\n",
|
|
sc->ax_unit);
|
|
ifp->if_baudrate = 100000000;
|
|
ifmedia_add(&sc->ifmedia,
|
|
IFM_ETHER|IFM_100_TX|IFM_FDX, 0, NULL);
|
|
sc->ifmedia.ifm_media = IFM_ETHER|IFM_100_TX|IFM_FDX;
|
|
}
|
|
|
|
/* Some also support 100BaseT4. */
|
|
if (bmsr & PHY_BMSR_100BT4) {
|
|
if (bootverbose)
|
|
printf("ax%d: 100baseT4 mode supported\n", sc->ax_unit);
|
|
ifp->if_baudrate = 100000000;
|
|
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_100_T4, 0, NULL);
|
|
sc->ifmedia.ifm_media = IFM_ETHER|IFM_100_T4;
|
|
#ifdef FORCE_AUTONEG_TFOUR
|
|
if (bootverbose)
|
|
printf("ax%d: forcing on autoneg support for BT4\n",
|
|
sc->ax_unit);
|
|
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_AUTO, 0 NULL):
|
|
sc->ifmedia.ifm_media = IFM_ETHER|IFM_AUTO;
|
|
#endif
|
|
}
|
|
|
|
if (bmsr & PHY_BMSR_CANAUTONEG) {
|
|
if (bootverbose)
|
|
printf("ax%d: autoneg supported\n", sc->ax_unit);
|
|
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_AUTO, 0, NULL);
|
|
sc->ifmedia.ifm_media = IFM_ETHER|IFM_AUTO;
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Set speed and duplex mode.
|
|
*/
|
|
static void ax_setmode_mii(sc, media)
|
|
struct ax_softc *sc;
|
|
int media;
|
|
{
|
|
u_int16_t bmcr;
|
|
struct ifnet *ifp;
|
|
|
|
ifp = &sc->arpcom.ac_if;
|
|
|
|
/*
|
|
* If an autoneg session is in progress, stop it.
|
|
*/
|
|
if (sc->ax_autoneg) {
|
|
printf("ax%d: canceling autoneg session\n", sc->ax_unit);
|
|
ifp->if_timer = sc->ax_autoneg = sc->ax_want_auto = 0;
|
|
bmcr = ax_phy_readreg(sc, PHY_BMCR);
|
|
bmcr &= ~PHY_BMCR_AUTONEGENBL;
|
|
ax_phy_writereg(sc, PHY_BMCR, bmcr);
|
|
}
|
|
|
|
printf("ax%d: selecting MII, ", sc->ax_unit);
|
|
|
|
bmcr = ax_phy_readreg(sc, PHY_BMCR);
|
|
|
|
bmcr &= ~(PHY_BMCR_AUTONEGENBL|PHY_BMCR_SPEEDSEL|
|
|
PHY_BMCR_DUPLEX|PHY_BMCR_LOOPBK);
|
|
|
|
if (IFM_SUBTYPE(media) == IFM_100_T4) {
|
|
printf("100Mbps/T4, half-duplex\n");
|
|
bmcr |= PHY_BMCR_SPEEDSEL;
|
|
bmcr &= ~PHY_BMCR_DUPLEX;
|
|
}
|
|
|
|
if (IFM_SUBTYPE(media) == IFM_100_TX) {
|
|
printf("100Mbps, ");
|
|
bmcr |= PHY_BMCR_SPEEDSEL;
|
|
}
|
|
|
|
if (IFM_SUBTYPE(media) == IFM_10_T) {
|
|
printf("10Mbps, ");
|
|
bmcr &= ~PHY_BMCR_SPEEDSEL;
|
|
}
|
|
|
|
if ((media & IFM_GMASK) == IFM_FDX) {
|
|
printf("full duplex\n");
|
|
bmcr |= PHY_BMCR_DUPLEX;
|
|
} else {
|
|
printf("half duplex\n");
|
|
bmcr &= ~PHY_BMCR_DUPLEX;
|
|
}
|
|
|
|
ax_setcfg(sc, bmcr);
|
|
ax_phy_writereg(sc, PHY_BMCR, bmcr);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Set speed and duplex mode on internal transceiver.
|
|
*/
|
|
static void ax_setmode(sc, media, verbose)
|
|
struct ax_softc *sc;
|
|
int media;
|
|
int verbose;
|
|
{
|
|
struct ifnet *ifp;
|
|
u_int32_t mode;
|
|
|
|
ifp = &sc->arpcom.ac_if;
|
|
|
|
if (verbose)
|
|
printf("ax%d: selecting internal xcvr, ", sc->ax_unit);
|
|
|
|
mode = CSR_READ_4(sc, AX_NETCFG);
|
|
|
|
mode &= ~(AX_NETCFG_FULLDUPLEX|AX_NETCFG_PORTSEL|
|
|
AX_NETCFG_PCS|AX_NETCFG_SCRAMBLER|AX_NETCFG_SPEEDSEL);
|
|
|
|
if (IFM_SUBTYPE(media) == IFM_100_T4) {
|
|
if (verbose)
|
|
printf("100Mbps/T4, half-duplex\n");
|
|
mode |= AX_NETCFG_PORTSEL|AX_NETCFG_PCS|AX_NETCFG_SCRAMBLER;
|
|
}
|
|
|
|
if (IFM_SUBTYPE(media) == IFM_100_TX) {
|
|
if (verbose)
|
|
printf("100Mbps, ");
|
|
mode |= AX_NETCFG_PORTSEL|AX_NETCFG_PCS|AX_NETCFG_SCRAMBLER;
|
|
}
|
|
|
|
if (IFM_SUBTYPE(media) == IFM_10_T) {
|
|
if (verbose)
|
|
printf("10Mbps, ");
|
|
mode &= ~AX_NETCFG_PORTSEL;
|
|
mode |= AX_NETCFG_SPEEDSEL;
|
|
}
|
|
|
|
if ((media & IFM_GMASK) == IFM_FDX) {
|
|
if (verbose)
|
|
printf("full duplex\n");
|
|
mode |= AX_NETCFG_FULLDUPLEX;
|
|
} else {
|
|
if (verbose)
|
|
printf("half duplex\n");
|
|
mode &= ~AX_NETCFG_FULLDUPLEX;
|
|
}
|
|
|
|
CSR_WRITE_4(sc, AX_NETCFG, mode);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* In order to fiddle with the
|
|
* 'full-duplex' and '100Mbps' bits in the netconfig register, we
|
|
* first have to put the transmit and/or receive logic in the idle state.
|
|
*/
|
|
static void ax_setcfg(sc, bmcr)
|
|
struct ax_softc *sc;
|
|
int bmcr;
|
|
{
|
|
int i, restart = 0;
|
|
|
|
if (CSR_READ_4(sc, AX_NETCFG) & (AX_NETCFG_TX_ON|AX_NETCFG_RX_ON)) {
|
|
restart = 1;
|
|
AX_CLRBIT(sc, AX_NETCFG, (AX_NETCFG_TX_ON|AX_NETCFG_RX_ON));
|
|
|
|
for (i = 0; i < AX_TIMEOUT; i++) {
|
|
DELAY(10);
|
|
if (CSR_READ_4(sc, AX_ISR) & AX_ISR_TX_IDLE)
|
|
break;
|
|
}
|
|
|
|
if (i == AX_TIMEOUT)
|
|
printf("ax%d: failed to force tx and "
|
|
"rx to idle state\n", sc->ax_unit);
|
|
|
|
}
|
|
|
|
if (bmcr & PHY_BMCR_SPEEDSEL)
|
|
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_SPEEDSEL);
|
|
else
|
|
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_SPEEDSEL);
|
|
|
|
if (bmcr & PHY_BMCR_DUPLEX)
|
|
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_FULLDUPLEX);
|
|
else
|
|
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_FULLDUPLEX);
|
|
|
|
if (restart)
|
|
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_TX_ON|AX_NETCFG_RX_ON);
|
|
|
|
return;
|
|
}
|
|
|
|
static void ax_reset(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
register int i;
|
|
|
|
AX_SETBIT(sc, AX_BUSCTL, AX_BUSCTL_RESET);
|
|
|
|
for (i = 0; i < AX_TIMEOUT; i++) {
|
|
DELAY(10);
|
|
if (!(CSR_READ_4(sc, AX_BUSCTL) & AX_BUSCTL_RESET))
|
|
break;
|
|
}
|
|
#ifdef notdef
|
|
if (i == AX_TIMEOUT)
|
|
printf("ax%d: reset never completed!\n", sc->ax_unit);
|
|
#endif
|
|
CSR_WRITE_4(sc, AX_BUSCTL, AX_BUSCTL_CONFIG);
|
|
|
|
/* Wait a little while for the chip to get its brains in order. */
|
|
DELAY(1000);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Probe for an ASIX chip. Check the PCI vendor and device
|
|
* IDs against our list and return a device name if we find a match.
|
|
*/
|
|
static const char *
|
|
ax_probe(config_id, device_id)
|
|
pcici_t config_id;
|
|
pcidi_t device_id;
|
|
{
|
|
struct ax_type *t;
|
|
|
|
t = ax_devs;
|
|
|
|
while(t->ax_name != NULL) {
|
|
if ((device_id & 0xFFFF) == t->ax_vid &&
|
|
((device_id >> 16) & 0xFFFF) == t->ax_did) {
|
|
return(t->ax_name);
|
|
}
|
|
t++;
|
|
}
|
|
|
|
return(NULL);
|
|
}
|
|
|
|
/*
|
|
* Attach the interface. Allocate softc structures, do ifmedia
|
|
* setup and ethernet/BPF attach.
|
|
*/
|
|
static void
|
|
ax_attach(config_id, unit)
|
|
pcici_t config_id;
|
|
int unit;
|
|
{
|
|
int s, i;
|
|
#ifndef AX_USEIOSPACE
|
|
vm_offset_t pbase, vbase;
|
|
#endif
|
|
u_char eaddr[ETHER_ADDR_LEN];
|
|
u_int32_t command;
|
|
struct ax_softc *sc;
|
|
struct ifnet *ifp;
|
|
int media = IFM_ETHER|IFM_100_TX|IFM_FDX;
|
|
unsigned int round;
|
|
caddr_t roundptr;
|
|
struct ax_type *p;
|
|
u_int16_t phy_vid, phy_did, phy_sts;
|
|
|
|
s = splimp();
|
|
|
|
sc = malloc(sizeof(struct ax_softc), M_DEVBUF, M_NOWAIT);
|
|
if (sc == NULL) {
|
|
printf("ax%d: no memory for softc struct!\n", unit);
|
|
goto fail;
|
|
}
|
|
bzero(sc, sizeof(struct ax_softc));
|
|
|
|
/*
|
|
* Handle power management nonsense.
|
|
*/
|
|
|
|
command = pci_conf_read(config_id, AX_PCI_CAPID) & 0x000000FF;
|
|
if (command == 0x01) {
|
|
|
|
command = pci_conf_read(config_id, AX_PCI_PWRMGMTCTRL);
|
|
if (command & AX_PSTATE_MASK) {
|
|
u_int32_t iobase, membase, irq;
|
|
|
|
/* Save important PCI config data. */
|
|
iobase = pci_conf_read(config_id, AX_PCI_LOIO);
|
|
membase = pci_conf_read(config_id, AX_PCI_LOMEM);
|
|
irq = pci_conf_read(config_id, AX_PCI_INTLINE);
|
|
|
|
/* Reset the power state. */
|
|
printf("ax%d: chip is in D%d power mode "
|
|
"-- setting to D0\n", unit, command & AX_PSTATE_MASK);
|
|
command &= 0xFFFFFFFC;
|
|
pci_conf_write(config_id, AX_PCI_PWRMGMTCTRL, command);
|
|
|
|
/* Restore PCI config data. */
|
|
pci_conf_write(config_id, AX_PCI_LOIO, iobase);
|
|
pci_conf_write(config_id, AX_PCI_LOMEM, membase);
|
|
pci_conf_write(config_id, AX_PCI_INTLINE, irq);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Map control/status registers.
|
|
*/
|
|
command = pci_conf_read(config_id, PCI_COMMAND_STATUS_REG);
|
|
command |= (PCIM_CMD_PORTEN|PCIM_CMD_MEMEN|PCIM_CMD_BUSMASTEREN);
|
|
pci_conf_write(config_id, PCI_COMMAND_STATUS_REG, command);
|
|
command = pci_conf_read(config_id, PCI_COMMAND_STATUS_REG);
|
|
|
|
#ifdef AX_USEIOSPACE
|
|
if (!(command & PCIM_CMD_PORTEN)) {
|
|
printf("ax%d: failed to enable I/O ports!\n", unit);
|
|
free(sc, M_DEVBUF);
|
|
goto fail;
|
|
}
|
|
|
|
if (!pci_map_port(config_id, AX_PCI_LOIO,
|
|
(u_short *)&(sc->ax_bhandle))) {
|
|
printf ("ax%d: couldn't map ports\n", unit);
|
|
goto fail;
|
|
}
|
|
sc->ax_btag = I386_BUS_SPACE_IO;
|
|
#else
|
|
if (!(command & PCIM_CMD_MEMEN)) {
|
|
printf("ax%d: failed to enable memory mapping!\n", unit);
|
|
goto fail;
|
|
}
|
|
|
|
if (!pci_map_mem(config_id, AX_PCI_LOMEM, &vbase, &pbase)) {
|
|
printf ("ax%d: couldn't map memory\n", unit);
|
|
goto fail;
|
|
}
|
|
sc->ax_btag = I386_BUS_SPACE_MEM;
|
|
sc->ax_bhandle = vbase;
|
|
#endif
|
|
|
|
/* Allocate interrupt */
|
|
if (!pci_map_int(config_id, ax_intr, sc, &net_imask)) {
|
|
printf("ax%d: couldn't map interrupt\n", unit);
|
|
goto fail;
|
|
}
|
|
|
|
/* Reset the adapter. */
|
|
ax_reset(sc);
|
|
|
|
/*
|
|
* Get station address from the EEPROM.
|
|
*/
|
|
ax_read_eeprom(sc, (caddr_t)&eaddr, AX_EE_NODEADDR, 3, 0);
|
|
|
|
/*
|
|
* An ASIX chip was detected. Inform the world.
|
|
*/
|
|
printf("ax%d: Ethernet address: %6D\n", unit, eaddr, ":");
|
|
|
|
sc->ax_unit = unit;
|
|
bcopy(eaddr, (char *)&sc->arpcom.ac_enaddr, ETHER_ADDR_LEN);
|
|
|
|
sc->ax_ldata_ptr = malloc(sizeof(struct ax_list_data) + 8,
|
|
M_DEVBUF, M_NOWAIT);
|
|
if (sc->ax_ldata_ptr == NULL) {
|
|
free(sc, M_DEVBUF);
|
|
printf("ax%d: no memory for list buffers!\n", unit);
|
|
goto fail;
|
|
}
|
|
|
|
sc->ax_ldata = (struct ax_list_data *)sc->ax_ldata_ptr;
|
|
round = (unsigned int)sc->ax_ldata_ptr & 0xF;
|
|
roundptr = sc->ax_ldata_ptr;
|
|
for (i = 0; i < 8; i++) {
|
|
if (round % 8) {
|
|
round++;
|
|
roundptr++;
|
|
} else
|
|
break;
|
|
}
|
|
sc->ax_ldata = (struct ax_list_data *)roundptr;
|
|
bzero(sc->ax_ldata, sizeof(struct ax_list_data));
|
|
|
|
ifp = &sc->arpcom.ac_if;
|
|
ifp->if_softc = sc;
|
|
ifp->if_unit = unit;
|
|
ifp->if_name = "ax";
|
|
ifp->if_mtu = ETHERMTU;
|
|
ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
|
|
ifp->if_ioctl = ax_ioctl;
|
|
ifp->if_output = ether_output;
|
|
ifp->if_start = ax_start;
|
|
ifp->if_watchdog = ax_watchdog;
|
|
ifp->if_init = ax_init;
|
|
ifp->if_baudrate = 10000000;
|
|
|
|
|
|
if (bootverbose)
|
|
printf("ax%d: probing for a PHY\n", sc->ax_unit);
|
|
for (i = AX_PHYADDR_MIN; i < AX_PHYADDR_MAX + 1; i++) {
|
|
if (bootverbose)
|
|
printf("ax%d: checking address: %d\n",
|
|
sc->ax_unit, i);
|
|
sc->ax_phy_addr = i;
|
|
ax_phy_writereg(sc, PHY_BMCR, PHY_BMCR_RESET);
|
|
DELAY(500);
|
|
while(ax_phy_readreg(sc, PHY_BMCR)
|
|
& PHY_BMCR_RESET);
|
|
if ((phy_sts = ax_phy_readreg(sc, PHY_BMSR)))
|
|
break;
|
|
}
|
|
if (phy_sts) {
|
|
phy_vid = ax_phy_readreg(sc, PHY_VENID);
|
|
phy_did = ax_phy_readreg(sc, PHY_DEVID);
|
|
if (bootverbose)
|
|
printf("ax%d: found PHY at address %d, ",
|
|
sc->ax_unit, sc->ax_phy_addr);
|
|
if (bootverbose)
|
|
printf("vendor id: %x device id: %x\n",
|
|
phy_vid, phy_did);
|
|
p = ax_phys;
|
|
while(p->ax_vid) {
|
|
if (phy_vid == p->ax_vid &&
|
|
(phy_did | 0x000F) == p->ax_did) {
|
|
sc->ax_pinfo = p;
|
|
break;
|
|
}
|
|
p++;
|
|
}
|
|
if (sc->ax_pinfo == NULL)
|
|
sc->ax_pinfo = &ax_phys[PHY_UNKNOWN];
|
|
if (bootverbose)
|
|
printf("ax%d: PHY type: %s\n",
|
|
sc->ax_unit, sc->ax_pinfo->ax_name);
|
|
} else {
|
|
#ifdef DIAGNOSTIC
|
|
printf("ax%d: MII without any phy!\n", sc->ax_unit);
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* Do ifmedia setup.
|
|
*/
|
|
ifmedia_init(&sc->ifmedia, 0, ax_ifmedia_upd, ax_ifmedia_sts);
|
|
|
|
if (sc->ax_pinfo != NULL) {
|
|
ax_getmode_mii(sc);
|
|
ax_autoneg_mii(sc, AX_FLAG_FORCEDELAY, 1);
|
|
} else {
|
|
ifmedia_add(&sc->ifmedia,
|
|
IFM_ETHER|IFM_10_T|IFM_HDX, 0, NULL);
|
|
ifmedia_add(&sc->ifmedia,
|
|
IFM_ETHER|IFM_10_T|IFM_FDX, 0, NULL);
|
|
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_10_T, 0, NULL);
|
|
ifmedia_add(&sc->ifmedia,
|
|
IFM_ETHER|IFM_100_TX|IFM_HDX, 0, NULL);
|
|
ifmedia_add(&sc->ifmedia,
|
|
IFM_ETHER|IFM_100_TX|IFM_FDX, 0, NULL);
|
|
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_100_TX, 0, NULL);
|
|
ifmedia_add(&sc->ifmedia, IFM_ETHER|IFM_AUTO, 0, NULL);
|
|
}
|
|
|
|
media = sc->ifmedia.ifm_media;
|
|
ax_stop(sc);
|
|
|
|
ifmedia_set(&sc->ifmedia, media);
|
|
|
|
/*
|
|
* Call MI attach routines.
|
|
*/
|
|
if_attach(ifp);
|
|
ether_ifattach(ifp);
|
|
|
|
#if NBPFILTER > 0
|
|
bpfattach(ifp, DLT_EN10MB, sizeof(struct ether_header));
|
|
#endif
|
|
at_shutdown(ax_shutdown, sc, SHUTDOWN_POST_SYNC);
|
|
|
|
fail:
|
|
splx(s);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Initialize the transmit descriptors.
|
|
*/
|
|
static int ax_list_tx_init(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
struct ax_chain_data *cd;
|
|
struct ax_list_data *ld;
|
|
int i;
|
|
|
|
cd = &sc->ax_cdata;
|
|
ld = sc->ax_ldata;
|
|
for (i = 0; i < AX_TX_LIST_CNT; i++) {
|
|
cd->ax_tx_chain[i].ax_ptr = &ld->ax_tx_list[i];
|
|
if (i == (AX_TX_LIST_CNT - 1))
|
|
cd->ax_tx_chain[i].ax_nextdesc =
|
|
&cd->ax_tx_chain[0];
|
|
else
|
|
cd->ax_tx_chain[i].ax_nextdesc =
|
|
&cd->ax_tx_chain[i + 1];
|
|
}
|
|
|
|
cd->ax_tx_free = &cd->ax_tx_chain[0];
|
|
cd->ax_tx_tail = cd->ax_tx_head = NULL;
|
|
|
|
return(0);
|
|
}
|
|
|
|
|
|
/*
|
|
* Initialize the RX descriptors and allocate mbufs for them. Note that
|
|
* we arrange the descriptors in a closed ring, so that the last descriptor
|
|
* points back to the first.
|
|
*/
|
|
static int ax_list_rx_init(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
struct ax_chain_data *cd;
|
|
struct ax_list_data *ld;
|
|
int i;
|
|
|
|
cd = &sc->ax_cdata;
|
|
ld = sc->ax_ldata;
|
|
|
|
for (i = 0; i < AX_RX_LIST_CNT; i++) {
|
|
cd->ax_rx_chain[i].ax_ptr =
|
|
(volatile struct ax_desc *)&ld->ax_rx_list[i];
|
|
if (ax_newbuf(sc, &cd->ax_rx_chain[i]) == ENOBUFS)
|
|
return(ENOBUFS);
|
|
if (i == (AX_RX_LIST_CNT - 1)) {
|
|
cd->ax_rx_chain[i].ax_nextdesc =
|
|
&cd->ax_rx_chain[0];
|
|
ld->ax_rx_list[i].ax_next =
|
|
vtophys(&ld->ax_rx_list[0]);
|
|
} else {
|
|
cd->ax_rx_chain[i].ax_nextdesc =
|
|
&cd->ax_rx_chain[i + 1];
|
|
ld->ax_rx_list[i].ax_next =
|
|
vtophys(&ld->ax_rx_list[i + 1]);
|
|
}
|
|
}
|
|
|
|
cd->ax_rx_head = &cd->ax_rx_chain[0];
|
|
|
|
return(0);
|
|
}
|
|
|
|
/*
|
|
* Initialize an RX descriptor and attach an MBUF cluster.
|
|
* Note: the length fields are only 11 bits wide, which means the
|
|
* largest size we can specify is 2047. This is important because
|
|
* MCLBYTES is 2048, so we have to subtract one otherwise we'll
|
|
* overflow the field and make a mess.
|
|
*/
|
|
static int ax_newbuf(sc, c)
|
|
struct ax_softc *sc;
|
|
struct ax_chain_onefrag *c;
|
|
{
|
|
struct mbuf *m_new = NULL;
|
|
|
|
MGETHDR(m_new, M_DONTWAIT, MT_DATA);
|
|
if (m_new == NULL) {
|
|
printf("ax%d: no memory for rx list -- packet dropped!\n",
|
|
sc->ax_unit);
|
|
return(ENOBUFS);
|
|
}
|
|
|
|
MCLGET(m_new, M_DONTWAIT);
|
|
if (!(m_new->m_flags & M_EXT)) {
|
|
printf("ax%d: no memory for rx list -- packet dropped!\n",
|
|
sc->ax_unit);
|
|
m_freem(m_new);
|
|
return(ENOBUFS);
|
|
}
|
|
|
|
c->ax_mbuf = m_new;
|
|
c->ax_ptr->ax_status = AX_RXSTAT;
|
|
c->ax_ptr->ax_data = vtophys(mtod(m_new, caddr_t));
|
|
c->ax_ptr->ax_ctl = MCLBYTES - 1;
|
|
|
|
return(0);
|
|
}
|
|
|
|
/*
|
|
* A frame has been uploaded: pass the resulting mbuf chain up to
|
|
* the higher level protocols.
|
|
*/
|
|
static void ax_rxeof(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
struct ether_header *eh;
|
|
struct mbuf *m;
|
|
struct ifnet *ifp;
|
|
struct ax_chain_onefrag *cur_rx;
|
|
int total_len = 0;
|
|
u_int32_t rxstat;
|
|
|
|
ifp = &sc->arpcom.ac_if;
|
|
|
|
while(!((rxstat = sc->ax_cdata.ax_rx_head->ax_ptr->ax_status) &
|
|
AX_RXSTAT_OWN)) {
|
|
cur_rx = sc->ax_cdata.ax_rx_head;
|
|
sc->ax_cdata.ax_rx_head = cur_rx->ax_nextdesc;
|
|
|
|
/*
|
|
* If an error occurs, update stats, clear the
|
|
* status word and leave the mbuf cluster in place:
|
|
* it should simply get re-used next time this descriptor
|
|
* comes up in the ring.
|
|
*/
|
|
if (rxstat & AX_RXSTAT_RXERR) {
|
|
ifp->if_ierrors++;
|
|
if (rxstat & AX_RXSTAT_COLLSEEN)
|
|
ifp->if_collisions++;
|
|
cur_rx->ax_ptr->ax_status = AX_RXSTAT;
|
|
cur_rx->ax_ptr->ax_ctl = (MCLBYTES - 1);
|
|
continue;
|
|
}
|
|
|
|
/* No errors; receive the packet. */
|
|
m = cur_rx->ax_mbuf;
|
|
total_len = AX_RXBYTES(cur_rx->ax_ptr->ax_status);
|
|
|
|
total_len -= ETHER_CRC_LEN;
|
|
|
|
if (total_len < MINCLSIZE) {
|
|
m = m_devget(mtod(cur_rx->ax_mbuf, char *),
|
|
total_len, 0, ifp, NULL);
|
|
cur_rx->ax_ptr->ax_status = AX_RXSTAT;
|
|
cur_rx->ax_ptr->ax_ctl = (MCLBYTES - 1);
|
|
if (m == NULL) {
|
|
ifp->if_ierrors++;
|
|
continue;
|
|
}
|
|
} else {
|
|
m = cur_rx->ax_mbuf;
|
|
/*
|
|
* Try to conjure up a new mbuf cluster. If that
|
|
* fails, it means we have an out of memory condition and
|
|
* should leave the buffer in place and continue. This will
|
|
* result in a lost packet, but there's little else we
|
|
* can do in this situation.
|
|
*/
|
|
if (ax_newbuf(sc, cur_rx) == ENOBUFS) {
|
|
ifp->if_ierrors++;
|
|
cur_rx->ax_ptr->ax_status = AX_RXSTAT;
|
|
cur_rx->ax_ptr->ax_ctl = (MCLBYTES - 1);
|
|
continue;
|
|
}
|
|
m->m_pkthdr.rcvif = ifp;
|
|
m->m_pkthdr.len = m->m_len = total_len;
|
|
}
|
|
|
|
ifp->if_ipackets++;
|
|
eh = mtod(m, struct ether_header *);
|
|
#if NBPFILTER > 0
|
|
/*
|
|
* Handle BPF listeners. Let the BPF user see the packet, but
|
|
* don't pass it up to the ether_input() layer unless it's
|
|
* a broadcast packet, multicast packet, matches our ethernet
|
|
* address or the interface is in promiscuous mode.
|
|
*/
|
|
if (ifp->if_bpf) {
|
|
bpf_mtap(ifp, m);
|
|
if (ifp->if_flags & IFF_PROMISC &&
|
|
(bcmp(eh->ether_dhost, sc->arpcom.ac_enaddr,
|
|
ETHER_ADDR_LEN) &&
|
|
(eh->ether_dhost[0] & 1) == 0)) {
|
|
m_freem(m);
|
|
continue;
|
|
}
|
|
}
|
|
#endif
|
|
/* Remove header from mbuf and pass it on. */
|
|
m_adj(m, sizeof(struct ether_header));
|
|
ether_input(ifp, eh, m);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
void ax_rxeoc(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
|
|
ax_rxeof(sc);
|
|
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_RX_ON);
|
|
CSR_WRITE_4(sc, AX_RXADDR, vtophys(sc->ax_cdata.ax_rx_head->ax_ptr));
|
|
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_RX_ON);
|
|
CSR_WRITE_4(sc, AX_RXSTART, 0xFFFFFFFF);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* A frame was downloaded to the chip. It's safe for us to clean up
|
|
* the list buffers.
|
|
*/
|
|
|
|
static void ax_txeof(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
struct ax_chain *cur_tx;
|
|
struct ifnet *ifp;
|
|
|
|
ifp = &sc->arpcom.ac_if;
|
|
|
|
/* Clear the timeout timer. */
|
|
ifp->if_timer = 0;
|
|
|
|
if (sc->ax_cdata.ax_tx_head == NULL)
|
|
return;
|
|
|
|
/*
|
|
* Go through our tx list and free mbufs for those
|
|
* frames that have been transmitted.
|
|
*/
|
|
while(sc->ax_cdata.ax_tx_head->ax_mbuf != NULL) {
|
|
u_int32_t txstat;
|
|
|
|
cur_tx = sc->ax_cdata.ax_tx_head;
|
|
txstat = AX_TXSTATUS(cur_tx);
|
|
|
|
if (txstat & AX_TXSTAT_OWN)
|
|
break;
|
|
|
|
if (txstat & AX_TXSTAT_ERRSUM) {
|
|
ifp->if_oerrors++;
|
|
if (txstat & AX_TXSTAT_EXCESSCOLL)
|
|
ifp->if_collisions++;
|
|
if (txstat & AX_TXSTAT_LATECOLL)
|
|
ifp->if_collisions++;
|
|
}
|
|
|
|
ifp->if_collisions += (txstat & AX_TXSTAT_COLLCNT) >> 3;
|
|
|
|
ifp->if_opackets++;
|
|
m_freem(cur_tx->ax_mbuf);
|
|
cur_tx->ax_mbuf = NULL;
|
|
|
|
if (sc->ax_cdata.ax_tx_head == sc->ax_cdata.ax_tx_tail) {
|
|
sc->ax_cdata.ax_tx_head = NULL;
|
|
sc->ax_cdata.ax_tx_tail = NULL;
|
|
break;
|
|
}
|
|
|
|
sc->ax_cdata.ax_tx_head = cur_tx->ax_nextdesc;
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* TX 'end of channel' interrupt handler.
|
|
*/
|
|
static void ax_txeoc(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
struct ifnet *ifp;
|
|
|
|
ifp = &sc->arpcom.ac_if;
|
|
|
|
ifp->if_timer = 0;
|
|
|
|
if (sc->ax_cdata.ax_tx_head == NULL) {
|
|
ifp->if_flags &= ~IFF_OACTIVE;
|
|
sc->ax_cdata.ax_tx_tail = NULL;
|
|
if (sc->ax_want_auto)
|
|
ax_autoneg_mii(sc, AX_FLAG_DELAYTIMEO, 1);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
static void ax_intr(arg)
|
|
void *arg;
|
|
{
|
|
struct ax_softc *sc;
|
|
struct ifnet *ifp;
|
|
u_int32_t status;
|
|
|
|
sc = arg;
|
|
ifp = &sc->arpcom.ac_if;
|
|
|
|
/* Supress unwanted interrupts */
|
|
if (!(ifp->if_flags & IFF_UP)) {
|
|
ax_stop(sc);
|
|
return;
|
|
}
|
|
|
|
/* Disable interrupts. */
|
|
CSR_WRITE_4(sc, AX_IMR, 0x00000000);
|
|
|
|
for (;;) {
|
|
status = CSR_READ_4(sc, AX_ISR);
|
|
if (status)
|
|
CSR_WRITE_4(sc, AX_ISR, status);
|
|
|
|
if ((status & AX_INTRS) == 0)
|
|
break;
|
|
|
|
if ((status & AX_ISR_TX_OK) || (status & AX_ISR_TX_EARLY))
|
|
ax_txeof(sc);
|
|
|
|
if (status & AX_ISR_TX_NOBUF)
|
|
ax_txeoc(sc);
|
|
|
|
if (status & AX_ISR_TX_IDLE) {
|
|
ax_txeof(sc);
|
|
if (sc->ax_cdata.ax_tx_head != NULL) {
|
|
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_TX_ON);
|
|
CSR_WRITE_4(sc, AX_TXSTART, 0xFFFFFFFF);
|
|
}
|
|
}
|
|
|
|
if (status & AX_ISR_TX_UNDERRUN) {
|
|
u_int32_t cfg;
|
|
cfg = CSR_READ_4(sc, AX_NETCFG);
|
|
if ((cfg & AX_NETCFG_TX_THRESH) == AX_TXTHRESH_160BYTES)
|
|
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_STORENFWD);
|
|
else
|
|
CSR_WRITE_4(sc, AX_NETCFG, cfg + 0x4000);
|
|
}
|
|
|
|
if (status & AX_ISR_RX_OK)
|
|
ax_rxeof(sc);
|
|
|
|
if ((status & AX_ISR_RX_WATDOGTIMEO)
|
|
|| (status & AX_ISR_RX_NOBUF))
|
|
ax_rxeoc(sc);
|
|
|
|
if (status & AX_ISR_BUS_ERR) {
|
|
ax_reset(sc);
|
|
ax_init(sc);
|
|
}
|
|
}
|
|
|
|
/* Re-enable interrupts. */
|
|
CSR_WRITE_4(sc, AX_IMR, AX_INTRS);
|
|
|
|
if (ifp->if_snd.ifq_head != NULL) {
|
|
ax_start(ifp);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Encapsulate an mbuf chain in a descriptor by coupling the mbuf data
|
|
* pointers to the fragment pointers.
|
|
*/
|
|
static int ax_encap(sc, c, m_head)
|
|
struct ax_softc *sc;
|
|
struct ax_chain *c;
|
|
struct mbuf *m_head;
|
|
{
|
|
int frag = 0;
|
|
volatile struct ax_desc *f = NULL;
|
|
int total_len;
|
|
struct mbuf *m;
|
|
|
|
/*
|
|
* Start packing the mbufs in this chain into
|
|
* the fragment pointers. Stop when we run out
|
|
* of fragments or hit the end of the mbuf chain.
|
|
*/
|
|
m = m_head;
|
|
total_len = 0;
|
|
|
|
for (m = m_head, frag = 0; m != NULL; m = m->m_next) {
|
|
if (m->m_len != 0) {
|
|
if (frag == AX_MAXFRAGS)
|
|
break;
|
|
total_len += m->m_len;
|
|
f = &c->ax_ptr->ax_frag[frag];
|
|
f->ax_ctl = m->m_len;
|
|
if (frag == 0) {
|
|
f->ax_status = 0;
|
|
f->ax_ctl |= AX_TXCTL_FIRSTFRAG;
|
|
} else
|
|
f->ax_status = AX_TXSTAT_OWN;
|
|
f->ax_next = vtophys(&c->ax_ptr->ax_frag[frag + 1]);
|
|
f->ax_data = vtophys(mtod(m, vm_offset_t));
|
|
frag++;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Handle special case: we ran out of fragments,
|
|
* but we have more mbufs left in the chain. Copy the
|
|
* data into an mbuf cluster. Note that we don't
|
|
* bother clearing the values in the other fragment
|
|
* pointers/counters; it wouldn't gain us anything,
|
|
* and would waste cycles.
|
|
*/
|
|
if (m != NULL) {
|
|
struct mbuf *m_new = NULL;
|
|
|
|
MGETHDR(m_new, M_DONTWAIT, MT_DATA);
|
|
if (m_new == NULL) {
|
|
printf("ax%d: no memory for tx list", sc->ax_unit);
|
|
return(1);
|
|
}
|
|
if (m_head->m_pkthdr.len > MHLEN) {
|
|
MCLGET(m_new, M_DONTWAIT);
|
|
if (!(m_new->m_flags & M_EXT)) {
|
|
m_freem(m_new);
|
|
printf("ax%d: no memory for tx list",
|
|
sc->ax_unit);
|
|
return(1);
|
|
}
|
|
}
|
|
m_copydata(m_head, 0, m_head->m_pkthdr.len,
|
|
mtod(m_new, caddr_t));
|
|
m_new->m_pkthdr.len = m_new->m_len = m_head->m_pkthdr.len;
|
|
m_freem(m_head);
|
|
m_head = m_new;
|
|
f = &c->ax_ptr->ax_frag[0];
|
|
f->ax_status = 0;
|
|
f->ax_data = vtophys(mtod(m_new, caddr_t));
|
|
f->ax_ctl = total_len = m_new->m_len;
|
|
f->ax_ctl |= AX_TXCTL_FIRSTFRAG;
|
|
frag = 1;
|
|
}
|
|
|
|
c->ax_mbuf = m_head;
|
|
c->ax_lastdesc = frag - 1;
|
|
AX_TXCTL(c) |= AX_TXCTL_LASTFRAG|AX_TXCTL_FINT;
|
|
c->ax_ptr->ax_frag[0].ax_ctl |= AX_TXCTL_FINT;
|
|
AX_TXNEXT(c) = vtophys(&c->ax_nextdesc->ax_ptr->ax_frag[0]);
|
|
return(0);
|
|
}
|
|
|
|
/*
|
|
* Main transmit routine. To avoid having to do mbuf copies, we put pointers
|
|
* to the mbuf data regions directly in the transmit lists. We also save a
|
|
* copy of the pointers since the transmit list fragment pointers are
|
|
* physical addresses.
|
|
*/
|
|
|
|
static void ax_start(ifp)
|
|
struct ifnet *ifp;
|
|
{
|
|
struct ax_softc *sc;
|
|
struct mbuf *m_head = NULL;
|
|
struct ax_chain *cur_tx = NULL, *start_tx;
|
|
|
|
sc = ifp->if_softc;
|
|
|
|
if (sc->ax_autoneg) {
|
|
sc->ax_tx_pend = 1;
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Check for an available queue slot. If there are none,
|
|
* punt.
|
|
*/
|
|
if (sc->ax_cdata.ax_tx_free->ax_mbuf != NULL) {
|
|
ifp->if_flags |= IFF_OACTIVE;
|
|
return;
|
|
}
|
|
|
|
start_tx = sc->ax_cdata.ax_tx_free;
|
|
|
|
while(sc->ax_cdata.ax_tx_free->ax_mbuf == NULL) {
|
|
IF_DEQUEUE(&ifp->if_snd, m_head);
|
|
if (m_head == NULL)
|
|
break;
|
|
|
|
/* Pick a descriptor off the free list. */
|
|
cur_tx = sc->ax_cdata.ax_tx_free;
|
|
sc->ax_cdata.ax_tx_free = cur_tx->ax_nextdesc;
|
|
|
|
/* Pack the data into the descriptor. */
|
|
ax_encap(sc, cur_tx, m_head);
|
|
if (cur_tx != start_tx)
|
|
AX_TXOWN(cur_tx) = AX_TXSTAT_OWN;
|
|
|
|
#if NBPFILTER > 0
|
|
/*
|
|
* If there's a BPF listener, bounce a copy of this frame
|
|
* to him.
|
|
*/
|
|
if (ifp->if_bpf)
|
|
bpf_mtap(ifp, cur_tx->ax_mbuf);
|
|
#endif
|
|
AX_TXOWN(cur_tx) = AX_TXSTAT_OWN;
|
|
CSR_WRITE_4(sc, AX_TXSTART, 0xFFFFFFFF);
|
|
}
|
|
|
|
sc->ax_cdata.ax_tx_tail = cur_tx;
|
|
if (sc->ax_cdata.ax_tx_head == NULL)
|
|
sc->ax_cdata.ax_tx_head = start_tx;
|
|
|
|
/*
|
|
* Set a timeout in case the chip goes out to lunch.
|
|
*/
|
|
ifp->if_timer = 5;
|
|
|
|
return;
|
|
}
|
|
|
|
static void ax_init(xsc)
|
|
void *xsc;
|
|
{
|
|
struct ax_softc *sc = xsc;
|
|
struct ifnet *ifp = &sc->arpcom.ac_if;
|
|
u_int16_t phy_bmcr = 0;
|
|
int s;
|
|
|
|
if (sc->ax_autoneg)
|
|
return;
|
|
|
|
s = splimp();
|
|
|
|
if (sc->ax_pinfo != NULL)
|
|
phy_bmcr = ax_phy_readreg(sc, PHY_BMCR);
|
|
|
|
/*
|
|
* Cancel pending I/O and free all RX/TX buffers.
|
|
*/
|
|
ax_stop(sc);
|
|
ax_reset(sc);
|
|
|
|
/*
|
|
* Set cache alignment and burst length.
|
|
*/
|
|
CSR_WRITE_4(sc, AX_BUSCTL, AX_BUSCTL_CONFIG);
|
|
|
|
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_HEARTBEAT);
|
|
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_STORENFWD);
|
|
|
|
if (sc->ax_pinfo != NULL) {
|
|
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_PORTSEL);
|
|
ax_setcfg(sc, ax_phy_readreg(sc, PHY_BMCR));
|
|
} else
|
|
ax_setmode(sc, sc->ifmedia.ifm_media, 0);
|
|
|
|
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_TX_THRESH);
|
|
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_SPEEDSEL);
|
|
|
|
if (IFM_SUBTYPE(sc->ifmedia.ifm_media) == IFM_10_T)
|
|
AX_SETBIT(sc, AX_NETCFG, AX_TXTHRESH_160BYTES);
|
|
else
|
|
AX_SETBIT(sc, AX_NETCFG, AX_TXTHRESH_72BYTES);
|
|
|
|
/* Init our MAC address */
|
|
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_PAR0);
|
|
CSR_WRITE_4(sc, AX_FILTDATA, *(u_int32_t *)(&sc->arpcom.ac_enaddr[0]));
|
|
CSR_WRITE_4(sc, AX_FILTIDX, AX_FILTIDX_PAR1);
|
|
CSR_WRITE_4(sc, AX_FILTDATA, *(u_int32_t *)(&sc->arpcom.ac_enaddr[4]));
|
|
|
|
/* Init circular RX list. */
|
|
if (ax_list_rx_init(sc) == ENOBUFS) {
|
|
printf("ax%d: initialization failed: no "
|
|
"memory for rx buffers\n", sc->ax_unit);
|
|
ax_stop(sc);
|
|
(void)splx(s);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Init tx descriptors.
|
|
*/
|
|
ax_list_tx_init(sc);
|
|
|
|
/* If we want promiscuous mode, set the allframes bit. */
|
|
if (ifp->if_flags & IFF_PROMISC) {
|
|
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_RX_PROMISC);
|
|
} else {
|
|
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_RX_PROMISC);
|
|
}
|
|
|
|
/*
|
|
* Set the capture broadcast bit to capture broadcast frames.
|
|
*/
|
|
if (ifp->if_flags & IFF_BROADCAST) {
|
|
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_RX_BROAD);
|
|
} else {
|
|
AX_CLRBIT(sc, AX_NETCFG, AX_NETCFG_RX_BROAD);
|
|
}
|
|
|
|
/*
|
|
* Load the multicast filter.
|
|
*/
|
|
ax_setmulti(sc);
|
|
|
|
/*
|
|
* Load the address of the RX list.
|
|
*/
|
|
CSR_WRITE_4(sc, AX_RXADDR, vtophys(sc->ax_cdata.ax_rx_head->ax_ptr));
|
|
CSR_WRITE_4(sc, AX_TXADDR, vtophys(&sc->ax_ldata->ax_tx_list[0]));
|
|
|
|
/*
|
|
* Enable interrupts.
|
|
*/
|
|
CSR_WRITE_4(sc, AX_IMR, AX_INTRS);
|
|
CSR_WRITE_4(sc, AX_ISR, 0xFFFFFFFF);
|
|
|
|
/* Enable receiver and transmitter. */
|
|
AX_SETBIT(sc, AX_NETCFG, AX_NETCFG_TX_ON|AX_NETCFG_RX_ON);
|
|
CSR_WRITE_4(sc, AX_RXSTART, 0xFFFFFFFF);
|
|
|
|
/* Restore state of BMCR */
|
|
if (sc->ax_pinfo != NULL)
|
|
ax_phy_writereg(sc, PHY_BMCR, phy_bmcr);
|
|
|
|
ifp->if_flags |= IFF_RUNNING;
|
|
ifp->if_flags &= ~IFF_OACTIVE;
|
|
|
|
(void)splx(s);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Set media options.
|
|
*/
|
|
static int ax_ifmedia_upd(ifp)
|
|
struct ifnet *ifp;
|
|
{
|
|
struct ax_softc *sc;
|
|
struct ifmedia *ifm;
|
|
|
|
sc = ifp->if_softc;
|
|
ifm = &sc->ifmedia;
|
|
|
|
if (IFM_TYPE(ifm->ifm_media) != IFM_ETHER)
|
|
return(EINVAL);
|
|
|
|
if (IFM_SUBTYPE(ifm->ifm_media) == IFM_AUTO)
|
|
ax_autoneg_mii(sc, AX_FLAG_SCHEDDELAY, 1);
|
|
else {
|
|
if (sc->ax_pinfo == NULL)
|
|
ax_setmode(sc, ifm->ifm_media, 1);
|
|
else
|
|
ax_setmode_mii(sc, ifm->ifm_media);
|
|
}
|
|
|
|
return(0);
|
|
}
|
|
|
|
/*
|
|
* Report current media status.
|
|
*/
|
|
static void ax_ifmedia_sts(ifp, ifmr)
|
|
struct ifnet *ifp;
|
|
struct ifmediareq *ifmr;
|
|
{
|
|
struct ax_softc *sc;
|
|
u_int16_t advert = 0, ability = 0;
|
|
u_int32_t media = 0;
|
|
|
|
sc = ifp->if_softc;
|
|
|
|
ifmr->ifm_active = IFM_ETHER;
|
|
|
|
if (sc->ax_pinfo == NULL) {
|
|
media = CSR_READ_4(sc, AX_NETCFG);
|
|
if (media & AX_NETCFG_PORTSEL)
|
|
ifmr->ifm_active = IFM_ETHER|IFM_100_TX;
|
|
else
|
|
ifmr->ifm_active = IFM_ETHER|IFM_10_T;
|
|
if (media & AX_NETCFG_FULLDUPLEX)
|
|
ifmr->ifm_active |= IFM_FDX;
|
|
else
|
|
ifmr->ifm_active |= IFM_HDX;
|
|
return;
|
|
}
|
|
|
|
if (!(ax_phy_readreg(sc, PHY_BMCR) & PHY_BMCR_AUTONEGENBL)) {
|
|
if (ax_phy_readreg(sc, PHY_BMCR) & PHY_BMCR_SPEEDSEL)
|
|
ifmr->ifm_active = IFM_ETHER|IFM_100_TX;
|
|
else
|
|
ifmr->ifm_active = IFM_ETHER|IFM_10_T;
|
|
if (ax_phy_readreg(sc, PHY_BMCR) & PHY_BMCR_DUPLEX)
|
|
ifmr->ifm_active |= IFM_FDX;
|
|
else
|
|
ifmr->ifm_active |= IFM_HDX;
|
|
return;
|
|
}
|
|
|
|
ability = ax_phy_readreg(sc, PHY_LPAR);
|
|
advert = ax_phy_readreg(sc, PHY_ANAR);
|
|
if (advert & PHY_ANAR_100BT4 &&
|
|
ability & PHY_ANAR_100BT4) {
|
|
ifmr->ifm_active = IFM_ETHER|IFM_100_T4;
|
|
} else if (advert & PHY_ANAR_100BTXFULL &&
|
|
ability & PHY_ANAR_100BTXFULL) {
|
|
ifmr->ifm_active = IFM_ETHER|IFM_100_TX|IFM_FDX;
|
|
} else if (advert & PHY_ANAR_100BTXHALF &&
|
|
ability & PHY_ANAR_100BTXHALF) {
|
|
ifmr->ifm_active = IFM_ETHER|IFM_100_TX|IFM_HDX;
|
|
} else if (advert & PHY_ANAR_10BTFULL &&
|
|
ability & PHY_ANAR_10BTFULL) {
|
|
ifmr->ifm_active = IFM_ETHER|IFM_10_T|IFM_FDX;
|
|
} else if (advert & PHY_ANAR_10BTHALF &&
|
|
ability & PHY_ANAR_10BTHALF) {
|
|
ifmr->ifm_active = IFM_ETHER|IFM_10_T|IFM_HDX;
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
static int ax_ioctl(ifp, command, data)
|
|
struct ifnet *ifp;
|
|
u_long command;
|
|
caddr_t data;
|
|
{
|
|
struct ax_softc *sc = ifp->if_softc;
|
|
struct ifreq *ifr = (struct ifreq *) data;
|
|
int s, error = 0;
|
|
|
|
s = splimp();
|
|
|
|
switch(command) {
|
|
case SIOCSIFADDR:
|
|
case SIOCGIFADDR:
|
|
case SIOCSIFMTU:
|
|
error = ether_ioctl(ifp, command, data);
|
|
break;
|
|
case SIOCSIFFLAGS:
|
|
if (ifp->if_flags & IFF_UP) {
|
|
ax_init(sc);
|
|
} else {
|
|
if (ifp->if_flags & IFF_RUNNING)
|
|
ax_stop(sc);
|
|
}
|
|
error = 0;
|
|
break;
|
|
case SIOCADDMULTI:
|
|
case SIOCDELMULTI:
|
|
ax_setmulti(sc);
|
|
error = 0;
|
|
break;
|
|
case SIOCGIFMEDIA:
|
|
case SIOCSIFMEDIA:
|
|
error = ifmedia_ioctl(ifp, ifr, &sc->ifmedia, command);
|
|
break;
|
|
default:
|
|
error = EINVAL;
|
|
break;
|
|
}
|
|
|
|
(void)splx(s);
|
|
|
|
return(error);
|
|
}
|
|
|
|
static void ax_watchdog(ifp)
|
|
struct ifnet *ifp;
|
|
{
|
|
struct ax_softc *sc;
|
|
|
|
sc = ifp->if_softc;
|
|
|
|
if (sc->ax_autoneg) {
|
|
ax_autoneg_mii(sc, AX_FLAG_DELAYTIMEO, 1);
|
|
return;
|
|
}
|
|
|
|
ifp->if_oerrors++;
|
|
printf("ax%d: watchdog timeout\n", sc->ax_unit);
|
|
|
|
if (sc->ax_pinfo != NULL) {
|
|
if (!(ax_phy_readreg(sc, PHY_BMSR) & PHY_BMSR_LINKSTAT))
|
|
printf("ax%d: no carrier - transceiver "
|
|
"cable problem?\n", sc->ax_unit);
|
|
}
|
|
|
|
ax_stop(sc);
|
|
ax_reset(sc);
|
|
ax_init(sc);
|
|
|
|
if (ifp->if_snd.ifq_head != NULL)
|
|
ax_start(ifp);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Stop the adapter and free any mbufs allocated to the
|
|
* RX and TX lists.
|
|
*/
|
|
static void ax_stop(sc)
|
|
struct ax_softc *sc;
|
|
{
|
|
register int i;
|
|
struct ifnet *ifp;
|
|
|
|
ifp = &sc->arpcom.ac_if;
|
|
ifp->if_timer = 0;
|
|
|
|
AX_CLRBIT(sc, AX_NETCFG, (AX_NETCFG_RX_ON|AX_NETCFG_TX_ON));
|
|
CSR_WRITE_4(sc, AX_IMR, 0x00000000);
|
|
CSR_WRITE_4(sc, AX_TXADDR, 0x00000000);
|
|
CSR_WRITE_4(sc, AX_RXADDR, 0x00000000);
|
|
|
|
/*
|
|
* Free data in the RX lists.
|
|
*/
|
|
for (i = 0; i < AX_RX_LIST_CNT; i++) {
|
|
if (sc->ax_cdata.ax_rx_chain[i].ax_mbuf != NULL) {
|
|
m_freem(sc->ax_cdata.ax_rx_chain[i].ax_mbuf);
|
|
sc->ax_cdata.ax_rx_chain[i].ax_mbuf = NULL;
|
|
}
|
|
}
|
|
bzero((volatile char *)&sc->ax_ldata->ax_rx_list,
|
|
sizeof(sc->ax_ldata->ax_rx_list));
|
|
|
|
/*
|
|
* Free the TX list buffers.
|
|
*/
|
|
for (i = 0; i < AX_TX_LIST_CNT; i++) {
|
|
if (sc->ax_cdata.ax_tx_chain[i].ax_mbuf != NULL) {
|
|
m_freem(sc->ax_cdata.ax_tx_chain[i].ax_mbuf);
|
|
sc->ax_cdata.ax_tx_chain[i].ax_mbuf = NULL;
|
|
}
|
|
}
|
|
|
|
bzero((volatile char *)&sc->ax_ldata->ax_tx_list,
|
|
sizeof(sc->ax_ldata->ax_tx_list));
|
|
|
|
ifp->if_flags &= ~(IFF_RUNNING | IFF_OACTIVE);
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Stop all chip I/O so that the kernel's probe routines don't
|
|
* get confused by errant DMAs when rebooting.
|
|
*/
|
|
static void ax_shutdown(howto, arg)
|
|
int howto;
|
|
void *arg;
|
|
{
|
|
struct ax_softc *sc = (struct ax_softc *)arg;
|
|
|
|
ax_stop(sc);
|
|
|
|
return;
|
|
}
|
|
|
|
static struct pci_device ax_device = {
|
|
"ax",
|
|
ax_probe,
|
|
ax_attach,
|
|
&ax_count,
|
|
NULL
|
|
};
|
|
DATA_SET(pcidevice_set, ax_device);
|