1544 lines
36 KiB
C
1544 lines
36 KiB
C
/*-
|
|
* Copyright (c) 1997 Semen Ustimenko (semen@iclub.nsu.ru)
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $Id: if_tx.c,v 1.11 1998/06/07 17:12:38 dfr Exp $
|
|
*
|
|
*/
|
|
|
|
/*
|
|
* EtherPower II 10/100 Fast Ethernet (tx0)
|
|
* (aka SMC9432TX based on SMC83c170 EPIC chip)
|
|
*
|
|
* TODO:
|
|
* Deal with TX threshold (probably we should calculate it depending
|
|
* on processor speed, as did the MS-DOS driver).
|
|
* Deal with bus mastering, i.e. i realy don't know what to do with
|
|
* it and how it can improve performance.
|
|
* Implement FULL IFF_MULTICAST support.
|
|
* Calculate optimal RX and TX rings size.
|
|
* Test, test and test again:-)
|
|
*
|
|
*/
|
|
|
|
/* We should define compile time options before smc83c170.h included */
|
|
/*#define EPIC_NOIFMEDIA 1*/
|
|
/*#define EPIC_USEIOSPACE 1*/
|
|
/*#define EPIC_DEBUG 1*/
|
|
#define RX_TO_MBUF 1 /* Receive directly to mbuf enstead of */
|
|
/* static allocated buffer */
|
|
#define TX_FRAG_LIST 1 /* Transmit directly from mbuf enstead */
|
|
/* of collecting mbuf's frags to one */
|
|
/* static allocated place */
|
|
|
|
#include "pci.h"
|
|
#if NPCI > 0
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/mbuf.h>
|
|
#include <sys/socket.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/kernel.h>
|
|
#include <sys/sockio.h>
|
|
#include <net/if.h>
|
|
#if defined(SIOCSIFMEDIA) && !defined(EPIC_NOIFMEDIA)
|
|
#include <net/if_media.h>
|
|
#endif
|
|
#include <net/if_mib.h>
|
|
#include <netinet/in.h>
|
|
#include <netinet/if_ether.h>
|
|
#include <vm/vm.h>
|
|
#include <vm/pmap.h>
|
|
#include <machine/clock.h>
|
|
|
|
#include <pci/pcivar.h>
|
|
#include <pci/smc83c170.h>
|
|
|
|
#include "bpfilter.h"
|
|
#if NBPFILTER > 0
|
|
#include <net/bpf.h>
|
|
#endif
|
|
|
|
/*
|
|
* Global variables
|
|
*/
|
|
static u_long epic_pci_count;
|
|
static epic_softc_t * epics[EPIC_MAX_DEVICES];
|
|
static struct pci_device txdevice = {
|
|
"tx",
|
|
epic_pci_probe,
|
|
epic_pci_attach,
|
|
&epic_pci_count,
|
|
NULL };
|
|
|
|
/*
|
|
* Append this driver to pci drivers list
|
|
*/
|
|
DATA_SET ( pcidevice_set, txdevice );
|
|
|
|
/*
|
|
* ifioctl function
|
|
*
|
|
* splimp() invoked here
|
|
*/
|
|
static int
|
|
epic_ifioctl __P((
|
|
register struct ifnet * ifp,
|
|
u_long command, caddr_t data))
|
|
{
|
|
epic_softc_t *sc = ifp->if_softc;
|
|
struct ifreq *ifr = (struct ifreq *) data;
|
|
int x, error = 0;
|
|
|
|
x = splimp();
|
|
|
|
switch (command) {
|
|
|
|
case SIOCSIFADDR:
|
|
case SIOCGIFADDR:
|
|
ether_ioctl(ifp, command, data);
|
|
break;
|
|
|
|
case SIOCSIFFLAGS:
|
|
/*
|
|
* If the interface is marked up and stopped, then start it.
|
|
* If it is marked down and running, then stop it.
|
|
*/
|
|
if (ifp->if_flags & IFF_UP) {
|
|
if ((ifp->if_flags & IFF_RUNNING) == 0) {
|
|
epic_init(sc);
|
|
break;
|
|
}
|
|
} else {
|
|
if (ifp->if_flags & IFF_RUNNING) {
|
|
epic_stop(sc);
|
|
ifp->if_flags &= ~IFF_RUNNING;
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* Handle IFF_PROMISC flag */
|
|
epic_set_rx_mode(sc);
|
|
|
|
#if !defined(_NET_IF_MEDIA_H_)
|
|
/* Handle IFF_LINKx flags */
|
|
epic_set_media_speed(sc);
|
|
#endif
|
|
|
|
break;
|
|
|
|
case SIOCADDMULTI:
|
|
case SIOCDELMULTI:
|
|
|
|
/* Update out multicast list */
|
|
#if defined(__FreeBSD__) && __FreeBSD__ >= 3
|
|
epic_set_mc_table(sc);
|
|
error = 0;
|
|
#else
|
|
error = (command == SIOCADDMULTI) ?
|
|
ether_addmulti(ifr, &sc->epic_ac) :
|
|
ether_delmulti(ifr, &sc->epic_ac);
|
|
|
|
if (error == ENETRESET) {
|
|
epic_set_mc_table(sc);
|
|
error = 0;
|
|
}
|
|
#endif
|
|
break;
|
|
|
|
case SIOCSIFMTU:
|
|
/*
|
|
* Set the interface MTU.
|
|
*/
|
|
if (ifr->ifr_mtu > ETHERMTU) {
|
|
error = EINVAL;
|
|
} else {
|
|
ifp->if_mtu = ifr->ifr_mtu;
|
|
}
|
|
break;
|
|
|
|
#if defined(_NET_IF_MEDIA_H_)
|
|
case SIOCSIFMEDIA:
|
|
case SIOCGIFMEDIA:
|
|
error = ifmedia_ioctl(ifp, ifr, &sc->ifmedia, command);
|
|
break;
|
|
#endif
|
|
|
|
default:
|
|
error = EINVAL;
|
|
}
|
|
splx(x);
|
|
|
|
return error;
|
|
}
|
|
|
|
/*
|
|
* ifstart function
|
|
*
|
|
* splimp() assumed to be done
|
|
*/
|
|
static void
|
|
epic_ifstart(struct ifnet * const ifp){
|
|
epic_softc_t *sc = ifp->if_softc;
|
|
|
|
while( sc->pending_txs < TX_RING_SIZE ){
|
|
struct epic_tx_buffer *buf = sc->tx_buffer + sc->cur_tx;
|
|
struct epic_tx_desc *desc = sc->tx_desc + sc->cur_tx;
|
|
#if defined(TX_FRAG_LIST)
|
|
struct epic_frag_list *flist = sc->tx_flist + sc->cur_tx;
|
|
#endif
|
|
struct mbuf *m,*m0;
|
|
int len;
|
|
|
|
/* If descriptor is busy, set IFF_OACTIVE and exit */
|
|
if( desc->status & 0x8000 ) break;
|
|
|
|
/* Get next packet to send */
|
|
IF_DEQUEUE( &(sc->epic_if.if_snd), m );
|
|
|
|
/* If no more mbuf's to send, return */
|
|
if( NULL == m ) return;
|
|
|
|
/* Save mbuf header */
|
|
m0 = m;
|
|
|
|
#if defined(TX_FRAG_LIST)
|
|
if( buf->mbuf ) m_freem( buf->mbuf );
|
|
buf->mbuf = m;
|
|
flist->numfrags = 0;
|
|
|
|
for(len=0;(m0!=0)&&(flist->numfrags<63);m0=m0->m_next) {
|
|
flist->frag[flist->numfrags].fraglen = m0->m_len;
|
|
flist->frag[flist->numfrags].fragaddr =
|
|
vtophys( mtod(m0, caddr_t) );
|
|
len += m0->m_len;
|
|
flist->numfrags++;
|
|
}
|
|
|
|
if( NULL != m0 ){
|
|
/* Copy packet to new allocated mbuf */
|
|
MGETHDR(m0,M_DONTWAIT,MT_DATA);
|
|
if( NULL == m0 ) {
|
|
printf("tx%d: cannot allocate mbuf header\n",sc->unit);
|
|
sc->epic_if.if_oerrors++;
|
|
m_freem(m);
|
|
continue;
|
|
}
|
|
MCLGET(m0,M_DONTWAIT);
|
|
if( NULL == (m0->m_flags & M_EXT) ){
|
|
printf("tx%d: cannot allocate mbuf cluster\n",sc->unit);
|
|
m_freem(m0);
|
|
m_freem(m);
|
|
sc->epic_if.if_oerrors++;
|
|
continue;
|
|
}
|
|
|
|
for (len = 0; m != 0; m = m->m_next) {
|
|
bcopy( mtod(m, caddr_t), mtod(m0, caddr_t) + len, m->m_len);
|
|
len += m->m_len;
|
|
}
|
|
|
|
m_freem(buf->mbuf);
|
|
buf->mbuf = m0;
|
|
flist->numfrags = 1;
|
|
flist->frag[0].fraglen = len;
|
|
flist->frag[0].fragaddr = vtophys( mtod(m0, caddr_t) );
|
|
}
|
|
|
|
/* Does not generate TXC unless ring is full more then a half */
|
|
desc->control = (sc->pending_txs>TX_RING_SIZE/2)?0x05:0x01;
|
|
#else
|
|
for (len = 0; m0 != 0; m0 = m0->m_next) {
|
|
bcopy(mtod(m0, caddr_t), buf->data + len, m0->m_len);
|
|
len += m0->m_len;
|
|
}
|
|
|
|
/* Does not generate TXC unless ring is full more then a half */
|
|
desc->control = (sc->pending_txs>TX_RING_SIZE/2)?0x14:0x10;
|
|
#endif
|
|
|
|
/* Packet should be at least ETHER_MIN_LEN */
|
|
desc->txlength = max(len,ETHER_MIN_LEN-ETHER_CRC_LEN);
|
|
|
|
/* Pass ownership to the chip */
|
|
desc->status = 0x8000;
|
|
|
|
/* Set watchdog timer */
|
|
ifp->if_timer = 2;
|
|
|
|
#if NBPFILTER > 0
|
|
if( ifp->if_bpf ) bpf_mtap( ifp, m );
|
|
#endif
|
|
|
|
#if !defined(TX_FRAG_LIST)
|
|
/* We don't need mbuf anyway */
|
|
m_freem( m );
|
|
#endif
|
|
/* Trigger an immediate transmit demand. */
|
|
CSR_WRITE_4( sc, COMMAND, COMMAND_TXQUEUED );
|
|
|
|
/* Packet queued successful */
|
|
sc->pending_txs++;
|
|
|
|
/* Switch to next descriptor */
|
|
sc->cur_tx = ( sc->cur_tx + 1 ) % TX_RING_SIZE;
|
|
}
|
|
|
|
sc->epic_if.if_flags |= IFF_OACTIVE;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
/*
|
|
* IFWATCHDOG function
|
|
*
|
|
* splimp() invoked here
|
|
*/
|
|
static void
|
|
epic_ifwatchdog(
|
|
struct ifnet *ifp)
|
|
{
|
|
epic_softc_t *sc = ifp->if_softc;
|
|
int x;
|
|
int i;
|
|
|
|
x = splimp();
|
|
|
|
printf("tx%d: device timeout %d packets\n",
|
|
sc->unit,sc->pending_txs);
|
|
|
|
ifp->if_oerrors+=sc->pending_txs;
|
|
|
|
epic_stop(sc);
|
|
epic_init(sc);
|
|
|
|
epic_ifstart(&sc->epic_if);
|
|
|
|
splx(x);
|
|
}
|
|
|
|
/*
|
|
*
|
|
* splimp() invoked before epic_intr_normal()
|
|
*/
|
|
static __inline void
|
|
epic_rx_done __P((
|
|
epic_softc_t *sc ))
|
|
{
|
|
int i = 0;
|
|
u_int16_t len;
|
|
struct epic_rx_buffer *buf;
|
|
struct epic_rx_desc *desc;
|
|
struct mbuf *m;
|
|
#if defined(RX_TO_MBUF)
|
|
struct mbuf *m0;
|
|
#endif
|
|
struct ether_header *eh;
|
|
int stt;
|
|
|
|
|
|
while( !(sc->rx_desc[sc->cur_rx].status & 0x8000) && \
|
|
i++ < RX_RING_SIZE ){
|
|
|
|
buf = sc->rx_buffer + sc->cur_rx;
|
|
desc = sc->rx_desc + sc->cur_rx;
|
|
|
|
stt = desc->status;
|
|
|
|
/* Check for errors */
|
|
if( !(desc->status & 1) ) {
|
|
sc->epic_if.if_ierrors++;
|
|
goto rxerror;
|
|
}
|
|
|
|
/* This is received frame actual length */
|
|
len = desc->rxlength - ETHER_CRC_LEN;
|
|
|
|
#if defined(RX_TO_MBUF)
|
|
/* Try to allocate mbuf cluster */
|
|
MGETHDR(m0,M_DONTWAIT,MT_DATA);
|
|
if( NULL == m0 ) {
|
|
printf("tx%d: cannot allocate mbuf header\n",sc->unit);
|
|
sc->epic_if.if_ierrors++;
|
|
goto rxerror;
|
|
}
|
|
MCLGET(m0,M_DONTWAIT);
|
|
if( NULL == (m0->m_flags & M_EXT) ){
|
|
printf("tx%d: cannot allocate mbuf cluster\n",sc->unit);
|
|
m_freem(m0);
|
|
sc->epic_if.if_ierrors++;
|
|
goto rxerror;
|
|
}
|
|
|
|
/* Swap new allocated mbuf with mbuf, containing packet */
|
|
m = buf->mbuf;
|
|
buf->mbuf = m0;
|
|
|
|
/* Insert new allocated mbuf into device queue */
|
|
desc->bufaddr = vtophys( mtod( buf->mbuf, caddr_t ) );
|
|
#else
|
|
/* Allocate mbuf to pass to OS */
|
|
MGETHDR(m, M_DONTWAIT, MT_DATA);
|
|
if( NULL == m ){
|
|
printf("tx%d: cannot allocate mbuf header\n",sc->unit);
|
|
sc->epic_if.if_ierrors++;
|
|
goto rxerror;
|
|
}
|
|
if( len > MHLEN ){
|
|
MCLGET(m,M_DONTWAIT);
|
|
if( NULL == (m->m_flags & M_EXT) ){
|
|
printf("tx%d: cannot allocate mbuf cluster\n",
|
|
sc->unit);
|
|
m_freem( m );
|
|
sc->epic_if.if_ierrors++;
|
|
goto rxerror;
|
|
}
|
|
}
|
|
|
|
/* Copy packet to new allocated mbuf */
|
|
memcpy( mtod(m,void*), buf->data, len );
|
|
|
|
#endif
|
|
|
|
/* First mbuf in packet holds the ethernet and packet headers */
|
|
eh = mtod( m, struct ether_header * );
|
|
m->m_pkthdr.rcvif = &(sc->epic_if);
|
|
m->m_pkthdr.len = len;
|
|
m->m_len = len;
|
|
|
|
#if NBPFILTER > 0
|
|
/* Give mbuf to BPFILTER */
|
|
if( sc->epic_if.if_bpf ) bpf_mtap( &sc->epic_if, m );
|
|
|
|
/* Accept only our packets, broadcasts and multicasts */
|
|
if( (eh->ether_dhost[0] & 1) == 0 &&
|
|
bcmp(eh->ether_dhost,sc->epic_ac.ac_enaddr,ETHER_ADDR_LEN)){
|
|
m_freem(m);
|
|
goto rxerror;
|
|
}
|
|
#endif
|
|
|
|
/* Second mbuf holds packet ifself */
|
|
m->m_pkthdr.len = len - sizeof(struct ether_header);
|
|
m->m_len = len - sizeof( struct ether_header );
|
|
m->m_data += sizeof( struct ether_header );
|
|
|
|
/* Give mbuf to OS */
|
|
ether_input(&sc->epic_if, eh, m);
|
|
|
|
/* Successfuly received frame */
|
|
sc->epic_if.if_ipackets++;
|
|
|
|
rxerror:
|
|
/* Mark current descriptor as free */
|
|
desc->rxlength = 0;
|
|
desc->status = 0x8000;
|
|
|
|
/* Switch to next descriptor */
|
|
sc->cur_rx = (sc->cur_rx+1) % RX_RING_SIZE;
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
*
|
|
* splimp() invoked before epic_intr_normal()
|
|
*/
|
|
static __inline void
|
|
epic_tx_done __P((
|
|
epic_softc_t *sc ))
|
|
{
|
|
int i = 0;
|
|
u_int32_t if_flags=~0;
|
|
int coll;
|
|
u_int16_t stt;
|
|
|
|
while( i++ < TX_RING_SIZE ){
|
|
struct epic_tx_buffer *buf = sc->tx_buffer + sc->dirty_tx;
|
|
struct epic_tx_desc *desc = sc->tx_desc + sc->dirty_tx;
|
|
#if defined(TX_FRAG_LIST)
|
|
struct epic_frag_list *flist = sc->tx_flist + sc->dirty_tx;
|
|
#endif
|
|
u_int16_t len = desc->txlength;
|
|
stt = desc->status;
|
|
|
|
if( stt & 0x8000 )
|
|
break; /* following packets are not Txed yet */
|
|
|
|
if( stt == 0 ){
|
|
if_flags = ~IFF_OACTIVE;
|
|
break;
|
|
}
|
|
|
|
sc->pending_txs--; /* packet is finished */
|
|
sc->dirty_tx = (sc->dirty_tx + 1) % TX_RING_SIZE;
|
|
|
|
coll = (stt >> 8) & 0xF; /* number of collisions*/
|
|
|
|
if( stt & 0x0001 ){
|
|
sc->epic_if.if_opackets++;
|
|
} else {
|
|
if(stt & 0x0008)
|
|
sc->dot3stats.dot3StatsCarrierSenseErrors++;
|
|
|
|
if(stt & 0x1050)
|
|
sc->dot3stats.dot3StatsInternalMacTransmitErrors++;
|
|
|
|
if(stt & 0x1000) coll = 16;
|
|
|
|
sc->epic_if.if_oerrors++;
|
|
}
|
|
|
|
if(stt & 0x0002) /* What does it mean? */
|
|
sc->dot3stats.dot3StatsDeferredTransmissions++;
|
|
|
|
sc->epic_if.if_collisions += coll;
|
|
|
|
switch( coll ){
|
|
case 0:
|
|
break;
|
|
case 16:
|
|
sc->dot3stats.dot3StatsExcessiveCollisions++;
|
|
sc->dot3stats.dot3StatsCollFrequencies[15]++;
|
|
break;
|
|
case 1:
|
|
sc->dot3stats.dot3StatsSingleCollisionFrames++;
|
|
sc->dot3stats.dot3StatsCollFrequencies[0]++;
|
|
break;
|
|
default:
|
|
sc->dot3stats.dot3StatsMultipleCollisionFrames++;
|
|
sc->dot3stats.dot3StatsCollFrequencies[coll-1]++;
|
|
break;
|
|
}
|
|
|
|
desc->status = 0;
|
|
desc->txlength = 0;
|
|
|
|
#if defined(TX_FRAG_LIST)
|
|
flist->numfrags = 0;
|
|
m_freem( buf->mbuf );
|
|
buf->mbuf = NULL;
|
|
#endif
|
|
|
|
if_flags = ~IFF_OACTIVE;
|
|
}
|
|
|
|
sc->epic_if.if_flags &= if_flags;
|
|
|
|
if( !(sc->epic_if.if_flags & IFF_OACTIVE) )
|
|
epic_ifstart( &sc->epic_if );
|
|
|
|
}
|
|
|
|
/*
|
|
* Interrupt function
|
|
*
|
|
* splimp() assumed to be done
|
|
*/
|
|
static void
|
|
epic_intr_normal(
|
|
void *arg)
|
|
{
|
|
epic_softc_t * sc = (epic_softc_t *) arg;
|
|
int status;
|
|
|
|
status = CSR_READ_4( sc, INTSTAT);
|
|
CSR_WRITE_4( sc, INTSTAT, status & (
|
|
INTSTAT_RQE|INTSTAT_HCC|INTSTAT_RCC|
|
|
INTSTAT_TXC|INTSTAT_TCC|INTSTAT_TQE|
|
|
INTSTAT_FATAL|INTSTAT_GP2|
|
|
INTSTAT_CNT|INTSTAT_TXU|INTSTAT_OVW|INTSTAT_RXE ) );
|
|
|
|
if( status & (INTSTAT_RQE|INTSTAT_HCC|INTSTAT_RCC) ) {
|
|
epic_rx_done( sc );
|
|
if( status & INTSTAT_RQE )
|
|
CSR_WRITE_4( sc, COMMAND, COMMAND_RXQUEUED );
|
|
}
|
|
|
|
if( status & (INTSTAT_TXC|INTSTAT_TCC|INTSTAT_TQE) )
|
|
epic_tx_done( sc );
|
|
|
|
if( (status & INTSTAT_TQE) && !(sc->epic_if.if_flags & IFF_OACTIVE) )
|
|
epic_ifstart( &sc->epic_if );
|
|
|
|
if( (status & INTSTAT_GP2) && (QS6612_OUI == sc->phyid) ) {
|
|
u_int32_t status;
|
|
|
|
status = epic_read_phy_register( sc, QS6612_INTSTAT );
|
|
|
|
if( (status & INTSTAT_AN_COMPLETE) &&
|
|
(epic_autoneg(sc) == EPIC_FULL_DUPLEX) ) {
|
|
status = BMCR_FULL_DUPLEX | epic_read_phy_register( sc, DP83840_BMCR );
|
|
CSR_WRITE_4( sc, TXCON,
|
|
TXCON_LOOPBACK_MODE_FULL_DUPLEX|TXCON_DEFAULT );
|
|
} else {
|
|
status = ~BMCR_FULL_DUPLEX & epic_read_phy_register( sc, DP83840_BMCR );
|
|
CSR_WRITE_4( sc, TXCON, TXCON_DEFAULT );
|
|
}
|
|
|
|
/* There is apparently QS6612 chip bug: */
|
|
/* BMCR_FULL_DUPLEX flag is not updated by */
|
|
/* autonegotiation process, so update it by hands */
|
|
epic_write_phy_register( sc, DP83840_BMCR, status);
|
|
|
|
/* We should clear GP2 int again after we clear it on PHY */
|
|
CSR_WRITE_4( sc, INTSTAT, INTSTAT_GP2 );
|
|
}
|
|
|
|
if( status & (INTSTAT_FATAL|INTSTAT_PMA|INTSTAT_PTA|INTSTAT_APE|INTSTAT_DPE) ){
|
|
int j;
|
|
struct epic_tx_desc *desc;
|
|
|
|
printf("tx%d: PCI fatal error occured (%s%s%s%s)\n",
|
|
sc->unit,
|
|
(status&INTSTAT_PMA)?"PMA":"",
|
|
(status&INTSTAT_PTA)?" PTA":"",
|
|
(status&INTSTAT_APE)?" APE":"",
|
|
(status&INTSTAT_DPE)?" DPE":"");
|
|
|
|
#if defined(EPIC_DEBUG)
|
|
printf("tx%d: dumping descriptors\n",sc->unit);
|
|
for(j=0;j<TX_RING_SIZE;j++){
|
|
desc = sc->tx_desc + j;
|
|
printf("desc%d: %d %04x, %08x, %04x %d, %08x\n",
|
|
j,
|
|
desc->txlength,desc->status,
|
|
desc->bufaddr,
|
|
desc->control,desc->buflength,
|
|
desc->next
|
|
);
|
|
}
|
|
#endif
|
|
epic_stop(sc);
|
|
epic_init(sc);
|
|
|
|
return;
|
|
}
|
|
|
|
/* UPDATE statistics */
|
|
if (status & (INTSTAT_CNT | INTSTAT_TXU | INTSTAT_OVW | INTSTAT_RXE)) {
|
|
|
|
/* update dot3 Rx statistics */
|
|
sc->dot3stats.dot3StatsMissedFrames += CSR_READ_1( sc, MPCNT);
|
|
sc->dot3stats.dot3StatsFrameTooLongs += CSR_READ_1( sc, ALICNT);
|
|
sc->dot3stats.dot3StatsFCSErrors += CSR_READ_1( sc, CRCCNT);
|
|
|
|
/* Update if Rx statistics */
|
|
if (status & (INTSTAT_OVW | INTSTAT_RXE))
|
|
sc->epic_if.if_ierrors++;
|
|
|
|
/* Tx FIFO underflow. */
|
|
if (status & INTSTAT_TXU) {
|
|
/* Inc. counters */
|
|
sc->dot3stats.dot3StatsInternalMacTransmitErrors++;
|
|
sc->epic_if.if_oerrors++;
|
|
|
|
/* Restart the transmit process. */
|
|
CSR_WRITE_4(sc, COMMAND, COMMAND_TXUGO);
|
|
}
|
|
}
|
|
|
|
/* If no packets are pending, thus no timeouts */
|
|
if( sc->pending_txs == 0 )
|
|
sc->epic_if.if_timer = 0;
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Probe function
|
|
*/
|
|
static char*
|
|
epic_pci_probe(
|
|
pcici_t config_id,
|
|
pcidi_t device_id)
|
|
{
|
|
if( PCI_VENDORID(device_id) != SMC_VENDORID )
|
|
return NULL;
|
|
|
|
if( PCI_CHIPID(device_id) == CHIPID_83C170 )
|
|
return "SMC 83c170";
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* PCI_Attach function
|
|
*
|
|
* splimp() invoked here
|
|
*/
|
|
static void
|
|
epic_pci_attach(
|
|
pcici_t config_id,
|
|
int unit)
|
|
{
|
|
struct ifnet * ifp;
|
|
epic_softc_t *sc;
|
|
#if defined(EPIC_USEIOSPACE)
|
|
u_int32_t iobase;
|
|
#else
|
|
caddr_t pmembase;
|
|
#endif
|
|
int i,s,media;
|
|
u_int32_t pool;
|
|
|
|
/* Allocate memory for softc and hardware descriptors */
|
|
sc = (epic_softc_t *) malloc(
|
|
sizeof(epic_softc_t) +
|
|
#if defined(TX_FRAG_LIST)
|
|
sizeof(struct epic_frag_list)*TX_RING_SIZE +
|
|
#endif
|
|
sizeof(struct epic_rx_desc)*RX_RING_SIZE +
|
|
sizeof(struct epic_tx_desc)*TX_RING_SIZE + PAGE_SIZE,
|
|
M_DEVBUF, M_NOWAIT);
|
|
|
|
if (sc == NULL) return;
|
|
|
|
/* Align pool on PAGE_SIZE */
|
|
pool = ((u_int32_t)sc)+sizeof(epic_softc_t);
|
|
pool = (pool + PAGE_SIZE - 1) & ~(PAGE_SIZE - 1);
|
|
|
|
/* Preinitialize softc structure */
|
|
bzero(sc, sizeof(epic_softc_t));
|
|
epics[ unit ] = sc;
|
|
sc->unit = unit;
|
|
|
|
/* Get iobase or membase */
|
|
#if defined(EPIC_USEIOSPACE)
|
|
if (!pci_map_port(config_id, PCI_CBIO,(u_short *) &(sc->iobase))) {
|
|
printf("tx%d: cannot map port\n",unit);
|
|
return;
|
|
}
|
|
#else
|
|
if (!pci_map_mem(config_id, PCI_CBMA,(vm_offset_t *) &(sc->csr),(vm_offset_t *) &pmembase)) {
|
|
printf("tx%d: cannot map memory\n",unit);
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
|
|
#if defined(TX_FRAG_LIST)
|
|
sc->tx_flist = (void *)pool;
|
|
pool += sizeof(struct epic_frag_list)*TX_RING_SIZE;
|
|
#endif
|
|
sc->rx_desc = (void *)pool;
|
|
pool += sizeof(struct epic_rx_desc)*RX_RING_SIZE;
|
|
sc->tx_desc = (void *)pool;
|
|
|
|
/* Bring the chip out of low-power mode. */
|
|
CSR_WRITE_4( sc, GENCTL, 0x0000 );
|
|
|
|
/* Magic?! If we don't set this bit the MII interface won't work. */
|
|
CSR_WRITE_4( sc, TEST1, 0x0008 );
|
|
|
|
/* Read mac address (may be better is read from EEPROM?) */
|
|
for (i = 0; i < ETHER_ADDR_LEN / sizeof( u_int16_t); i++)
|
|
((u_int16_t *)sc->epic_macaddr)[i] = CSR_READ_2( sc, LAN0 + i*4 );
|
|
|
|
/* Display some info */
|
|
printf("tx%d: address %02x:%02x:%02x:%02x:%02x:%02x,",sc->unit,
|
|
sc->epic_macaddr[0],sc->epic_macaddr[1],sc->epic_macaddr[2],
|
|
sc->epic_macaddr[3],sc->epic_macaddr[4],sc->epic_macaddr[5]);
|
|
|
|
|
|
s = splimp();
|
|
|
|
/* Map interrupt */
|
|
if( !pci_map_int(config_id, epic_intr_normal, (void*)sc, &net_imask) ) {
|
|
printf("tx%d: couldn't map interrupt\n",unit);
|
|
epics[ unit ] = NULL;
|
|
free(sc, M_DEVBUF);
|
|
return;
|
|
}
|
|
|
|
/* Fill ifnet structure */
|
|
ifp = &sc->epic_if;
|
|
|
|
ifp->if_unit = unit;
|
|
ifp->if_name = "tx";
|
|
ifp->if_softc = sc;
|
|
ifp->if_flags = IFF_BROADCAST|IFF_SIMPLEX|IFF_MULTICAST;
|
|
ifp->if_ioctl = epic_ifioctl;
|
|
ifp->if_start = epic_ifstart;
|
|
ifp->if_watchdog = epic_ifwatchdog;
|
|
ifp->if_init = (if_init_f_t*)epic_init;
|
|
ifp->if_timer = 0;
|
|
ifp->if_output = ether_output;
|
|
ifp->if_linkmib = &sc->dot3stats;
|
|
ifp->if_linkmiblen = sizeof(struct ifmib_iso_8802_3);
|
|
|
|
sc->dot3stats.dot3StatsEtherChipSet =
|
|
DOT3CHIPSET(dot3VendorSMC,
|
|
dot3ChipSetSMC83c170);
|
|
|
|
sc->dot3stats.dot3Compliance = DOT3COMPLIANCE_COLLS;
|
|
|
|
printf(" type SMC9432TX, ");
|
|
|
|
/* Identify PHY */
|
|
sc->phyid = epic_read_phy_register( sc, DP83840_PHYIDR1 )<<6;
|
|
sc->phyid|= (epic_read_phy_register( sc, DP83840_PHYIDR2 )>>10)&0x3F;
|
|
|
|
if( QS6612_OUI == sc->phyid ){
|
|
printf("phy QS6612, ");
|
|
} else if( DP83840_OUI == sc->phyid ){
|
|
printf("phy DP83840, ");
|
|
} else {
|
|
printf("phy unknown (%x), ",sc->phyid);
|
|
sc->phyid = DP83840_OUI;
|
|
}
|
|
|
|
/* Read current config */
|
|
i = epic_read_phy_register( sc, DP83840_BMCR );
|
|
|
|
#if defined(_NET_IF_MEDIA_H_)
|
|
media = IFM_ETHER;
|
|
#endif
|
|
|
|
if( i & BMCR_AUTONEGOTIATION ){
|
|
i = epic_read_phy_register( sc, DP83840_LPAR );
|
|
|
|
printf("Auto-Neg ");
|
|
|
|
if( i & (ANAR_100_TX|ANAR_100_TX_FD) )
|
|
printf("100Mbps ");
|
|
else
|
|
printf("10Mbps ");
|
|
|
|
if( i & (ANAR_10_FD|ANAR_100_TX_FD) ) printf("FD");
|
|
|
|
#if defined(_NET_IF_MEDIA_H_)
|
|
media |= IFM_AUTO;
|
|
#endif
|
|
} else {
|
|
#if !defined(_NET_IF_MEDIA_H_)
|
|
ifp->if_flags |= IFF_LINK0;
|
|
#endif
|
|
if( i & BMCR_100MBPS ) {
|
|
printf("100Mbps ");
|
|
#if defined(_NET_IF_MEDIA_H_)
|
|
media |= IFM_100_TX;
|
|
#else
|
|
ifp->if_flags |= IFF_LINK2;
|
|
#endif
|
|
} else {
|
|
printf("10Mbps ");
|
|
#if defined(_NET_IF_MEDIA_H_)
|
|
media |= IFM_10_T;
|
|
#endif
|
|
}
|
|
if( i & BMCR_FULL_DUPLEX ) {
|
|
printf("FD");
|
|
#if defined(_NET_IF_MEDIA_H_)
|
|
media |= IFM_FDX;
|
|
#else
|
|
ifp->if_flags |= IFF_LINK1;
|
|
#endif
|
|
}
|
|
}
|
|
|
|
printf("\n");
|
|
|
|
#if defined(SIOCSIFMEDIA) && !defined(EPIC_NOIFMEDIA)
|
|
/* init ifmedia interface */
|
|
ifmedia_init(&sc->ifmedia,0,epic_ifmedia_change,epic_ifmedia_status);
|
|
ifmedia_add(&sc->ifmedia,IFM_ETHER|IFM_10_T,0,NULL);
|
|
ifmedia_add(&sc->ifmedia,IFM_ETHER|IFM_10_T|IFM_FDX,0,NULL);
|
|
ifmedia_add(&sc->ifmedia,IFM_ETHER|IFM_100_TX,0,NULL);
|
|
ifmedia_add(&sc->ifmedia,IFM_ETHER|IFM_100_TX|IFM_FDX,0,NULL);
|
|
ifmedia_add(&sc->ifmedia,IFM_ETHER|IFM_AUTO,0,NULL);
|
|
ifmedia_set(&sc->ifmedia, media);
|
|
#endif
|
|
|
|
/* Read MBSR twice to update latched bits */
|
|
epic_read_phy_register( sc, DP83840_BMSR );
|
|
i=epic_read_phy_register( sc, DP83840_BMSR );
|
|
|
|
if( !(i & BMSR_LINK_STATUS) )
|
|
printf("tx%d: WARNING! no link established\n",sc->unit);
|
|
|
|
/* Set shut down routine to stop DMA processes on reboot */
|
|
at_shutdown(epic_shutdown, sc, SHUTDOWN_POST_SYNC);
|
|
|
|
/*
|
|
* Attach to if manager
|
|
*/
|
|
if_attach(ifp);
|
|
ether_ifattach(ifp);
|
|
|
|
#if NBPFILTER > 0
|
|
bpfattach(ifp,DLT_EN10MB, sizeof(struct ether_header));
|
|
#endif
|
|
|
|
splx(s);
|
|
|
|
return;
|
|
}
|
|
|
|
#if defined(SIOCSIFMEDIA) && !defined(EPIC_NOIFMEDIA)
|
|
static int
|
|
epic_ifmedia_change __P((
|
|
struct ifnet * ifp))
|
|
{
|
|
if (IFM_TYPE(((epic_softc_t *)(ifp->if_softc))->ifmedia.ifm_media) != IFM_ETHER)
|
|
return (EINVAL);
|
|
|
|
epic_set_media_speed( ifp->if_softc );
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
epic_ifmedia_status __P((
|
|
struct ifnet * ifp,
|
|
struct ifmediareq *ifmr))
|
|
{
|
|
epic_softc_t *sc = ifp->if_softc;
|
|
u_int32_t bmcr;
|
|
u_int32_t bmsr;
|
|
|
|
bmcr = epic_read_phy_register( sc, DP83840_BMCR );
|
|
|
|
epic_read_phy_register( sc, DP83840_BMSR );
|
|
bmsr = epic_read_phy_register( sc, DP83840_BMSR );
|
|
|
|
ifmr->ifm_active = IFM_ETHER;
|
|
ifmr->ifm_status = IFM_AVALID;
|
|
|
|
if( !(bmsr & BMSR_LINK_STATUS) ) {
|
|
ifmr->ifm_active |= bmcr&BMCR_AUTONEGOTIATION?IFM_AUTO:IFM_NONE;
|
|
return;
|
|
}
|
|
|
|
ifmr->ifm_status |= IFM_ACTIVE;
|
|
ifmr->ifm_active |= (bmcr&BMCR_100MBPS)?IFM_100_TX:IFM_10_T;
|
|
ifmr->ifm_active |= (bmcr&BMCR_FULL_DUPLEX)?IFM_FDX:0;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* IFINIT function
|
|
*
|
|
* splimp() invoked here
|
|
*/
|
|
static int
|
|
epic_init __P((
|
|
epic_softc_t * sc))
|
|
{
|
|
struct ifnet *ifp = &sc->epic_if;
|
|
int i,s;
|
|
|
|
s = splimp();
|
|
|
|
/* Soft reset the chip */
|
|
CSR_WRITE_4( sc, GENCTL, GENCTL_SOFT_RESET );
|
|
|
|
/* Reset takes 15 pci ticks which depends on processor speed*/
|
|
DELAY(1);
|
|
|
|
/* Wake up */
|
|
CSR_WRITE_4( sc, GENCTL, 0 );
|
|
|
|
/* ?????? */
|
|
CSR_WRITE_4( sc, TEST1, 0x0008);
|
|
|
|
/* Initialize rings */
|
|
if( -1 == epic_init_rings( sc ) ) {
|
|
printf("tx%d: failed to initialize rings\n",sc->unit);
|
|
epic_free_rings( sc );
|
|
splx(s);
|
|
return -1;
|
|
}
|
|
|
|
/* Put node address to EPIC */
|
|
CSR_WRITE_4( sc, LAN0, ((u_int16_t *)sc->epic_macaddr)[0] );
|
|
CSR_WRITE_4( sc, LAN1, ((u_int16_t *)sc->epic_macaddr)[1] );
|
|
CSR_WRITE_4( sc, LAN2, ((u_int16_t *)sc->epic_macaddr)[2] );
|
|
|
|
/* Set transmit threshold */
|
|
CSR_WRITE_4( sc, ETXTHR, 0x40 );
|
|
|
|
/* Compute and set RXCON. */
|
|
epic_set_rx_mode( sc );
|
|
|
|
/* Set media speed mode */
|
|
epic_init_phy( sc );
|
|
epic_set_media_speed( sc );
|
|
|
|
/* Set multicast table */
|
|
epic_set_mc_table( sc );
|
|
|
|
/* Enable interrupts by setting the interrupt mask. */
|
|
if( QS6612_OUI == sc->phyid ) {
|
|
CSR_WRITE_4( sc, INTMASK,
|
|
INTSTAT_RCC | INTSTAT_RQE | INTSTAT_OVW | INTSTAT_RXE |
|
|
INTSTAT_TXC | INTSTAT_TCC | INTSTAT_TQE | INTSTAT_TXU |
|
|
INTSTAT_CNT | INTSTAT_GP2 | INTSTAT_FATAL |
|
|
INTSTAT_PTA | INTSTAT_PMA | INTSTAT_APE | INTSTAT_DPE );
|
|
} else {
|
|
CSR_WRITE_4( sc, INTMASK,
|
|
INTSTAT_RCC | INTSTAT_RQE | INTSTAT_OVW | INTSTAT_RXE |
|
|
INTSTAT_TXC | INTSTAT_TCC | INTSTAT_TQE | INTSTAT_TXU |
|
|
INTSTAT_CNT | INTSTAT_FATAL |
|
|
INTSTAT_PTA | INTSTAT_PMA | INTSTAT_APE | INTSTAT_DPE );
|
|
}
|
|
|
|
/* Enable interrupts, set for PCI read multiple and etc */
|
|
CSR_WRITE_4( sc, GENCTL,
|
|
GENCTL_ENABLE_INTERRUPT | GENCTL_MEMORY_READ_MULTIPLE |
|
|
GENCTL_ONECOPY | GENCTL_RECEIVE_FIFO_THRESHOLD64 );
|
|
|
|
/* Start rx process */
|
|
CSR_WRITE_4( sc, COMMAND, COMMAND_RXQUEUED | COMMAND_START_RX );
|
|
|
|
/* Mark interface running ... */
|
|
if( ifp->if_flags & IFF_UP ) ifp->if_flags |= IFF_RUNNING;
|
|
else ifp->if_flags &= ~IFF_RUNNING;
|
|
|
|
/* ... and free */
|
|
ifp->if_flags &= ~IFF_OACTIVE;
|
|
|
|
splx(s);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* This function should set EPIC's registers according IFF_* flags
|
|
*/
|
|
static void
|
|
epic_set_rx_mode(
|
|
epic_softc_t * sc)
|
|
{
|
|
struct ifnet *ifp = &sc->epic_if;
|
|
u_int16_t rxcon = 0;
|
|
|
|
#if NBPFILTER > 0
|
|
if( sc->epic_if.if_flags & IFF_PROMISC )
|
|
rxcon |= RXCON_PROMISCUOUS_MODE;
|
|
#endif
|
|
|
|
if( sc->epic_if.if_flags & IFF_BROADCAST )
|
|
rxcon |= RXCON_RECEIVE_BROADCAST_FRAMES;
|
|
|
|
if( sc->epic_if.if_flags & IFF_MULTICAST )
|
|
rxcon |= RXCON_RECEIVE_MULTICAST_FRAMES;
|
|
|
|
CSR_WRITE_4( sc, RXCON, rxcon );
|
|
|
|
return;
|
|
}
|
|
|
|
static void
|
|
epic_init_phy __P((
|
|
epic_softc_t * sc))
|
|
{
|
|
u_int32_t i;
|
|
|
|
/* Reset PHY */
|
|
epic_write_phy_register( sc, DP83840_BMCR, BMCR_RESET );
|
|
for(i=0;i<0x100000;i++)
|
|
if( !(epic_read_phy_register( sc, DP83840_BMCR ) & BMCR_RESET) )
|
|
break;
|
|
|
|
if( epic_read_phy_register( sc, DP83840_BMCR ) & BMCR_RESET )
|
|
printf("tx%d: WARNING! cannot reset PHY\n",sc->unit);
|
|
|
|
switch( sc->phyid ){
|
|
case QS6612_OUI:
|
|
/* Init QS6612 to generate interrupt when AutoNeg complete */
|
|
CSR_WRITE_4( sc, NVCTL, NVCTL_GP1_OUTPUT_ENABLE );
|
|
epic_read_phy_register( sc, QS6612_INTSTAT );
|
|
epic_write_phy_register( sc, QS6612_INTMASK,
|
|
INTMASK_THUNDERLAN|INTSTAT_AN_COMPLETE );
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* This function should set MII to mode specified by IFF_LINK* flags or
|
|
* ifmedia structure.
|
|
*/
|
|
static void
|
|
epic_set_media_speed __P((
|
|
epic_softc_t * sc))
|
|
{
|
|
#if defined(_NET_IF_MEDIA_H_)
|
|
u_int32_t tgtmedia = sc->ifmedia.ifm_cur->ifm_media;
|
|
#else
|
|
struct ifnet *ifp = &sc->epic_if;
|
|
#endif
|
|
u_int16_t media;
|
|
|
|
/* Set media speed */
|
|
|
|
#if defined(_NET_IF_MEDIA_H_)
|
|
if( IFM_SUBTYPE(tgtmedia) != IFM_AUTO ){
|
|
/* Set mode */
|
|
media = (IFM_SUBTYPE(tgtmedia)==IFM_100_TX) ? BMCR_100MBPS : 0;
|
|
media|= (tgtmedia&IFM_FDX) ? BMCR_FULL_DUPLEX : 0;
|
|
|
|
sc->epic_if.if_baudrate =
|
|
(IFM_SUBTYPE(tgtmedia)==IFM_100_TX)?100000000:10000000;
|
|
|
|
epic_write_phy_register( sc, DP83840_BMCR, media );
|
|
|
|
CSR_WRITE_4( sc, TXCON,(tgtmedia&IFM_FDX)?TXCON_LOOPBACK_MODE_FULL_DUPLEX|TXCON_DEFAULT:TXCON_DEFAULT );
|
|
}
|
|
#else
|
|
if( ifp->if_flags & IFF_LINK0 ) {
|
|
/* Set mode */
|
|
media = (ifp->if_flags & IFF_LINK2) ? BMCR_100MBPS : 0;
|
|
media|= (ifp->if_flags & IFF_LINK1) ? BMCR_FULL_DUPLEX : 0;
|
|
|
|
sc->epic_if.if_baudrate =
|
|
(ifp->if_flags & IFF_LINK2)?100000000:10000000;
|
|
|
|
epic_write_phy_register( sc, DP83840_BMCR, media );
|
|
|
|
CSR_WRITE_4( sc, TXCON, (ifp->if_flags & IFF_LINK2) ? TXCON_LOOPBACK_MODE_FULL_DUPLEX|TXCON_DEFAULT : TXCON_DEFAULT );
|
|
}
|
|
#endif
|
|
else {
|
|
sc->epic_if.if_baudrate = 100000000;
|
|
|
|
CSR_WRITE_4( sc, TXCON, TXCON_DEFAULT );
|
|
|
|
/* Set and restart autoneg */
|
|
epic_write_phy_register( sc, DP83840_BMCR,
|
|
BMCR_AUTONEGOTIATION | BMCR_RESTART_AUTONEG );
|
|
|
|
/* If it is not QS6612 PHY, try to get result of autoneg. */
|
|
if( QS6612_OUI != sc->phyid ) {
|
|
/* Wait 3 seconds for the autoneg to finish
|
|
* This is the recommended time from the DP83840A data
|
|
* sheet Section 7.1
|
|
*/
|
|
DELAY(3000000);
|
|
|
|
if( epic_autoneg(sc) == EPIC_FULL_DUPLEX )
|
|
CSR_WRITE_4( sc, TXCON,
|
|
TXCON_LOOPBACK_MODE_FULL_DUPLEX|TXCON_DEFAULT);
|
|
}
|
|
/* Else it will be done when GP2 int occured */
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* This functions get results of the autoneg processes of the phy
|
|
* It implements the workaround that is described in section 7.2 & 7.3 of the
|
|
* DP83840A data sheet
|
|
* http://www.national.com/ds/DP/DP83840A.pdf
|
|
*/
|
|
static int
|
|
epic_autoneg(
|
|
epic_softc_t * sc)
|
|
{
|
|
struct ifnet *ifp = &sc->epic_if;
|
|
u_int16_t media;
|
|
u_int16_t i;
|
|
|
|
/* BMSR must be read twice to update the link status bit
|
|
* since that bit is a latch bit
|
|
*/
|
|
epic_read_phy_register( sc, DP83840_BMSR);
|
|
i = epic_read_phy_register( sc, DP83840_BMSR);
|
|
|
|
if ((i & BMSR_LINK_STATUS) && (i & BMSR_AUTONEG_COMPLETE)){
|
|
i = epic_read_phy_register( sc, DP83840_LPAR );
|
|
|
|
if ( i & (ANAR_100_TX_FD|ANAR_10_FD) )
|
|
return EPIC_FULL_DUPLEX;
|
|
else
|
|
return EPIC_HALF_DUPLEX;
|
|
}
|
|
else { /*Auto-negotiation or link status is not 1
|
|
Thus the auto-negotiation failed and one
|
|
must take other means to fix it.
|
|
*/
|
|
|
|
/* ANER must be read twice to get the correct reading for the
|
|
* Multiple link fault bit -- it is a latched bit
|
|
*/
|
|
epic_read_phy_register( sc, DP83840_ANER );
|
|
i = epic_read_phy_register( sc, DP83840_ANER );
|
|
|
|
if ( i & ANER_MULTIPLE_LINK_FAULT ) {
|
|
/* it can be forced to 100Mb/s Half-Duplex */
|
|
media = epic_read_phy_register( sc, DP83840_BMCR );
|
|
media &= ~(BMCR_AUTONEGOTIATION | BMCR_FULL_DUPLEX);
|
|
media |= BMCR_100MBPS;
|
|
epic_write_phy_register( sc, DP83840_BMCR, media );
|
|
|
|
/* read BMSR again to determine link status */
|
|
epic_read_phy_register( sc, DP83840_BMSR );
|
|
i=epic_read_phy_register( sc, DP83840_BMSR );
|
|
|
|
if (i & BMSR_LINK_STATUS){
|
|
/* port is linked to the non Auto-Negotiation
|
|
* 100Mbs partner.
|
|
*/
|
|
return EPIC_HALF_DUPLEX;
|
|
}
|
|
else {
|
|
media = epic_read_phy_register( sc, DP83840_BMCR);
|
|
media &= ~(BMCR_AUTONEGOTIATION | BMCR_FULL_DUPLEX | BMCR_100MBPS);
|
|
epic_write_phy_register( sc, DP83840_BMCR, media);
|
|
epic_read_phy_register( sc, DP83840_BMSR );
|
|
i=epic_read_phy_register( sc, DP83840_BMSR );
|
|
|
|
if (i & BMSR_LINK_STATUS) {
|
|
/*port is linked to the non
|
|
* Auto-Negotiation10Mbs partner
|
|
*/
|
|
return EPIC_HALF_DUPLEX;
|
|
}
|
|
}
|
|
}
|
|
/* If we get here we are most likely not connected
|
|
* so lets default it to half duplex
|
|
*/
|
|
return EPIC_HALF_DUPLEX;
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
* This function sets EPIC multicast table
|
|
*/
|
|
static void
|
|
epic_set_mc_table (
|
|
epic_softc_t * sc)
|
|
{
|
|
struct ifnet *ifp = &sc->epic_if;
|
|
|
|
if( ifp->if_flags & IFF_MULTICAST ){
|
|
CSR_WRITE_4( sc, MC0, 0xFFFF );
|
|
CSR_WRITE_4( sc, MC1, 0xFFFF );
|
|
CSR_WRITE_4( sc, MC2, 0xFFFF );
|
|
CSR_WRITE_4( sc, MC3, 0xFFFF );
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
static void
|
|
epic_shutdown(
|
|
int howto,
|
|
void *sc)
|
|
{
|
|
epic_stop(sc);
|
|
}
|
|
|
|
/*
|
|
* This function should completely stop rx and tx processes
|
|
*
|
|
* splimp() invoked here
|
|
*/
|
|
static void
|
|
epic_stop(
|
|
epic_softc_t * sc)
|
|
{
|
|
int i,s;
|
|
|
|
s = splimp();
|
|
sc->epic_if.if_timer = 0;
|
|
|
|
/* Disable interrupts, stop processes */
|
|
CSR_WRITE_4( sc, INTMASK, 0 );
|
|
CSR_WRITE_4( sc, GENCTL, 0 );
|
|
CSR_WRITE_4( sc, COMMAND,
|
|
COMMAND_STOP_RX | COMMAND_STOP_RDMA | COMMAND_STOP_TDMA );
|
|
|
|
/* Wait RX and TX DMA to stop */
|
|
for(i=0;i<0x100000;i++){
|
|
if( (CSR_READ_4(sc,INTSTAT)&(INTSTAT_RXIDLE|INTSTAT_TXIDLE)) ==
|
|
(INTSTAT_RXIDLE|INTSTAT_TXIDLE) ) break;
|
|
}
|
|
|
|
if( !(CSR_READ_4(sc,INTSTAT)&INTSTAT_RXIDLE) )
|
|
printf("tx%d: can't stop RX DMA\n",sc->unit);
|
|
|
|
if( !(CSR_READ_4(sc,INTSTAT)&INTSTAT_TXIDLE) )
|
|
printf("tx%d: can't stop TX DMA\n",sc->unit);
|
|
|
|
/* Reset chip and phy */
|
|
CSR_WRITE_4( sc, GENCTL, GENCTL_SOFT_RESET );
|
|
|
|
/* Need to wait for 15 pci ticks to pass before accessing again*/
|
|
DELAY(1);
|
|
|
|
/* Free memory allocated for rings */
|
|
epic_free_rings( sc );
|
|
|
|
splx(s);
|
|
|
|
}
|
|
|
|
/*
|
|
* This function should free all allocated for rings memory.
|
|
* NB: The DMA processes must be stopped.
|
|
*
|
|
* splimp() assumed to be done
|
|
*/
|
|
static void
|
|
epic_free_rings(epic_softc_t * sc){
|
|
int i;
|
|
|
|
for(i=0;i<RX_RING_SIZE;i++){
|
|
struct epic_rx_buffer *buf = sc->rx_buffer + i;
|
|
struct epic_rx_desc *desc = sc->rx_desc + i;
|
|
|
|
desc->status = 0;
|
|
desc->buflength = 0;
|
|
desc->bufaddr = 0;
|
|
|
|
#if defined(RX_TO_MBUF)
|
|
if( buf->mbuf ) m_freem( buf->mbuf );
|
|
buf->mbuf = NULL;
|
|
#else
|
|
if( buf->data ) free( buf->data, M_DEVBUF );
|
|
buf->data = NULL;
|
|
#endif
|
|
}
|
|
|
|
for(i=0;i<TX_RING_SIZE;i++){
|
|
struct epic_tx_buffer *buf = sc->tx_buffer + i;
|
|
struct epic_tx_desc *desc = sc->tx_desc + i;
|
|
|
|
desc->status = 0;
|
|
desc->buflength = 0;
|
|
desc->bufaddr = 0;
|
|
|
|
#if defined(TX_FRAG_LIST)
|
|
if( buf->mbuf ) m_freem( buf->mbuf );
|
|
buf->mbuf = NULL;
|
|
#else
|
|
if( buf->data ) free( buf->data, M_DEVBUF );
|
|
buf->data = NULL;
|
|
#endif
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Initialize Rx and Tx rings and give them to EPIC
|
|
*
|
|
* If RX_TO_MBUF option is enabled, mbuf cluster is allocated instead of
|
|
* static buffer for RX ringi element.
|
|
* If TX_FRAG_LIST option is enabled, nothig is done, except chaining
|
|
* descriptors to ring and point them to static fraglists.
|
|
*
|
|
* splimp() assumed to be done
|
|
*/
|
|
static int
|
|
epic_init_rings(epic_softc_t * sc){
|
|
int i;
|
|
struct mbuf *m;
|
|
|
|
sc->cur_rx = sc->cur_tx = sc->dirty_tx = sc->pending_txs = 0;
|
|
|
|
for (i = 0; i < RX_RING_SIZE; i++) {
|
|
struct epic_rx_buffer *buf = sc->rx_buffer + i;
|
|
struct epic_rx_desc *desc = sc->rx_desc + i;
|
|
|
|
desc->status = 0; /* Owned by driver */
|
|
desc->next = vtophys( sc->rx_desc + ((i+1)%RX_RING_SIZE) );
|
|
|
|
if( (desc->next & 3) ||
|
|
((desc->next & 0xFFF) + sizeof(struct epic_rx_desc) > 0x1000 ) )
|
|
printf("tx%d: WARNING! frag_list is misbound or misaligned\n",sc->unit);
|
|
|
|
#if defined(RX_TO_MBUF)
|
|
MGETHDR(buf->mbuf,M_DONTWAIT,MT_DATA);
|
|
if( NULL == buf->mbuf ) return -1;
|
|
MCLGET(buf->mbuf,M_DONTWAIT);
|
|
if( NULL == (buf->mbuf->m_flags & M_EXT) ) return -1;
|
|
desc->bufaddr = vtophys( mtod(buf->mbuf,caddr_t) );
|
|
#else
|
|
buf->data = malloc(ETHER_MAX_FRAME_LEN, M_DEVBUF, M_NOWAIT);
|
|
if( buf->data == NULL ) return -1;
|
|
desc->bufaddr = vtophys( buf->data );
|
|
#endif
|
|
|
|
desc->buflength = ETHER_MAX_FRAME_LEN;
|
|
desc->status = 0x8000; /* Give to EPIC */
|
|
|
|
}
|
|
|
|
for (i = 0; i < TX_RING_SIZE; i++) {
|
|
struct epic_tx_buffer *buf = sc->tx_buffer + i;
|
|
struct epic_tx_desc *desc = sc->tx_desc + i;
|
|
|
|
desc->status = 0;
|
|
desc->next = vtophys( sc->tx_desc + ( (i+1)%TX_RING_SIZE ) );
|
|
|
|
if( (desc->next & 3) ||
|
|
((desc->next & 0xFFF) + sizeof(struct epic_tx_desc) > 0x1000 ) )
|
|
printf("tx%d: WARNING! frag_list is misbound or misaligned\n",sc->unit);
|
|
|
|
#if defined(TX_FRAG_LIST)
|
|
buf->mbuf = NULL;
|
|
desc->bufaddr = vtophys( sc->tx_flist + i );
|
|
if( (desc->bufaddr & 3) ||
|
|
((desc->bufaddr & 0xFFF) + sizeof(struct epic_frag_list) > 0x1000 ) )
|
|
printf("tx%d: WARNING! frag_list is misbound or misaligned\n",sc->unit);
|
|
#else
|
|
/* Allocate buffer */
|
|
buf->data = malloc(ETHER_MAX_FRAME_LEN, M_DEVBUF, M_NOWAIT);
|
|
|
|
if( buf->data == NULL ) return -1;
|
|
|
|
desc->bufaddr = vtophys( buf->data );
|
|
desc->buflength = ETHER_MAX_FRAME_LEN;
|
|
#endif
|
|
}
|
|
|
|
/* Give rings to EPIC */
|
|
CSR_WRITE_4( sc, PRCDAR, vtophys( sc->rx_desc ) );
|
|
CSR_WRITE_4( sc, PTCDAR, vtophys( sc->tx_desc ) );
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* EEPROM operation functions
|
|
*/
|
|
static void epic_write_eepromreg __P((
|
|
epic_softc_t *sc,
|
|
u_int8_t val))
|
|
{
|
|
u_int16_t i;
|
|
|
|
CSR_WRITE_1( sc, EECTL, val );
|
|
|
|
for( i=0;i<0xFF; i++)
|
|
if( !(CSR_READ_1( sc, EECTL ) & 0x20) ) break;
|
|
|
|
return;
|
|
}
|
|
|
|
static u_int8_t epic_read_eepromreg __P((
|
|
epic_softc_t *sc))
|
|
{
|
|
return CSR_READ_1( sc,EECTL );
|
|
}
|
|
|
|
static u_int8_t epic_eeprom_clock __P((
|
|
epic_softc_t *sc,
|
|
u_int8_t val))
|
|
{
|
|
epic_write_eepromreg( sc, val );
|
|
epic_write_eepromreg( sc, (val | 0x4) );
|
|
epic_write_eepromreg( sc, val );
|
|
|
|
return epic_read_eepromreg( sc );
|
|
}
|
|
|
|
static void epic_output_eepromw __P((
|
|
epic_softc_t * sc,
|
|
u_int16_t val))
|
|
{
|
|
int i;
|
|
for( i = 0xF; i >= 0; i--){
|
|
if( (val & (1 << i)) ) epic_eeprom_clock( sc, 0x0B );
|
|
else epic_eeprom_clock( sc, 3);
|
|
}
|
|
}
|
|
|
|
static u_int16_t epic_input_eepromw __P((
|
|
epic_softc_t *sc))
|
|
{
|
|
int i;
|
|
int tmp;
|
|
u_int16_t retval = 0;
|
|
|
|
for( i = 0xF; i >= 0; i--) {
|
|
tmp = epic_eeprom_clock( sc, 0x3 );
|
|
if( tmp & 0x10 ){
|
|
retval |= (1 << i);
|
|
}
|
|
}
|
|
return retval;
|
|
}
|
|
|
|
static int epic_read_eeprom __P((
|
|
epic_softc_t *sc,
|
|
u_int16_t loc))
|
|
{
|
|
int i;
|
|
u_int16_t dataval;
|
|
u_int16_t read_cmd;
|
|
|
|
epic_write_eepromreg( sc , 3);
|
|
|
|
if( epic_read_eepromreg( sc ) & 0x40 )
|
|
read_cmd = ( loc & 0x3F ) | 0x180;
|
|
else
|
|
read_cmd = ( loc & 0xFF ) | 0x600;
|
|
|
|
epic_output_eepromw( sc, read_cmd );
|
|
|
|
dataval = epic_input_eepromw( sc );
|
|
|
|
epic_write_eepromreg( sc, 1 );
|
|
|
|
return dataval;
|
|
}
|
|
|
|
static int epic_read_phy_register __P((
|
|
epic_softc_t *sc,
|
|
u_int16_t loc))
|
|
{
|
|
int i;
|
|
|
|
CSR_WRITE_4( sc, MIICTL, ((loc << 4) | 0x0601) );
|
|
|
|
for( i=0;i<0x1000;i++) if( !(CSR_READ_4( sc, MIICTL )&1) ) break;
|
|
|
|
return CSR_READ_4( sc, MIIDATA );
|
|
}
|
|
|
|
static void epic_write_phy_register __P((
|
|
epic_softc_t * sc,
|
|
u_int16_t loc,
|
|
u_int16_t val))
|
|
{
|
|
int i;
|
|
|
|
CSR_WRITE_4( sc, MIIDATA, val );
|
|
CSR_WRITE_4( sc, MIICTL, ((loc << 4) | 0x0602) );
|
|
|
|
for( i=0;i<0x1000;i++) if( !(CSR_READ_4( sc, MIICTL )&2) ) break;
|
|
|
|
return;
|
|
}
|
|
|
|
#endif /* NPCI > 0 */
|