/*
* marvell kirkwood gigabit ethernet (88e1116 and 88e1121) driver
* (as found in the sheevaplug, openrd and guruplug).
* the main difference is the flavour of phy kludgery necessary.
*
* from /public/doc/marvell/88f61xx.kirkwood.pdf,
* /public/doc/marvell/88e1116.pdf, and
* /public/doc/marvell/88e1121r.pdf.
*/
#include "u.h"
#include "../port/lib.h"
#include "mem.h"
#include "dat.h"
#include "fns.h"
#include "io.h"
#include "../port/error.h"
#include "../port/netif.h"
#include "etherif.h"
#include "ethermii.h"
#include "../ip/ip.h"
#define MIIDBG if(0)iprint
#define WINATTR(v) (((v) & MASK(8)) << 8)
#define WINSIZE(v) (((v)/(64*1024) - 1) << 16)
enum {
Nrx = 512,
Ntx = 32,
Nrxblks = 1024,
Rxblklen = 2+1522, /* ifc. supplies first 2 bytes as padding */
Maxrxintrsec = 20*1000, /* max. rx intrs. / sec */
Etherstuck = 70, /* must send or receive a packet in this many sec.s */
Descralign = 16,
Bufalign = 8,
Pass = 1, /* accept packets */
Qno = 0, /* do everything on queue zero */
};
typedef struct Ctlr Ctlr;
typedef struct Gbereg Gbereg;
typedef struct Mibstats Mibstats;
typedef struct Rx Rx;
typedef struct Tx Tx;
static struct {
Lock;
Block *head;
} freeblocks;
/* hardware receive buffer descriptor */
struct Rx {
ulong cs;
ulong countsize; /* bytes, buffer size */
ulong buf; /* phys. addr. of packet buffer */
ulong next; /* phys. addr. of next Rx */
};
/* hardware transmit buffer descriptor */
struct Tx {
ulong cs;
ulong countchk; /* bytes, checksum */
ulong buf; /* phys. addr. of packet buffer */
ulong next; /* phys. addr. of next Tx */
};
/* fixed by hw; part of Gberegs */
struct Mibstats {
union {
uvlong rxby; /* good bytes rcv'd */
struct {
ulong rxbylo;
ulong rxbyhi;
};
};
ulong badrxby; /* bad bytes rcv'd */
ulong mactxerr; /* tx err pkts */
ulong rxpkt; /* good pkts rcv'd */
ulong badrxpkt; /* bad pkts rcv'd */
ulong rxbcastpkt; /* b'cast pkts rcv'd */
ulong rxmcastpkt; /* m'cast pkts rcv'd */
ulong rx64; /* pkts <= 64 bytes */
ulong rx65_127; /* pkts 65—127 bytes */
ulong rx128_255; /* pkts 128—255 bytes */
ulong rx256_511; /* pkts 256—511 bytes */
ulong rx512_1023; /* pkts 512—1023 bytes */
ulong rx1024_max; /* pkts >= 1024 bytes */
union {
uvlong txby; /* good bytes sent */
struct {
ulong txbylo;
ulong txbyhi;
};
};
ulong txpkt; /* good pkts sent */
/* half-duplex: pkts dropped due to excessive collisions */
ulong txcollpktdrop;
ulong txmcastpkt; /* m'cast pkts sent */
ulong txbcastpkt; /* b'cast pkts sent */
ulong badmacctlpkts; /* bad mac ctl pkts */
ulong txflctl; /* flow-control pkts sent */
ulong rxflctl; /* good flow-control pkts rcv'd */
ulong badrxflctl; /* bad flow-control pkts rcv'd */
ulong rxundersized; /* runts */
ulong rxfrags; /* fragments rcv'd */
ulong rxtoobig; /* oversized pkts rcv'd */
ulong rxjabber; /* jabber pkts rcv'd */
ulong rxerr; /* rx error events */
ulong crcerr; /* crc error events */
ulong collisions; /* collision events */
ulong latecoll; /* late collisions */
};
struct Ctlr {
Lock;
Ether *ether;
Gbereg *reg;
Lock initlock;
int init;
Rx *rx; /* receive descriptors */
Block *rxb[Nrx]; /* blocks belonging to the descriptors */
int rxhead; /* descr ethernet will write to next */
int rxtail; /* next descr that might need a buffer */
Rendez rrendez; /* interrupt wakes up read process */
int haveinput;
Tx *tx;
Block *txb[Ntx];
int txhead; /* next descr we can use for new packet */
int txtail; /* next descr to reclaim on tx complete */
Mii *mii;
int port;
/* stats */
ulong intrs;
ulong newintrs;
ulong txunderrun;
ulong txringfull;
ulong rxdiscard;
ulong rxoverrun;
ulong nofirstlast;
Mibstats;
};
#define Rxqon(q) (1<<(q))
#define Txqon(q) (1<<(q))
enum {
/* euc bits */
Portreset = 1 << 20,
/* sdma config, sdc bits */
Burst1 = 0,
Burst2,
Burst4,
Burst8,
Burst16,
SDCrifb = 1<<0, /* rx intr on pkt boundaries */
#define SDCrxburst(v) ((v)<<1)
SDCrxnobyteswap = 1<<4,
SDCtxnobyteswap = 1<<5,
SDCswap64byte = 1<<6,
#define SDCtxburst(v) ((v)<<22)
/* rx intr ipg (inter packet gap) */
#define SDCipgintrx(v) ((((v)>>15) & 1)<<25) | (((v) & MASK(15))<<7)
/* portcfg bits */
PCFGupromisc = 1<<0, /* unicast promiscuous mode */
#define Rxqdefault(q) ((q)<<1)
#define Rxqarp(q) ((q)<<4)
PCFGbcrejectnoiparp = 1<<7,
PCFGbcrejectip = 1<<8,
PCFGbcrejectarp = 1<<9,
PCFGamnotxes = 1<<12, /* auto mode, no summary update on tx */
PCFGtcpq = 1<<14, /* capture tcp frames to tcpq */
PCFGudpq = 1<<15, /* capture udp frames to udpq */
#define Rxqtcp(q) ((q)<<16)
#define Rxqudp(q) ((q)<<19)
#define Rxqbpdu(q) ((q)<<22)
PCFGrxcs = 1<<25, /* rx tcp checksum mode with header */
/* portcfgx bits */
PCFGXspanq = 1<<1,
PCFGXcrcoff = 1<<2, /* no ethernet crc */
/* port serial control0, psc0 bits */
PSC0porton = 1<<0,
PSC0forcelinkup = 1<<1,
PSC0an_dplxoff = 1<<2, /* an_ = auto. negotiate */
PSC0an_flctloff = 1<<3,
PSC0an_pauseadv = 1<<4,
PSC0nofrclinkdown = 1<<10,
PSC0an_spdoff = 1<<13,
PSC0dteadv = 1<<14, /* dte advertise */
/* max. input pkt size */
#define PSC0mru(v) ((v)<<17)
PSC0mrumask = PSC0mru(MASK(3)),
PSC0mru1518 = 0, /* 1500+2* 6(addrs) +2 + 4(crc) */
PSC0mru1522, /* 1518 + 4(vlan tags) */
PSC0mru1552, /* `baby giant' */
PSC0mru9022, /* `jumbo' */
PSC0mru9192, /* bigger jumbo */
PSC0mru9700, /* still bigger jumbo */
PSC0fd_frc = 1<<21, /* force full duplex */
PSC0flctlfrc = 1<<22,
PSC0gmiispd_gbfrc = 1<<23,
PSC0miispdfrc100mbps = 1<<24,
/* port status 0, ps0 bits */
PS0linkup = 1<<1,
PS0fd = 1<<2, /* full duplex */
PS0flctl = 1<<3,
PS0gmii_gb = 1<<4,
PS0mii100mbps = 1<<5,
PS0txbusy = 1<<7,
PS0txfifoempty = 1<<10,
PS0rxfifo1empty = 1<<11,
PS0rxfifo2empty = 1<<12,
/* port serial control 1, psc1 bits */
PSC1loopback = 1<<1,
PSC1mii = 0<<2,
PSC1rgmii = 1<<3, /* enable RGMII */
PSC1portreset = 1<<4,
PSC1clockbypass = 1<<5,
PSC1iban = 1<<6,
PSC1iban_bypass = 1<<7,
PSC1iban_restart= 1<<8,
PSC1_gbonly = 1<<11,
PSC1encolonbp = 1<<15, /* "collision during back-pressure mib counting" */
PSC1coldomlimmask= MASK(6)<<16,
#define PSC1coldomlim(v) (((v) & MASK(6))<<16)
PSC1miiallowoddpreamble = 1<<22,
/* port status 1, ps1 bits */
PS1rxpause = 1<<0,
PS1txpause = 1<<1,
PS1pressure = 1<<2,
PS1syncfail10ms = 1<<3,
PS1an_done = 1<<4,
PS1inbandan_bypassed = 1<<5,
PS1serdesplllocked = 1<<6,
PS1syncok = 1<<7,
PS1nosquelch = 1<<8,
/* irq bits */
/* rx buf returned to cpu ownership, or frame reception finished */
Irx = 1<<0,
Iextend = 1<<1, /* IEsum of irqe set */
#define Irxbufferq(q) (1<<((q)+2)) /* rx buf returned to cpu ownership */
Irxerr = 1<<10, /* input ring full, usually */
#define Irxerrq(q) (1<<((q)+11))
#define Itxendq(q) (1<<((q)+19)) /* tx dma stopped for q */
Isum = 1<<31,
/* irq extended, irqe bits */
#define IEtxbufferq(q) (1<<((q)+0)) /* tx buf returned to cpu ownership */
#define IEtxerrq(q) (1<<((q)+8))
IEphystschg = 1<<16,
IEptp = 1<<17,
IErxoverrun = 1<<18,
IEtxunderrun = 1<<19,
IElinkchg = 1<<20,
IEintaddrerr = 1<<23,
IEprbserr = 1<<25,
IEsum = 1<<31,
/* tx fifo urgent threshold (tx interrupt coalescing), pxtfut */
#define TFUTipginttx(v) (((v) & MASK(16))<<4);
/* minimal frame size, mfs */
MFS40by = 10<<2,
MFS44by = 11<<2,
MFS48by = 12<<2,
MFS52by = 13<<2,
MFS56by = 14<<2,
MFS60by = 15<<2,
MFS64by = 16<<2,
/* receive descriptor status */
RCSmacerr = 1<<0,
RCSmacmask = 3<<1,
RCSmacce = 0<<1,
RCSmacor = 1<<1,
RCSmacmf = 2<<1,
RCSl4chkshift = 3,
RCSl4chkmask = MASK(16),
RCSvlan = 1<<17,
RCSbpdu = 1<<18,
RCSl4mask = 3<<21,
RCSl4tcp4 = 0<<21,
RCSl4udp4 = 1<<21,
RCSl4other = 2<<21,
RCSl4rsvd = 3<<21,
RCSl2ev2 = 1<<23,
RCSl3ip4 = 1<<24,
RCSip4headok = 1<<25,
RCSlast = 1<<26,
RCSfirst = 1<<27,
RCSunknownaddr = 1<<28,
RCSenableintr = 1<<29,
RCSl4chkok = 1<<30,
RCSdmaown = 1<<31,
/* transmit descriptor status */
TCSmacerr = 1<<0,
TCSmacmask = 3<<1,
TCSmaclc = 0<<1,
TCSmacur = 1<<1,
TCSmacrl = 2<<1,
TCSllc = 1<<9,
TCSl4chkmode = 1<<10,
TCSipv4hdlenshift= 11,
TCSvlan = 1<<15,
TCSl4type = 1<<16,
TCSgl4chk = 1<<17,
TCSgip4chk = 1<<18,
TCSpadding = 1<<19,
TCSlast = 1<<20,
TCSfirst = 1<<21,
TCSenableintr = 1<<23,
TCSautomode = 1<<30,
TCSdmaown = 1<<31,
};
enum {
/* SMI regs */
PhysmiTimeout = 10000, /* what units? in ms. */
Physmidataoff = 0, /* Data */
Physmidatamask = 0xffff<<Physmidataoff,
Physmiaddroff = 16, /* PHY device addr */
Physmiaddrmask = 0x1f << Physmiaddroff,
Physmiop = 26,
Physmiopmask = 3<<Physmiop,
PhysmiopWr = 0<<Physmiop,
PhysmiopRd = 1<<Physmiop,
PhysmiReadok = 1<<27,
PhysmiBusy = 1<<28,
SmiRegaddroff = 21, /* PHY device register addr */
SmiRegaddrmask = 0x1f << SmiRegaddroff,
};
struct Gbereg {
ulong phy; /* PHY address */
ulong smi; /* serial mgmt. interface */
ulong euda; /* ether default address */
ulong eudid; /* ether default id */
uchar _pad0[0x80-0x10];
/* dma stuff */
ulong euirq; /* interrupt cause */
ulong euirqmask; /* interrupt mask */
uchar _pad1[0x94-0x88];
ulong euea; /* error address */
ulong euiae; /* internal error address */
uchar _pad2[0xb0-0x9c];
ulong euc; /* control */
uchar _pad3[0x200-0xb4];
struct {
ulong base; /* window base */
ulong size; /* window size */
} base[6];
uchar _pad4[0x280-0x230];
ulong harr[4]; /* high address remap */
ulong bare; /* base address enable */
ulong epap; /* port access protect */
uchar _pad5[0x400-0x298];
ulong portcfg; /* port configuration */
ulong portcfgx; /* port config. extend */
ulong mii; /* mii serial parameters */
ulong _pad6;
ulong evlane; /* vlan ether type */
ulong macal; /* mac address low */
ulong macah; /* mac address high */
ulong sdc; /* sdma config. */
ulong dscp[7]; /* ip diff. serv. code point -> pri */
ulong psc0; /* port serial control 0 */
ulong vpt2p; /* vlan priority tag -> pri */
ulong ps0; /* ether port status 0 */
ulong tqc; /* transmit queue command */
ulong psc1; /* port serial control 1 */
ulong ps1; /* ether port status 1 */
ulong mvhdr; /* marvell header */
ulong _pad8[2];
/* interrupts */
ulong irq; /* interrupt cause; some rw0c bits */
ulong irqe; /* " " extended; some rw0c bits */
ulong irqmask; /* interrupt mask (actually enable) */
ulong irqemask; /* " " extended */
ulong _pad9;
ulong pxtfut; /* port tx fifo urgent threshold */
ulong _pad10;
ulong pxmfs; /* port rx minimum frame size */
ulong _pad11;
/*
* # of input frames discarded by addr filtering or lack of resources;
* zeroed upon read.
*/
ulong pxdfc; /* port rx discard frame counter */
ulong pxofc; /* port overrun frame counter */
ulong _pad12[2];
ulong piae; /* port internal address error */
uchar _pad13[0x4bc-0x498];
ulong etherprio; /* ether type priority */
uchar _pad14[0x4dc-0x4c0];
ulong tqfpc; /* tx queue fixed priority config. */
ulong pttbrc; /* port tx token-bucket rate config. */
ulong tqc1; /* tx queue command 1 */
ulong pmtu; /* port maximum transmit unit */
ulong pmtbs; /* port maximum token bucket size */
uchar _pad15[0x600-0x4f0];
struct {
ulong _pad[3];
ulong r; /* phys. addr.: cur. rx desc. ptrs */
} crdp[8];
ulong rqc; /* rx queue command */
ulong tcsdp; /* phys. addr.: cur. tx desc. ptr */
uchar _pad16[0x6c0-0x688];
ulong tcqdp[8]; /* phys. addr.: cur. tx q. desc. ptr */
uchar _pad17[0x700-0x6e0];
struct {
ulong tbctr; /* queue tx token-bucket counter */
ulong tbcfg; /* tx queue token-bucket config. */
ulong acfg; /* tx queue arbiter config. */
ulong _pad;
} tq[8];
ulong pttbc; /* port tx token-bucket counter */
uchar _pad18[0x7a8-0x784];
ulong ipg2; /* tx queue ipg */
ulong _pad19[3];
ulong ipg3;
ulong _pad20;
ulong htlp; /* high token in low packet */
ulong htap; /* high token in async packet */
ulong ltap; /* low token in async packet */
ulong _pad21;
ulong ts; /* tx speed */
uchar _pad22[0x1000-0x7d4];
/* mac mib counters: statistics */
Mibstats;
uchar _pad23[0x1400-0x1080];
/* multicast filtering; each byte: Qno<<1 | Pass */
ulong dfsmt[64]; /* dest addr filter special m'cast table */
ulong dfomt[64]; /* dest addr filter other m'cast table */
/* unicast filtering */
ulong dfut[4]; /* dest addr filter unicast table */
};
static Ctlr *ctlrs[MaxEther];
static uchar zeroea[Eaddrlen];
static void getmibstats(Ctlr *);
static void
rxfreeb(Block *b)
{
/* freeb(b) will have previously decremented b->ref to 0; raise to 1 */
_xinc(&b->ref);
b->wp = b->rp =
(uchar*)((uintptr)(b->lim - Rxblklen) & ~(Bufalign - 1));
assert(((uintptr)b->rp & (Bufalign - 1)) == 0);
b->free = rxfreeb;
ilock(&freeblocks);
b->next = freeblocks.head;
freeblocks.head = b;
iunlock(&freeblocks);
}
static Block *
rxallocb(void)
{
Block *b;
ilock(&freeblocks);
b = freeblocks.head;
if(b != nil) {
freeblocks.head = b->next;
b->next = nil;
b->free = rxfreeb;
}
iunlock(&freeblocks);
return b;
}
static void
rxkick(Ctlr *ctlr)
{
Gbereg *reg = ctlr->reg;
if (reg->crdp[Qno].r == 0)
reg->crdp[Qno].r = PADDR(&ctlr->rx[ctlr->rxhead]);
if ((reg->rqc & 0xff) == 0) /* all queues are stopped? */
reg->rqc = Rxqon(Qno); /* restart */
coherence();
}
static void
txkick(Ctlr *ctlr)
{
Gbereg *reg = ctlr->reg;
if (reg->tcqdp[Qno] == 0)
reg->tcqdp[Qno] = PADDR(&ctlr->tx[ctlr->txhead]);
if ((reg->tqc & 0xff) == 0) /* all q's stopped? */
reg->tqc = Txqon(Qno); /* restart */
coherence();
}
static void
rxreplenish(Ctlr *ctlr)
{
Rx *r;
Block *b;
while(ctlr->rxb[ctlr->rxtail] == nil) {
b = rxallocb();
if(b == nil) {
iprint("#l%d: rxreplenish out of buffers\n",
ctlr->ether->ctlrno);
break;
}
ctlr->rxb[ctlr->rxtail] = b;
/* set up uncached receive descriptor */
r = &ctlr->rx[ctlr->rxtail];
assert(((uintptr)r & (Descralign - 1)) == 0);
r->countsize = ROUNDUP(Rxblklen, 8);
r->buf = PADDR(b->rp);
coherence();
/* and fire */
r->cs = RCSdmaown | RCSenableintr;
coherence();
ctlr->rxtail = NEXT(ctlr->rxtail, Nrx);
}
}
static void
dump(uchar *bp, long max)
{
if (max > 64)
max = 64;
for (; max > 0; max--, bp++)
iprint("%02.2ux ", *bp);
print("...\n");
}
static void
etheractive(Ether *ether)
{
ether->starttime = TK2MS(MACHP(0)->ticks)/1000;
}
static void
ethercheck(Ether *ether)
{
if (ether->starttime != 0 &&
TK2MS(MACHP(0)->ticks)/1000 - ether->starttime > Etherstuck) {
etheractive(ether);
if (ether->ctlrno == 0) /* only complain about main ether */
iprint("#l%d: ethernet stuck\n", ether->ctlrno);
}
}
static void
receive(Ether *ether)
{
int i;
ulong n;
Block *b;
Ctlr *ctlr = ether->ctlr;
Rx *r;
ethercheck(ether);
for (i = Nrx-2; i > 0; i--) {
r = &ctlr->rx[ctlr->rxhead]; /* *r is uncached */
assert(((uintptr)r & (Descralign - 1)) == 0);
if(r->cs & RCSdmaown) /* descriptor busy? */
break;
b = ctlr->rxb[ctlr->rxhead]; /* got input buffer? */
if (b == nil)
panic("ether1116: nil ctlr->rxb[ctlr->rxhead] "
"in receive");
ctlr->rxb[ctlr->rxhead] = nil;
ctlr->rxhead = NEXT(ctlr->rxhead, Nrx);
if((r->cs & (RCSfirst|RCSlast)) != (RCSfirst|RCSlast)) {
ctlr->nofirstlast++; /* partial packet */
freeb(b);
continue;
}
if(r->cs & RCSmacerr) {
freeb(b);
continue;
}
n = r->countsize >> 16; /* TODO includes 2 pad bytes? */
assert(n >= 2 && n < 2048);
/* clear any cached packet or part thereof */
l2cacheuinvse(b->rp, n+2);
cachedinvse(b->rp, n+2);
b->wp = b->rp + n;
/*
* skip hardware padding intended to align ipv4 address
* in memory (mv-s104860-u0 §8.3.4.1)
*/
b->rp += 2;
etheriq(ether, b, 1);
etheractive(ether);
if (i % (Nrx / 2) == 0) {
rxreplenish(ctlr);
rxkick(ctlr);
}
}
rxreplenish(ctlr);
rxkick(ctlr);
}
static void
txreplenish(Ether *ether) /* free transmitted packets */
{
Ctlr *ctlr;
ctlr = ether->ctlr;
while(ctlr->txtail != ctlr->txhead) {
/* ctlr->tx is uncached */
if(ctlr->tx[ctlr->txtail].cs & TCSdmaown)
break;
if(ctlr->txb[ctlr->txtail] == nil)
panic("no block for sent packet?!");
freeb(ctlr->txb[ctlr->txtail]);
ctlr->txb[ctlr->txtail] = nil;
ctlr->txtail = NEXT(ctlr->txtail, Ntx);
etheractive(ether);
}
}
/*
* transmit strategy: fill the output ring as far as possible,
* perhaps leaving a few spare; kick off the output and take
* an interrupt only when the transmit queue is empty.
*/
static void
transmit(Ether *ether)
{
int i, kick, len;
Block *b;
Ctlr *ctlr = ether->ctlr;
Gbereg *reg = ctlr->reg;
Tx *t;
ethercheck(ether);
ilock(ctlr);
txreplenish(ether); /* reap old packets */
/* queue new packets; use at most half the tx descs to avoid livelock */
kick = 0;
for (i = Ntx/2 - 2; i > 0; i--) {
t = &ctlr->tx[ctlr->txhead]; /* *t is uncached */
assert(((uintptr)t & (Descralign - 1)) == 0);
if(t->cs & TCSdmaown) { /* descriptor busy? */
ctlr->txringfull++;
break;
}
b = qget(ether->oq); /* outgoing packet? */
if (b == nil)
break;
len = BLEN(b);
if(len < ether->minmtu || len > ether->maxmtu) {
freeb(b);
continue;
}
ctlr->txb[ctlr->txhead] = b;
/* make sure the whole packet is in memory */
cachedwbse(b->rp, len);
l2cacheuwbse(b->rp, len);
/* set up the transmit descriptor */
t->buf = PADDR(b->rp);
t->countchk = len << 16;
coherence();
/* and fire */
t->cs = TCSpadding | TCSfirst | TCSlast | TCSdmaown |
TCSenableintr;
coherence();
kick++;
ctlr->txhead = NEXT(ctlr->txhead, Ntx);
}
if (kick) {
txkick(ctlr);
reg->irqmask |= Itxendq(Qno);
reg->irqemask |= IEtxerrq(Qno) | IEtxunderrun;
}
iunlock(ctlr);
}
static void
dumprxdescs(Ctlr *ctlr)
{
int i;
Gbereg *reg = ctlr->reg;
iprint("\nrxhead %d rxtail %d; txcdp %#p rxcdp %#p\n",
ctlr->rxhead, ctlr->rxtail, reg->tcqdp[Qno], reg->crdp[Qno].r);
for (i = 0; i < Nrx; i++) {
iprint("rxb %d @ %#p: %#p\n", i, &ctlr->rxb[i], ctlr->rxb[i]);
delay(50);
}
for (i = 0; i < Nrx; i++) {
iprint("rx %d @ %#p: cs %#lux countsize %lud buf %#lux next %#lux\n",
i, &ctlr->rx[i], ctlr->rx[i].cs,
ctlr->rx[i].countsize >> 3, ctlr->rx[i].buf,
ctlr->rx[i].next);
delay(50);
}
delay(1000);
}
static int
gotinput(void* ctlr)
{
return ((Ctlr*)ctlr)->haveinput != 0;
}
/*
* process any packets in the input ring.
* also sum mib stats frequently to avoid the overflow
* mentioned in the errata.
*/
static void
rcvproc(void* arg)
{
Ctlr *ctlr;
Ether *ether;
ether = arg;
ctlr = ether->ctlr;
for(;;){
tsleep(&ctlr->rrendez, gotinput, ctlr, 10*1000);
ilock(ctlr);
getmibstats(ctlr);
if (ctlr->haveinput) {
ctlr->haveinput = 0;
iunlock(ctlr);
receive(ether);
} else
iunlock(ctlr);
}
}
static void
interrupt(Ureg*, void *arg)
{
ulong irq, irqe, handled;
Ether *ether = arg;
Ctlr *ctlr = ether->ctlr;
Gbereg *reg = ctlr->reg;
handled = 0;
irq = reg->irq;
irqe = reg->irqe;
reg->irqe = 0; /* extinguish intr causes */
reg->irq = 0; /* extinguish intr causes */
ethercheck(ether);
if(irq & (Irx | Irxbufferq(Qno))) {
/*
* letting a kproc process the input takes far less real time
* than doing it all at interrupt level.
*/
ctlr->haveinput = 1;
wakeup(&ctlr->rrendez);
irq &= ~(Irx | Irxbufferq(Qno));
handled++;
} else
rxkick(ctlr);
if(irq & Itxendq(Qno)) { /* transmit ring empty? */
reg->irqmask &= ~Itxendq(Qno); /* prevent more interrupts */
reg->irqemask &= ~(IEtxerrq(Qno) | IEtxunderrun);
transmit(ether);
irq &= ~Itxendq(Qno);
handled++;
}
if(irqe & IEsum) {
/*
* IElinkchg appears to only be set when unplugging.
* autonegotiation is likely not done yet, so linkup not valid,
* thus we note the link change here, and check for
* that and autonegotiation done below.
*/
if(irqe & IEphystschg) {
ether->link = (reg->ps0 & PS0linkup) != 0;
ether->linkchg = 1;
}
if(irqe & IEtxerrq(Qno))
ether->oerrs++;
if(irqe & IErxoverrun)
ether->overflows++;
if(irqe & IEtxunderrun)
ctlr->txunderrun++;
if(irqe & (IEphystschg | IEtxerrq(Qno) | IErxoverrun |
IEtxunderrun))
handled++;
}
if (irq & Isum) {
if (irq & Irxerr) { /* nil desc. ptr. or desc. owned by cpu */
ether->buffs++; /* approx. error */
/* if the input ring is full, drain it */
ctlr->haveinput = 1;
wakeup(&ctlr->rrendez);
}
if(irq & (Irxerr | Irxerrq(Qno)))
handled++;
irq &= ~(Irxerr | Irxerrq(Qno));
}
if(ether->linkchg && (reg->ps1 & PS1an_done)) {
handled++;
ether->link = (reg->ps0 & PS0linkup) != 0;
ether->linkchg = 0;
}
ctlr->newintrs++;
if (!handled) {
irq &= ~Isum;
irqe &= ~IEtxbufferq(Qno);
if (irq == 0 && irqe == 0) {
/* seems to be triggered by continuous output */
// iprint("ether1116: spurious interrupt\n");
} else
iprint("ether1116: interrupt cause unknown; "
"irq %#lux irqe %#lux\n", irq, irqe);
}
intrclear(Irqlo, ether->irq);
}
void
promiscuous(void *arg, int on)
{
Ether *ether = arg;
Ctlr *ctlr = ether->ctlr;
Gbereg *reg = ctlr->reg;
ilock(ctlr);
ether->prom = on;
if(on)
reg->portcfg |= PCFGupromisc;
else
reg->portcfg &= ~PCFGupromisc;
iunlock(ctlr);
}
void
multicast(void *, uchar *, int)
{
/* nothing to do; we always accept multicast */
}
static void quiesce(Gbereg *reg);
static void
shutdown(Ether *ether)
{
int i;
Ctlr *ctlr = ether->ctlr;
Gbereg *reg = ctlr->reg;
ilock(ctlr);
quiesce(reg);
reg->euc |= Portreset;
coherence();
iunlock(ctlr);
delay(100);
ilock(ctlr);
reg->euc &= ~Portreset;
coherence();
delay(20);
reg->psc0 = 0; /* no PSC0porton */
reg->psc1 |= PSC1portreset;
coherence();
delay(50);
reg->psc1 &= ~PSC1portreset;
coherence();
for (i = 0; i < nelem(reg->tcqdp); i++)
reg->tcqdp[i] = 0;
for (i = 0; i < nelem(reg->crdp); i++)
reg->crdp[i].r = 0;
coherence();
iunlock(ctlr);
}
enum {
CMjumbo,
};
static Cmdtab ctlmsg[] = {
CMjumbo, "jumbo", 2,
};
long
ctl(Ether *e, void *p, long n)
{
Cmdbuf *cb;
Cmdtab *ct;
Ctlr *ctlr = e->ctlr;
Gbereg *reg = ctlr->reg;
cb = parsecmd(p, n);
if(waserror()) {
free(cb);
nexterror();
}
ct = lookupcmd(cb, ctlmsg, nelem(ctlmsg));
switch(ct->index) {
case CMjumbo:
if(strcmp(cb->f[1], "on") == 0) {
/* incoming packet queue doesn't expect jumbo frames */
error("jumbo disabled");
reg->psc0 = (reg->psc0 & ~PSC0mrumask) |
PSC0mru(PSC0mru9022);
e->maxmtu = 9022;
} else if(strcmp(cb->f[1], "off") == 0) {
reg->psc0 = (reg->psc0 & ~PSC0mrumask) |
PSC0mru(PSC0mru1522);
e->maxmtu = ETHERMAXTU;
} else
error(Ebadctl);
break;
default:
error(Ebadctl);
break;
}
free(cb);
poperror();
return n;
}
/*
* phy/mii goo
*/
static int
smibusywait(Gbereg *reg, ulong waitbit)
{
ulong timeout, smi_reg;
timeout = PhysmiTimeout;
/* wait till the SMI is not busy */
do {
/* read smi register */
smi_reg = reg->smi;
if (timeout-- == 0) {
MIIDBG("SMI busy timeout\n");
return -1;
}
// delay(1);
} while (smi_reg & waitbit);
return 0;
}
static int
miird(Mii *mii, int pa, int ra)
{
ulong smi_reg, timeout;
Gbereg *reg;
reg = ((Ctlr*)mii->ctlr)->reg;
/* check params */
if ((pa<<Physmiaddroff) & ~Physmiaddrmask ||
(ra<<SmiRegaddroff) & ~SmiRegaddrmask)
return -1;
smibusywait(reg, PhysmiBusy);
/* fill the phy address and register offset and read opcode */
reg->smi = pa << Physmiaddroff | ra << SmiRegaddroff | PhysmiopRd;
coherence();
/* wait til read value is ready */
timeout = PhysmiTimeout;
do {
smi_reg = reg->smi;
if (timeout-- == 0) {
MIIDBG("SMI read-valid timeout\n");
return -1;
}
} while (!(smi_reg & PhysmiReadok));
/* Wait for the data to update in the SMI register */
for (timeout = 0; timeout < PhysmiTimeout; timeout++)
;
return reg->smi & Physmidatamask;
}
static int
miiwr(Mii *mii, int pa, int ra, int v)
{
Gbereg *reg;
ulong smi_reg;
reg = ((Ctlr*)mii->ctlr)->reg;
/* check params */
if (((pa<<Physmiaddroff) & ~Physmiaddrmask) ||
((ra<<SmiRegaddroff) & ~SmiRegaddrmask))
return -1;
smibusywait(reg, PhysmiBusy);
/* fill the phy address and register offset and read opcode */
smi_reg = v << Physmidataoff | pa << Physmiaddroff | ra << SmiRegaddroff;
reg->smi = smi_reg & ~PhysmiopRd;
coherence();
return 0;
}
#define MIIMODEL(idr2) (((idr2) >> 4) & MASK(6))
enum {
Hacknone,
Hackdual,
Ouimarvell = 0x005043,
/* idr2 mii/phy model numbers */
Phy1000 = 0x00, /* 88E1000 Gb */
Phy1011 = 0x02, /* 88E1011 Gb */
Phy1000_3 = 0x03, /* 88E1000 Gb */
Phy1000s = 0x04, /* 88E1000S Gb */
Phy1000_5 = 0x05, /* 88E1000 Gb */
Phy1000_6 = 0x06, /* 88E1000 Gb */
Phy3082 = 0x08, /* 88E3082 10/100 */
Phy1112 = 0x09, /* 88E1112 Gb */
Phy1121r = 0x0b, /* says the 1121r manual */
Phy1149 = 0x0b, /* 88E1149 Gb */
Phy1111 = 0x0c, /* 88E1111 Gb */
Phy1116 = 0x21, /* 88E1116 Gb */
Phy1116r = 0x24, /* 88E1116R Gb */
Phy1118 = 0x22, /* 88E1118 Gb */
Phy3016 = 0x26, /* 88E3016 10/100 */
};
static int hackflavour;
/*
* on openrd, ether0's phy has address 8, ether1's is ether0's 24.
* on guruplug, ether0's is phy 0 and ether1's is ether0's phy 1.
*/
int
mymii(Mii* mii, int mask)
{
Ctlr *ctlr;
MiiPhy *miiphy;
int bit, ctlrno, oui, model, phyno, r, rmask;
static int dualport, phyidx;
static int phynos[NMiiPhy];
ctlr = mii->ctlr;
ctlrno = ctlr->ether->ctlrno;
/* first pass: figure out what kind of phy(s) we have. */
dualport = 0;
if (ctlrno == 0) {
for(phyno = 0; phyno < NMiiPhy; phyno++){
bit = 1<<phyno;
if(!(mask & bit) || mii->mask & bit)
continue;
if(mii->mir(mii, phyno, Bmsr) == -1)
continue;
r = mii->mir(mii, phyno, Phyidr1);
oui = (r & 0x3FFF)<<6;
r = mii->mir(mii, phyno, Phyidr2);
oui |= r>>10;
model = MIIMODEL(r);
if (oui == 0xfffff && model == 0x3f)
continue;
MIIDBG("ctlrno %d phy %d oui %#ux model %#ux\n",
ctlrno, phyno, oui, model);
if (oui == Ouimarvell &&
(model == Phy1121r || model == Phy1116r))
++dualport;
phynos[phyidx++] = phyno;
}
hackflavour = dualport == 2 && phyidx == 2? Hackdual: Hacknone;
MIIDBG("ether1116: %s-port phy\n",
hackflavour == Hackdual? "dual": "single");
}
/*
* Probe through mii for PHYs in mask;
* return the mask of those found in the current probe.
* If the PHY has not already been probed, update
* the Mii information.
*/
rmask = 0;
if (hackflavour == Hackdual && ctlrno < phyidx) {
/*
* openrd, guruplug or the like: use ether0's phys.
* this is a nasty hack, but so is the hardware.
*/
MIIDBG("ctlrno %d using ctlrno 0's phyno %d\n",
ctlrno, phynos[ctlrno]);
ctlr->mii = mii = ctlrs[0]->mii;
mask = 1 << phynos[ctlrno];
mii->mask = ~mask;
}
for(phyno = 0; phyno < NMiiPhy; phyno++){
bit = 1<<phyno;
if(!(mask & bit))
continue;
if(mii->mask & bit){
rmask |= bit;
continue;
}
if(mii->mir(mii, phyno, Bmsr) == -1)
continue;
r = mii->mir(mii, phyno, Phyidr1);
oui = (r & 0x3FFF)<<6;
r = mii->mir(mii, phyno, Phyidr2);
oui |= r>>10;
if(oui == 0xFFFFF || oui == 0)
continue;
if((miiphy = malloc(sizeof(MiiPhy))) == nil)
continue;
miiphy->mii = mii;
miiphy->oui = oui;
miiphy->phyno = phyno;
miiphy->anar = ~0;
miiphy->fc = ~0;
miiphy->mscr = ~0;
mii->phy[phyno] = miiphy;
if(ctlrno == 0 || hackflavour != Hackdual && mii->curphy == nil)
mii->curphy = miiphy;
mii->mask |= bit;
mii->nphy++;
rmask |= bit;
}
return rmask;
}
static int
kirkwoodmii(Ether *ether)
{
int i;
Ctlr *ctlr;
MiiPhy *phy;
MIIDBG("mii\n");
ctlr = ether->ctlr;
if((ctlr->mii = malloc(sizeof(Mii))) == nil)
return -1;
ctlr->mii->ctlr = ctlr;
ctlr->mii->mir = miird;
ctlr->mii->miw = miiwr;
if(mymii(ctlr->mii, ~0) == 0 || (phy = ctlr->mii->curphy) == nil){
print("#l%d: ether1116: init mii failure\n", ether->ctlrno);
free(ctlr->mii);
ctlr->mii = nil;
return -1;
}
/* oui 005043 is marvell */
MIIDBG("oui %#X phyno %d\n", phy->oui, phy->phyno);
// TODO: does this make sense? shouldn't each phy be initialised?
if((ctlr->ether->ctlrno == 0 || hackflavour != Hackdual) &&
miistatus(ctlr->mii) < 0){
miireset(ctlr->mii);
MIIDBG("miireset\n");
if(miiane(ctlr->mii, ~0, 0, ~0) < 0){
iprint("miiane failed\n");
return -1;
}
MIIDBG("miistatus\n");
miistatus(ctlr->mii);
if(miird(ctlr->mii, phy->phyno, Bmsr) & BmsrLs){
for(i = 0; ; i++){
if(i > 600){
iprint("ether1116: autonegotiation failed\n");
break;
}
if(miird(ctlr->mii, phy->phyno, Bmsr) & BmsrAnc)
break;
delay(10);
}
if(miistatus(ctlr->mii) < 0)
iprint("miistatus failed\n");
}else{
iprint("ether1116: no link\n");
phy->speed = 10; /* simple default */
}
}
ether->mbps = phy->speed;
MIIDBG("#l%d: kirkwoodmii: fd %d speed %d tfc %d rfc %d\n",
ctlr->port, phy->fd, phy->speed, phy->tfc, phy->rfc);
MIIDBG("mii done\n");
return 0;
}
enum { /* PHY register pages */
Pagcopper,
Pagfiber,
Pagrgmii,
Pagled,
Pagrsvd1,
Pagvct,
Pagtest,
Pagrsvd2,
Pagfactest,
};
static void
miiregpage(Mii *mii, ulong dev, ulong page)
{
miiwr(mii, dev, Eadr, page);
}
static int
miiphyinit(Mii *mii)
{
ulong dev;
Ctlr *ctlr;
Gbereg *reg;
ctlr = (Ctlr*)mii->ctlr;
reg = ctlr->reg;
dev = reg->phy;
MIIDBG("phy dev addr %lux\n", dev);
/* leds link & activity */
miiregpage(mii, dev, Pagled);
/* low 4 bits == 1: on - link, blink - activity, off - no link */
miiwr(mii, dev, Scr, (miird(mii, dev, Scr) & ~0xf) | 1);
miiregpage(mii, dev, Pagrgmii);
miiwr(mii, dev, Scr, miird(mii, dev, Scr) | Rgmiipwrup);
/* must now do a software reset, says the manual */
miireset(ctlr->mii);
/* enable RGMII delay on Tx and Rx for CPU port */
miiwr(mii, dev, Recr, miird(mii, dev, Recr) | Rxtiming | Rxtiming);
/* must now do a software reset, says the manual */
miireset(ctlr->mii);
miiregpage(mii, dev, Pagcopper);
miiwr(mii, dev, Scr,
(miird(mii, dev, Scr) & ~(Pwrdown|Endetect)) | Mdix);
return 0;
}
/*
* initialisation
*/
static void
quiesce(Gbereg *reg)
{
ulong v;
v = reg->tqc;
if (v & 0xFF)
reg->tqc = v << 8; /* stop active channels */
v = reg->rqc;
if (v & 0xFF)
reg->rqc = v << 8; /* stop active channels */
/* wait for all queues to stop */
while (reg->tqc & 0xFF || reg->rqc & 0xFF)
;
}
static void
p16(uchar *p, ulong v) /* convert big-endian short to bytes */
{
*p++ = v>>8;
*p = v;
}
static void
p32(uchar *p, ulong v) /* convert big-endian long to bytes */
{
*p++ = v>>24;
*p++ = v>>16;
*p++ = v>>8;
*p = v;
}
/*
* set ether->ea from hw mac address,
* configure unicast filtering to accept it.
*/
void
archetheraddr(Ether *ether, Gbereg *reg, int rxqno)
{
uchar *ea;
ulong nibble, ucreg, tbloff, regoff;
ea = ether->ea;
p32(ea, reg->macah);
p16(ea+4, reg->macal);
if (memcmp(ea, zeroea, sizeof zeroea) == 0 && ether->ctlrno > 0) {
/* hack: use ctlr[0]'s + ctlrno */
memmove(ea, ctlrs[0]->ether->ea, Eaddrlen);
ea[Eaddrlen-1] += ether->ctlrno;
reg->macah = ea[0] << 24 | ea[1] << 16 | ea[2] << 8 | ea[3];
reg->macal = ea[4] << 8 | ea[5];
coherence();
}
/* accept frames on ea */
nibble = ea[5] & 0xf;
tbloff = nibble / 4;
regoff = nibble % 4;
regoff *= 8;
ucreg = reg->dfut[tbloff] & (0xff << regoff);
ucreg |= (rxqno << 1 | Pass) << regoff;
reg->dfut[tbloff] = ucreg;
/* accept all multicast too. set up special & other tables. */
memset(reg->dfsmt, Qno<<1 | Pass, sizeof reg->dfsmt);
memset(reg->dfomt, Qno<<1 | Pass, sizeof reg->dfomt);
coherence();
}
static void
cfgdramacc(Gbereg *reg)
{
memset(reg->harr, 0, sizeof reg->harr);
memset(reg->base, 0, sizeof reg->base);
reg->bare = MASK(6) - MASK(2); /* disable wins 2-5 */
/* this doesn't make any sense, but it's required */
reg->epap = 3 << 2 | 3; /* full access for wins 0 & 1 */
// reg->epap = 0; /* no access on access violation for all wins */
coherence();
reg->base[0].base = PHYSDRAM | WINATTR(Attrcs0) | Targdram;
reg->base[0].size = WINSIZE(256*MB);
reg->base[1].base = (PHYSDRAM + 256*MB) | WINATTR(Attrcs1) | Targdram;
reg->base[1].size = WINSIZE(256*MB);
coherence();
}
static void
ctlralloc(Ctlr *ctlr)
{
int i;
Block *b;
Rx *r;
Tx *t;
ilock(&freeblocks);
for(i = 0; i < Nrxblks; i++) {
b = iallocb(Rxblklen+Bufalign-1);
if(b == nil) {
iprint("ether1116: no memory for rx buffers\n");
break;
}
assert(b->ref == 1);
b->wp = b->rp = (uchar*)
((uintptr)(b->lim - Rxblklen) & ~(Bufalign - 1));
assert(((uintptr)b->rp & (Bufalign - 1)) == 0);
b->free = rxfreeb;
b->next = freeblocks.head;
freeblocks.head = b;
}
iunlock(&freeblocks);
/*
* allocate uncached rx ring descriptors because rings are shared
* with the ethernet controller and more than one fits in a cache line.
*/
ctlr->rx = ucallocalign(Nrx * sizeof(Rx), Descralign, 0);
if(ctlr->rx == nil)
panic("ether1116: no memory for rx ring");
for(i = 0; i < Nrx; i++) {
r = &ctlr->rx[i];
assert(((uintptr)r & (Descralign - 1)) == 0);
r->cs = 0; /* owned by software until r->buf is non-nil */
r->buf = 0;
r->next = PADDR(&ctlr->rx[NEXT(i, Nrx)]);
ctlr->rxb[i] = nil;
}
ctlr->rxtail = ctlr->rxhead = 0;
rxreplenish(ctlr);
/* allocate uncached tx ring descriptors */
ctlr->tx = ucallocalign(Ntx * sizeof(Tx), Descralign, 0);
if(ctlr->tx == nil)
panic("ether1116: no memory for tx ring");
for(i = 0; i < Ntx; i++) {
t = &ctlr->tx[i];
assert(((uintptr)t & (Descralign - 1)) == 0);
t->cs = 0;
t->buf = 0;
t->next = PADDR(&ctlr->tx[NEXT(i, Ntx)]);
ctlr->txb[i] = nil;
}
ctlr->txtail = ctlr->txhead = 0;
}
static void
ctlrinit(Ether *ether)
{
int i;
Ctlr *ctlr = ether->ctlr;
Gbereg *reg = ctlr->reg;
static char name[KNAMELEN];
static Ctlr fakectlr; /* bigger than 4K; keep off the stack */
for (i = 0; i < nelem(reg->tcqdp); i++)
reg->tcqdp[i] = 0;
for (i = 0; i < nelem(reg->crdp); i++)
reg->crdp[i].r = 0;
coherence();
cfgdramacc(reg);
ctlralloc(ctlr);
reg->tcqdp[Qno] = PADDR(&ctlr->tx[ctlr->txhead]);
reg->crdp[Qno].r = PADDR(&ctlr->rx[ctlr->rxhead]);
coherence();
// dumprxdescs(ctlr);
/* clear stats by reading them into fake ctlr */
getmibstats(&fakectlr);
reg->pxmfs = MFS40by; /* allow runts in */
/*
* ipg's (inter packet gaps) for interrupt coalescing,
* values in units of 64 clock cycles. A full-sized
* packet (1514 bytes) takes just over 12µs to transmit.
*/
if (CLOCKFREQ/(Maxrxintrsec*64) >= (1<<16))
panic("rx coalescing value %d too big for short",
CLOCKFREQ/(Maxrxintrsec*64));
reg->sdc = SDCrifb | SDCrxburst(Burst16) | SDCtxburst(Burst16) |
SDCrxnobyteswap | SDCtxnobyteswap |
SDCipgintrx(CLOCKFREQ/(Maxrxintrsec*64));
reg->pxtfut = 0; /* TFUTipginttx(CLOCKFREQ/(Maxrxintrsec*64)) */
/* allow just these interrupts */
/* guruplug generates Irxerr interrupts continually */
reg->irqmask = Isum | Irx | Irxbufferq(Qno) | Irxerr | Itxendq(Qno);
reg->irqemask = IEsum | IEtxerrq(Qno) | IEphystschg | IErxoverrun |
IEtxunderrun;
reg->irqe = 0;
reg->euirqmask = 0;
coherence();
reg->irq = 0;
reg->euirq = 0;
/* send errors to end of memory */
// reg->euda = PHYSDRAM + 512*MB - 8*1024;
reg->euda = 0;
reg->eudid = Attrcs1 << 4 | Targdram;
// archetheraddr(ether, ctlr->reg, Qno); /* 2nd location */
reg->portcfg = Rxqdefault(Qno) | Rxqarp(Qno);
reg->portcfgx = 0;
coherence();
/*
* start the controller running.
* turn the port on, kick the receiver.
*/
reg->psc1 = PSC1rgmii | PSC1encolonbp | PSC1coldomlim(0x23);
/* do this only when the controller is quiescent */
reg->psc0 = PSC0porton | PSC0an_flctloff |
PSC0an_pauseadv | PSC0nofrclinkdown | PSC0mru(PSC0mru1522);
coherence();
for (i = 0; i < 4000; i++) /* magic delay */
;
ether->link = (reg->ps0 & PS0linkup) != 0;
/* set ethernet MTU for leaky bucket mechanism to 0 (disabled) */
reg->pmtu = 0;
etheractive(ether);
snprint(name, sizeof name, "#l%drproc", ether->ctlrno);
kproc(name, rcvproc, ether);
reg->rqc = Rxqon(Qno);
coherence();
}
static void
attach(Ether* ether)
{
Ctlr *ctlr = ether->ctlr;
lock(&ctlr->initlock);
if(ctlr->init == 0) {
ctlrinit(ether);
ctlr->init = 1;
}
unlock(&ctlr->initlock);
}
/*
* statistics goo.
* mib registers clear on read.
*/
static void
getmibstats(Ctlr *ctlr)
{
Gbereg *reg = ctlr->reg;
/*
* Marvell 88f6281 errata FE-ETH-120: high long of rxby and txby
* can't be read correctly, so read the low long frequently
* (every 30 seconds or less), thus avoiding overflow into high long.
*/
ctlr->rxby += reg->rxbylo;
ctlr->txby += reg->txbylo;
ctlr->badrxby += reg->badrxby;
ctlr->mactxerr += reg->mactxerr;
ctlr->rxpkt += reg->rxpkt;
ctlr->badrxpkt += reg->badrxpkt;
ctlr->rxbcastpkt+= reg->rxbcastpkt;
ctlr->rxmcastpkt+= reg->rxmcastpkt;
ctlr->rx64 += reg->rx64;
ctlr->rx65_127 += reg->rx65_127;
ctlr->rx128_255 += reg->rx128_255;
ctlr->rx256_511 += reg->rx256_511;
ctlr->rx512_1023+= reg->rx512_1023;
ctlr->rx1024_max+= reg->rx1024_max;
ctlr->txpkt += reg->txpkt;
ctlr->txcollpktdrop+= reg->txcollpktdrop;
ctlr->txmcastpkt+= reg->txmcastpkt;
ctlr->txbcastpkt+= reg->txbcastpkt;
ctlr->badmacctlpkts+= reg->badmacctlpkts;
ctlr->txflctl += reg->txflctl;
ctlr->rxflctl += reg->rxflctl;
ctlr->badrxflctl+= reg->badrxflctl;
ctlr->rxundersized+= reg->rxundersized;
ctlr->rxfrags += reg->rxfrags;
ctlr->rxtoobig += reg->rxtoobig;
ctlr->rxjabber += reg->rxjabber;
ctlr->rxerr += reg->rxerr;
ctlr->crcerr += reg->crcerr;
ctlr->collisions+= reg->collisions;
ctlr->latecoll += reg->latecoll;
}
long
ifstat(Ether *ether, void *a, long n, ulong off)
{
Ctlr *ctlr = ether->ctlr;
Gbereg *reg = ctlr->reg;
char *buf, *p, *e;
buf = p = malloc(READSTR);
if(p == nil)
panic("ether1116 ifstat: no memory");
e = p + READSTR;
ilock(ctlr);
getmibstats(ctlr);
ctlr->intrs += ctlr->newintrs;
p = seprint(p, e, "interrupts: %lud\n", ctlr->intrs);
p = seprint(p, e, "new interrupts: %lud\n", ctlr->newintrs);
ctlr->newintrs = 0;
p = seprint(p, e, "tx underrun: %lud\n", ctlr->txunderrun);
p = seprint(p, e, "tx ring full: %lud\n", ctlr->txringfull);
ctlr->rxdiscard += reg->pxdfc;
ctlr->rxoverrun += reg->pxofc;
p = seprint(p, e, "rx discarded frames: %lud\n", ctlr->rxdiscard);
p = seprint(p, e, "rx overrun frames: %lud\n", ctlr->rxoverrun);
p = seprint(p, e, "no first+last flag: %lud\n", ctlr->nofirstlast);
p = seprint(p, e, "duplex: %s\n", (reg->ps0 & PS0fd)? "full": "half");
p = seprint(p, e, "flow control: %s\n", (reg->ps0 & PS0flctl)? "on": "off");
/* p = seprint(p, e, "speed: %d mbps\n", ); */
p = seprint(p, e, "received bytes: %llud\n", ctlr->rxby);
p = seprint(p, e, "bad received bytes: %lud\n", ctlr->badrxby);
p = seprint(p, e, "internal mac transmit errors: %lud\n", ctlr->mactxerr);
p = seprint(p, e, "total received frames: %lud\n", ctlr->rxpkt);
p = seprint(p, e, "received broadcast frames: %lud\n", ctlr->rxbcastpkt);
p = seprint(p, e, "received multicast frames: %lud\n", ctlr->rxmcastpkt);
p = seprint(p, e, "bad received frames: %lud\n", ctlr->badrxpkt);
p = seprint(p, e, "received frames 0-64: %lud\n", ctlr->rx64);
p = seprint(p, e, "received frames 65-127: %lud\n", ctlr->rx65_127);
p = seprint(p, e, "received frames 128-255: %lud\n", ctlr->rx128_255);
p = seprint(p, e, "received frames 256-511: %lud\n", ctlr->rx256_511);
p = seprint(p, e, "received frames 512-1023: %lud\n", ctlr->rx512_1023);
p = seprint(p, e, "received frames 1024-max: %lud\n", ctlr->rx1024_max);
p = seprint(p, e, "transmitted bytes: %llud\n", ctlr->txby);
p = seprint(p, e, "total transmitted frames: %lud\n", ctlr->txpkt);
p = seprint(p, e, "transmitted broadcast frames: %lud\n", ctlr->txbcastpkt);
p = seprint(p, e, "transmitted multicast frames: %lud\n", ctlr->txmcastpkt);
p = seprint(p, e, "transmit frames dropped by collision: %lud\n", ctlr->txcollpktdrop);
p = seprint(p, e, "misaligned buffers: %lud\n", ether->pktsmisaligned);
p = seprint(p, e, "bad mac control frames: %lud\n", ctlr->badmacctlpkts);
p = seprint(p, e, "transmitted flow control messages: %lud\n", ctlr->txflctl);
p = seprint(p, e, "received flow control messages: %lud\n", ctlr->rxflctl);
p = seprint(p, e, "bad received flow control messages: %lud\n", ctlr->badrxflctl);
p = seprint(p, e, "received undersized packets: %lud\n", ctlr->rxundersized);
p = seprint(p, e, "received fragments: %lud\n", ctlr->rxfrags);
p = seprint(p, e, "received oversized packets: %lud\n", ctlr->rxtoobig);
p = seprint(p, e, "received jabber packets: %lud\n", ctlr->rxjabber);
p = seprint(p, e, "mac receive errors: %lud\n", ctlr->rxerr);
p = seprint(p, e, "crc errors: %lud\n", ctlr->crcerr);
p = seprint(p, e, "collisions: %lud\n", ctlr->collisions);
p = seprint(p, e, "late collisions: %lud\n", ctlr->latecoll);
USED(p);
iunlock(ctlr);
n = readstr(off, a, n, buf);
free(buf);
return n;
}
static int
reset(Ether *ether)
{
Ctlr *ctlr;
ether->ctlr = ctlr = malloc(sizeof *ctlr);
if (ctlr == nil)
panic("ether1116 reset: no memory");
switch(ether->ctlrno) {
case 0:
ether->irq = IRQ0gbe0sum;
break;
case 1:
ether->irq = IRQ0gbe1sum;
break;
default:
panic("ether1116: bad ether ctlr #%d", ether->ctlrno);
}
ctlr->reg = (Gbereg*)soc.ether[ether->ctlrno];
/* need this for guruplug, at least */
*(ulong *)soc.iocfg |= 1 << 7 | 1 << 15; /* io cfg 0: 1.8v gbe */
coherence();
ctlr->ether = ether;
ctlrs[ether->ctlrno] = ctlr;
shutdown(ether);
/* ensure that both interfaces are set to RGMII before calling mii */
((Gbereg*)soc.ether[0])->psc1 |= PSC1rgmii;
((Gbereg*)soc.ether[1])->psc1 |= PSC1rgmii;
coherence();
/* Set phy address of the port */
ctlr->port = ether->ctlrno;
ctlr->reg->phy = ether->ctlrno;
coherence();
ether->port = (uintptr)ctlr->reg;
if(kirkwoodmii(ether) < 0){
free(ctlr);
ether->ctlr = nil;
return -1;
}
miiphyinit(ctlr->mii);
archetheraddr(ether, ctlr->reg, Qno); /* original location */
if (memcmp(ether->ea, zeroea, sizeof zeroea) == 0){
iprint("ether1116: reset: zero ether->ea\n");
free(ctlr);
ether->ctlr = nil;
return -1; /* no rj45 for this ether */
}
ether->attach = attach;
ether->transmit = transmit;
ether->interrupt = interrupt;
ether->ifstat = ifstat;
ether->shutdown = shutdown;
ether->ctl = ctl;
ether->arg = ether;
ether->promiscuous = promiscuous;
ether->multicast = multicast;
return 0;
}
void
ether1116link(void)
{
addethercard("88e1116", reset);
}
|