From 9d685a85978abb46c26ff6838837c90b065f75d1 Mon Sep 17 00:00:00 2001 From: Gabor Juhos Date: Mon, 15 Oct 2007 06:22:34 +0000 Subject: [PATCH] switch driver cleanup, 1st phase SVN-Revision: 9324 --- .../adm5120/files/drivers/net/adm5120sw.c | 895 +++++++++++------- .../adm5120/files/drivers/net/adm5120sw.h | 87 +- .../asm-mips/mach-adm5120/adm5120_switch.h | 26 + 3 files changed, 598 insertions(+), 410 deletions(-) diff --git a/target/linux/adm5120/files/drivers/net/adm5120sw.c b/target/linux/adm5120/files/drivers/net/adm5120sw.c index 80c81abf74..5acdf393bd 100644 --- a/target/linux/adm5120/files/drivers/net/adm5120sw.c +++ b/target/linux/adm5120/files/drivers/net/adm5120sw.c @@ -12,90 +12,286 @@ * TODO: Add support of high prio queues (currently disabled) * */ -#include -#include #include -#include -#include -#include +#include #include #include #include +#include + +#include +#include +#include + +#include +#include + #include -#include -#include + +#include +#include +#include +#include + #include "adm5120sw.h" -#include -#include +#define DRV_NAME "adm5120-switch" +#define DRV_DESC "ADM5120 built-in ethernet switch driver" +#define DRV_VERSION "0.1.0" MODULE_AUTHOR("Jeroen Vreeken (pe1rxq@amsat.org)"); MODULE_DESCRIPTION("ADM5120 ethernet switch driver"); MODULE_LICENSE("GPL"); +/* ------------------------------------------------------------------------ */ + +#if 1 /*def ADM5120_SWITCH_DEBUG*/ +#define SW_DBG(f, a...) printk(KERN_DEBUG "%s: " f, DRV_NAME , ## a) +#else +#define SW_DBG(f, a...) do {} while (0) +#endif +#define SW_ERR(f, a...) printk(KERN_ERR "%s: " f, DRV_NAME , ## a) +#define SW_INFO(f, a...) printk(KERN_INFO "%s: " f, DRV_NAME , ## a) + +#define SWITCH_NUM_PORTS 6 +#define ETH_CSUM_LEN 4 + +#define RX_MAX_PKTLEN 1550 +#define RX_RING_SIZE 64 + +#define TX_RING_SIZE 32 +#define TX_QUEUE_LEN 28 /* Limit ring entries actually used. */ +#define TX_TIMEOUT HZ*400 + +#define SKB_ALLOC_LEN (RX_MAX_PKTLEN + 32) +#define SKB_RESERVE_LEN (NET_IP_ALIGN + NET_SKB_PAD) + +#define SWITCH_INTS_HIGH (SWITCH_INT_SHD | SWITCH_INT_RHD | SWITCH_INT_HDF) +#define SWITCH_INTS_LOW (SWITCH_INT_SLD | SWITCH_INT_RLD | SWITCH_INT_LDF) +#define SWITCH_INTS_ERR (SWITCH_INT_RDE | SWITCH_INT_SDE | SWITCH_INT_CPUH) +#define SWITCH_INTS_Q (SWITCH_INT_P0QF | SWITCH_INT_P1QF | SWITCH_INT_P2QF | \ + SWITCH_INT_P3QF | SWITCH_INT_P4QF | SWITCH_INT_P5QF | \ + SWITCH_INT_CPQF | SWITCH_INT_GQF) + +#define SWITCH_INTS_ALL (SWITCH_INTS_HIGH | SWITCH_INTS_LOW | \ + SWITCH_INTS_ERR | SWITCH_INTS_Q | \ + SWITCH_INT_MD | SWITCH_INT_PSC) + +#define SWITCH_INTS_USED (SWITCH_INTS_LOW | SWITCH_INT_PSC) +#define SWITCH_INTS_POLL (SWITCH_INT_RLD | SWITCH_INT_LDF) + +/* ------------------------------------------------------------------------ */ + +struct dma_desc { + __u32 buf1; +#define DESC_OWN (1UL << 31) /* Owned by the switch */ +#define DESC_EOR (1UL << 28) /* End of Ring */ +#define DESC_ADDR_MASK 0x1FFFFFF +#define DESC_ADDR(x) ((__u32)(x) & DESC_ADDR_MASK) + __u32 buf2; +#define DESC_BUF2_EN (1UL << 31) /* Buffer 2 enable */ + __u32 buflen; + __u32 misc; +/* definitions for tx/rx descriptors */ +#define DESC_PKTLEN_SHIFT 16 +#define DESC_PKTLEN_MASK 0x7FF +/* tx descriptor specific part */ +#define DESC_CSUM (1UL << 31) /* Append checksum */ +#define DESC_DSTPORT_SHIFT 8 +#define DESC_DSTPORT_MASK 0x3F +#define DESC_VLAN_MASK 0x3F +/* rx descriptor specific part */ +#define DESC_SRCPORT_SHIFT 12 +#define DESC_SRCPORT_MASK 0x7 +#define DESC_DA_MASK 0x3 +#define DESC_DA_SHIFT 4 +#define DESC_IPCSUM_FAIL (1UL << 3) /* IP checksum fail */ +#define DESC_VLAN_TAG (1UL << 2) /* VLAN tag present */ +#define DESC_TYPE_MASK 0x3 /* mask for Packet type */ +#define DESC_TYPE_IP 0x0 /* IP packet */ +#define DESC_TYPE_PPPoE 0x1 /* PPPoE packet */ +} __attribute__ ((aligned(16))); + +static inline u32 desc_get_srcport(struct dma_desc *desc) +{ + return (desc->misc >> DESC_SRCPORT_SHIFT) & DESC_SRCPORT_MASK; +} + +static inline u32 desc_get_pktlen(struct dma_desc *desc) +{ + return (desc->misc >> DESC_PKTLEN_SHIFT) & DESC_PKTLEN_MASK; +} + +static inline int desc_ipcsum_fail(struct dma_desc *desc) +{ + return ((desc->misc & DESC_IPCSUM_FAIL) != 0); +} + +/* ------------------------------------------------------------------------ */ + /* default settings - unlimited TX and RX on all ports, default shaper mode */ -static unsigned char bw_matrix[SW_DEVS] = { +static unsigned char bw_matrix[SWITCH_NUM_PORTS] = { 0, 0, 0, 0, 0, 0 }; static int adm5120_nrdevs; -static struct net_device *adm5120_devs[SW_DEVS]; +static struct net_device *adm5120_devs[SWITCH_NUM_PORTS]; /* Lookup table port -> device */ -static struct net_device *adm5120_port[SW_DEVS]; +static struct net_device *adm5120_port[SWITCH_NUM_PORTS]; -static struct adm5120_dma - adm5120_dma_txh_v[ADM5120_DMA_TXH] __attribute__((aligned(16))), - adm5120_dma_txl_v[ADM5120_DMA_TXL] __attribute__((aligned(16))), - adm5120_dma_rxh_v[ADM5120_DMA_RXH] __attribute__((aligned(16))), - adm5120_dma_rxl_v[ADM5120_DMA_RXL] __attribute__((aligned(16))), - *adm5120_dma_txh, - *adm5120_dma_txl, - *adm5120_dma_rxh, - *adm5120_dma_rxl; -static struct sk_buff - *adm5120_skb_rxh[ADM5120_DMA_RXH], - *adm5120_skb_rxl[ADM5120_DMA_RXL], - *adm5120_skb_txh[ADM5120_DMA_TXH], - *adm5120_skb_txl[ADM5120_DMA_TXL]; -static int adm5120_rxli = 0; -static int adm5120_txli = 0; -/*static int adm5120_txhi = 0;*/ -static int adm5120_if_open = 0; +static struct dma_desc txh_descs_v[TX_RING_SIZE] __attribute__((aligned(16))); +static struct dma_desc txl_descs_v[TX_RING_SIZE] __attribute__((aligned(16))); +static struct dma_desc rxh_descs_v[RX_RING_SIZE] __attribute__((aligned(16))); +static struct dma_desc rxl_descs_v[RX_RING_SIZE] __attribute__((aligned(16))); +static struct dma_desc *txh_descs; +static struct dma_desc *txl_descs; +static struct dma_desc *rxh_descs; +static struct dma_desc *rxl_descs; -static inline void adm5120_set_reg(unsigned int reg, unsigned long val) +static struct sk_buff *rxl_skbuff[RX_RING_SIZE]; +static struct sk_buff *rxh_skbuff[RX_RING_SIZE]; +static struct sk_buff *txl_skbuff[TX_RING_SIZE]; +static struct sk_buff *txh_skbuff[TX_RING_SIZE]; + +static unsigned int cur_rxl, dirty_rxl; /* producer/consumer ring indices */ +static unsigned int cur_txl, dirty_txl; + +static unsigned int sw_used; + +static spinlock_t sw_lock = SPIN_LOCK_UNLOCKED; + +static struct net_device sw_dev; + +/* ------------------------------------------------------------------------ */ + +static inline u32 sw_read_reg(u32 reg) { - *(volatile unsigned long*)(SW_BASE+reg) = val; + return __raw_readl((void __iomem *)KSEG1ADDR(ADM5120_SWITCH_BASE)+reg); } -static inline unsigned long adm5120_get_reg(unsigned int reg) +static inline void sw_write_reg(u32 reg, u32 val) { - return *(volatile unsigned long*)(SW_BASE+reg); + __raw_writel(val, (void __iomem *)KSEG1ADDR(ADM5120_SWITCH_BASE)+reg); } -static inline void adm5120_rx_dma_update(struct adm5120_dma *dma, +static inline void sw_int_disable(u32 mask) +{ + u32 t; + + t = sw_read_reg(SWITCH_REG_INT_MASK); + t |= mask; + sw_write_reg(SWITCH_REG_INT_MASK, t); +} + +static inline void sw_int_enable(u32 mask) +{ + u32 t; + + t = sw_read_reg(SWITCH_REG_INT_MASK); + t &= ~mask; + sw_write_reg(SWITCH_REG_INT_MASK, t); +} + +static inline void sw_int_ack(u32 mask) +{ + sw_write_reg(SWITCH_REG_INT_STATUS, mask); +} + +/* ------------------------------------------------------------------------ */ + +static void sw_dump_desc(char *label, struct dma_desc *desc, int tx) +{ + u32 t; + + SW_DBG("%s %s desc/%p\n", label, tx ? "tx" : "rx", desc); + + t = desc->buf1; + SW_DBG(" buf1 %08X addr=%08X; len=%08X %s%s\n", t, + t & DESC_ADDR_MASK, + desc->buflen, + (t & DESC_OWN) ? "SWITCH" : "CPU", + (t & DESC_EOR) ? " RE" : ""); + + t = desc->buf2; + SW_DBG(" buf2 %08X addr=%08X%s\n", desc->buf2, + t & DESC_ADDR_MASK, + (t & DESC_BUF2_EN) ? " EN" : "" ); + + t = desc->misc; + if (tx) + SW_DBG(" misc %08X%s pktlen=%04X ports=%02X vlan=%02X\n", t, + (t & DESC_CSUM) ? " CSUM" : "", + (t >> DESC_PKTLEN_SHIFT) & DESC_PKTLEN_MASK, + (t >> DESC_DSTPORT_SHIFT) & DESC_DSTPORT_MASK, + t & DESC_VLAN_MASK); + else + SW_DBG(" misc %08X pktlen=%04X port=%d DA=%d%s%s type=%d\n", + t, + (t >> DESC_PKTLEN_SHIFT) & DESC_PKTLEN_MASK, + (t >> DESC_SRCPORT_SHIFT) & DESC_SRCPORT_MASK, + (t >> DESC_DA_SHIFT) & DESC_DA_MASK, + (t & DESC_IPCSUM_FAIL) ? " IPCF" : "", + (t & DESC_VLAN_TAG) ? " VLAN" : "", + (t & DESC_TYPE_MASK)); +} + +static void sw_dump_intr_mask(char *label, u32 mask) +{ + SW_DBG("%s %08X%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s\n", + label, mask, + (mask & SWITCH_INT_SHD) ? " SHD" : "", + (mask & SWITCH_INT_SLD) ? " SLD" : "", + (mask & SWITCH_INT_RHD) ? " RHD" : "", + (mask & SWITCH_INT_RLD) ? " RLD" : "", + (mask & SWITCH_INT_HDF) ? " HDF" : "", + (mask & SWITCH_INT_LDF) ? " LDF" : "", + (mask & SWITCH_INT_P0QF) ? " P0QF" : "", + (mask & SWITCH_INT_P1QF) ? " P1QF" : "", + (mask & SWITCH_INT_P2QF) ? " P2QF" : "", + (mask & SWITCH_INT_P3QF) ? " P3QF" : "", + (mask & SWITCH_INT_P4QF) ? " P4QF" : "", + (mask & SWITCH_INT_CPQF) ? " CPQF" : "", + (mask & SWITCH_INT_GQF) ? " GQF" : "", + (mask & SWITCH_INT_MD) ? " MD" : "", + (mask & SWITCH_INT_BCS) ? " BCS" : "", + (mask & SWITCH_INT_PSC) ? " PSC" : "", + (mask & SWITCH_INT_ID) ? " ID" : "", + (mask & SWITCH_INT_W0TE) ? " W0TE" : "", + (mask & SWITCH_INT_W1TE) ? " W1TE" : "", + (mask & SWITCH_INT_RDE) ? " RDE" : "", + (mask & SWITCH_INT_SDE) ? " SDE" : "", + (mask & SWITCH_INT_CPUH) ? " CPUH" : ""); +} + +/* ------------------------------------------------------------------------ */ + +static inline void adm5120_rx_dma_update(struct dma_desc *desc, struct sk_buff *skb, int end) { - dma->status = 0; - dma->cntl = 0; - dma->len = ADM5120_DMA_RXSIZE; - dma->data = ADM5120_DMA_ADDR(skb->data) | - ADM5120_DMA_OWN | (end ? ADM5120_DMA_RINGEND : 0); + desc->misc = 0; + desc->buf2 = 0; + desc->buflen = RX_MAX_PKTLEN; + desc->buf1 = DESC_ADDR(skb->data) | + DESC_OWN | (end ? DESC_EOR : 0); } -static int adm5120_rx(struct net_device *dev,int *budget) +static int adm5120_switch_rx(struct net_device *dev, int *budget) { struct sk_buff *skb, *skbn; struct adm5120_sw *priv; struct net_device *cdev; - struct adm5120_dma *dma; - int port, len, quota; + struct dma_desc *desc; + int len, quota; quota = min(dev->quota, *budget); - dma = &adm5120_dma_rxl[adm5120_rxli]; - while (!(dma->data & ADM5120_DMA_OWN) && quota) { - port = (dma->status & ADM5120_DMA_PORTID); - port >>= ADM5120_DMA_PORTSHIFT; + SW_DBG("%s polling, quota=%d\n", dev->name, quota); + + sw_int_ack(SWITCH_INTS_POLL); + + desc = &rxl_descs[cur_rxl]; + while (!(desc->buf1 & DESC_OWN) && quota) { + u32 port = desc_get_srcport(desc); cdev = adm5120_port[port]; if (cdev != dev) { /* The current packet belongs to a different device */ if ((cdev==NULL) || !netif_running(cdev)) { @@ -109,39 +305,38 @@ static int adm5120_rx(struct net_device *dev,int *budget) } } - skb = adm5120_skb_rxl[adm5120_rxli]; - len = (dma->status & ADM5120_DMA_LEN); - len >>= ADM5120_DMA_LENSHIFT; - len -= ETH_FCS; + skb = rxl_skbuff[cur_rxl]; + len = desc_get_pktlen(desc); + len -= ETH_CSUM_LEN; priv = netdev_priv(dev); - if (len <= 0 || len > ADM5120_DMA_RXSIZE || - dma->status & ADM5120_DMA_FCSERR) { - priv->stats.rx_errors++; + if (len <= 0 || len > RX_MAX_PKTLEN || + desc_ipcsum_fail(desc)) { + dev->stats.rx_errors++; skbn = NULL; } else { - skbn = dev_alloc_skb(ADM5120_DMA_RXSIZE+16); + skbn = dev_alloc_skb(SKB_ALLOC_LEN); if (skbn) { skb_put(skb, len); skb->dev = dev; skb->protocol = eth_type_trans(skb, dev); skb->ip_summed = CHECKSUM_UNNECESSARY; dev->last_rx = jiffies; - priv->stats.rx_packets++; - priv->stats.rx_bytes += len; - skb_reserve(skbn, NET_IP_ALIGN); - adm5120_skb_rxl[adm5120_rxli] = skbn; + dev->stats.rx_packets++; + dev->stats.rx_bytes += len; + skb_reserve(skbn, SKB_RESERVE_LEN); + rxl_skbuff[cur_rxl] = skbn; } else { - printk(KERN_INFO "%s recycling!\n", dev->name); + SW_INFO("%s recycling!\n", dev->name); } } rx_skip: - adm5120_rx_dma_update(&adm5120_dma_rxl[adm5120_rxli], - adm5120_skb_rxl[adm5120_rxli], - (ADM5120_DMA_RXL-1==adm5120_rxli)); - if (ADM5120_DMA_RXL == ++adm5120_rxli) - adm5120_rxli = 0; - dma = &adm5120_dma_rxl[adm5120_rxli]; + adm5120_rx_dma_update(&rxl_descs[cur_rxl], + rxl_skbuff[cur_rxl], + (RX_RING_SIZE-1==cur_rxl)); + if (RX_RING_SIZE == ++cur_rxl) + cur_rxl = 0; + desc = &rxl_descs[cur_rxl]; if (skbn){ netif_receive_skb(skb); dev->quota--; @@ -149,59 +344,92 @@ rx_skip: quota--; } } /* while */ - /* If there are still packets to process, return 1 */ - if (quota){ - /* No more packets to process, so disable the polling and reenable the interrupts */ + + if (quota) { netif_rx_complete(dev); - adm5120_set_reg(ADM5120_INT_MASK, - adm5120_get_reg(ADM5120_INT_MASK) & - ~(ADM5120_INT_RXL|ADM5120_INT_LFULL)); + sw_int_enable(SWITCH_INTS_POLL); return 0; - - } + + /* If there are still packets to process, return 1 */ return 1; } -static irqreturn_t adm5120_sw_irq(int irq, void *dev_id) +static void adm5120_switch_tx(void) { - unsigned long intreg, intmask; - int port; - struct net_device *dev; + unsigned int entry; - intmask = adm5120_get_reg(ADM5120_INT_MASK); /* Remember interrupt mask */ - adm5120_set_reg(ADM5120_INT_MASK, ADM5120_INTMASKALL); /* Disable interrupts */ + /* find and cleanup dirty tx descriptors */ + entry = dirty_txl % TX_RING_SIZE; + while (dirty_txl != cur_txl) { + struct dma_desc *desc = &txl_descs[entry]; + struct sk_buff *skb = txl_skbuff[entry]; - intreg = adm5120_get_reg(ADM5120_INT_ST); /* Read interrupt status */ - adm5120_set_reg(ADM5120_INT_ST, intreg); /* Clear interrupt status */ + if (desc->buf1 & DESC_OWN) + break; - /* In NAPI operation the interrupts are disabled and the polling mechanism - * is activated. The interrupts are finally enabled again in the polling routine. - */ - if (intreg & (ADM5120_INT_RXL|ADM5120_INT_LFULL)) { - /* check rx buffer for port number */ - port = adm5120_dma_rxl[adm5120_rxli].status & ADM5120_DMA_PORTID; - port >>= ADM5120_DMA_PORTSHIFT; - dev = adm5120_port[port]; - if ((dev==NULL) || !netif_running(dev)) { - /* discard (update with old skb) */ - adm5120_rx_dma_update(&adm5120_dma_rxl[adm5120_rxli], - adm5120_skb_rxl[adm5120_rxli], - (ADM5120_DMA_RXL-1==adm5120_rxli)); - if (ADM5120_DMA_RXL == ++adm5120_rxli) - adm5120_rxli = 0; + sw_dump_desc("tx done", desc, 1); + if (netif_running(skb->dev)) { + skb->dev->stats.tx_bytes += skb->len; + skb->dev->stats.tx_packets++; } - else { - netif_rx_schedule(dev); - intmask |= (ADM5120_INT_RXL|ADM5120_INT_LFULL); /* Disable RX interrupts */ + + dev_kfree_skb_irq(skb); + txl_skbuff[entry] = NULL; + entry = (++dirty_txl) % TX_RING_SIZE; + } + + if ((cur_txl - dirty_txl) < TX_QUEUE_LEN - 4) { + /* wake up queue of all devices */ + int i; + for (i = 0; i < SWITCH_NUM_PORTS; i++) { + if (!adm5120_devs[i]) + continue; + netif_wake_queue(adm5120_devs[i]); } } -#ifdef CONFIG_DEBUG - if (intreg & ~(intmask)) - printk(KERN_INFO "adm5120sw: IRQ 0x%08X unexpected!\n", (unsigned int)(intreg & ~(intmask))); -#endif +} - adm5120_set_reg(ADM5120_INT_MASK, intmask); +static irqreturn_t adm5120_poll_irq(int irq, void *dev_id) +{ + struct net_device *dev = dev_id; + u32 status; + + status = sw_read_reg(SWITCH_REG_INT_STATUS); + status &= ~(sw_read_reg(SWITCH_REG_INT_MASK)); + + status &= SWITCH_INTS_POLL; + if (!status) + return IRQ_NONE; + + sw_int_disable(SWITCH_INTS_POLL); + netif_rx_schedule(dev); + + SW_DBG("%s handling IRQ%d\n", dev->name, irq); + return IRQ_HANDLED; +} + +static irqreturn_t adm5120_switch_irq(int irq, void *dev_id) +{ + irqreturn_t ret; + u32 status; + + status = sw_read_reg(SWITCH_REG_INT_STATUS); + status &= ~(sw_read_reg(SWITCH_REG_INT_MASK)); + + sw_dump_intr_mask("sw irq", status); + + status &= SWITCH_INTS_ALL & ~SWITCH_INTS_POLL; + if (!status) + return IRQ_NONE; + + sw_int_ack(status); + + if (status & SWITCH_INT_SLD) { + spin_lock(&sw_lock); + adm5120_switch_tx(); + spin_unlock(&sw_lock); + } return IRQ_HANDLED; } @@ -212,13 +440,14 @@ static void adm5120_set_vlan(char *matrix) int vlan_port, port; val = matrix[0] + (matrix[1]<<8) + (matrix[2]<<16) + (matrix[3]<<24); - adm5120_set_reg(ADM5120_VLAN_GI, val); + sw_write_reg(SWITCH_REG_VLAN_G1, val); val = matrix[4] + (matrix[5]<<8); - adm5120_set_reg(ADM5120_VLAN_GII, val); + sw_write_reg(SWITCH_REG_VLAN_G2, val); + /* Now set/update the port vs. device lookup table */ - for (port=0; porttrans_start = jiffies; - dma = &adm5120_dma_txl[adm5120_txli]; - if (dma->data & ADM5120_DMA_OWN) { + /* calculate the next TX descriptor entry. */ + entry = cur_txl % TX_RING_SIZE; + + desc = &txl_descs[entry]; + if (desc->buf1 & DESC_OWN) { /* We want to write a packet but the TX queue is still * occupied by the DMA. We are faster than the DMA... */ dev_kfree_skb(skb); - priv->stats.tx_dropped++; + dev->stats.tx_dropped++; return 0; } - data = ADM5120_DMA_ADDR(skb->data) | ADM5120_DMA_OWN; - if (adm5120_txli == ADM5120_DMA_TXL-1) - data |= ADM5120_DMA_RINGEND; - dma->status = - ((skb->lenlen) << ADM5120_DMA_LENSHIFT) | + + txl_skbuff[entry] = skb; + data = (desc->buf1 & DESC_EOR); + data |= DESC_ADDR(skb->data); + + desc->misc = + ((skb->lenlen) << DESC_PKTLEN_SHIFT) | (0x1 << priv->port); - dma->len = skb->len < ETH_ZLEN ? ETH_ZLEN : skb->len; - priv->stats.tx_packets++; - priv->stats.tx_bytes += skb->len; + desc->buflen = skb->len < ETH_ZLEN ? ETH_ZLEN : skb->len; - /* free old skbs here instead of tx completion interrupt: - * will hold some more memory allocated but reduces interrupts */ - if (skbl[adm5120_txli]){ - dev_kfree_skb(skbl[adm5120_txli]); + /* lock switch irq */ + spin_lock_irq(&sw_lock); + + desc->buf1 = data | DESC_OWN; + sw_write_reg(SWITCH_REG_SEND_TRIG, SEND_TRIG_STL); + + cur_txl++; + if (cur_txl == dirty_txl + TX_QUEUE_LEN) { + /* FIXME: stop queue for all devices */ + netif_stop_queue(dev); } - skbl[adm5120_txli] = skb; - dma->data = data; /* Here we enable the buffer for the TX DMA machine */ - adm5120_set_reg(ADM5120_SEND_TRIG, ADM5120_SEND_TRIG_L); - if (++adm5120_txli == ADM5120_DMA_TXL) - adm5120_txli = 0; + dev->trans_start = jiffies; + + spin_unlock_irq(&sw_lock); + return 0; } static void adm5120_tx_timeout(struct net_device *dev) { - printk(KERN_INFO "%s: TX timeout\n",dev->name); -} - -static struct net_device_stats *adm5120_sw_stats(struct net_device *dev) -{ - struct adm5120_sw *priv = netdev_priv(dev); - int portmask; - unsigned long adm5120_cpup_conf_reg; - - portmask = adm5120_eth_vlans[priv->port] & 0x3f; - - adm5120_cpup_conf_reg = adm5120_get_reg(ADM5120_CPUP_CONF); - - if (dev->flags & IFF_PROMISC) - adm5120_cpup_conf_reg &= ~((portmask << ADM5120_DISUNSHIFT) & ADM5120_DISUNALL); - else - adm5120_cpup_conf_reg |= (portmask << ADM5120_DISUNSHIFT); - - if (dev->flags & IFF_PROMISC || dev->flags & IFF_ALLMULTI || dev->mc_count) - adm5120_cpup_conf_reg &= ~((portmask << ADM5120_DISMCSHIFT) & ADM5120_DISMCALL); - else - adm5120_cpup_conf_reg |= (portmask << ADM5120_DISMCSHIFT); - - /* If there is any port configured to be in promiscuous mode, then the */ - /* Bridge Test Mode has to be activated. This will result in */ - /* transporting also packets learned in another VLAN to be forwarded */ - /* to the CPU. */ - /* The difficult scenario is when we want to build a bridge on the CPU.*/ - /* Assume we have port0 and the CPU port in VLAN0 and port1 and the */ - /* CPU port in VLAN1. Now we build a bridge on the CPU between */ - /* VLAN0 and VLAN1. Both ports of the VLANs are set in promisc mode. */ - /* Now assume a packet with ethernet source address 99 enters port 0 */ - /* It will be forwarded to the CPU because it is unknown. Then the */ - /* bridge in the CPU will send it to VLAN1 and it goes out at port 1. */ - /* When now a packet with ethernet destination address 99 comes in at */ - /* port 1 in VLAN1, then the switch has learned that this address is */ - /* located at port 0 in VLAN0. Therefore the switch will drop */ - /* this packet. In order to avoid this and to send the packet still */ - /* to the CPU, the Bridge Test Mode has to be activated. */ - - /* Check if there is any vlan in promisc mode. */ - if (~adm5120_cpup_conf_reg & ADM5120_DISUNALL) - adm5120_cpup_conf_reg |= ADM5120_BTM; /* Set the BTM */ - else - adm5120_cpup_conf_reg &= ~ADM5120_BTM; /* Disable the BTM */ - - adm5120_set_reg(ADM5120_CPUP_CONF,adm5120_cpup_conf_reg); - - return &((struct adm5120_sw *)netdev_priv(dev))->stats; + SW_INFO("TX timeout on %s\n",dev->name); } static void adm5120_set_multicast_list(struct net_device *dev) { struct adm5120_sw *priv = netdev_priv(dev); - int portmask; + u32 ports; + u32 t; - portmask = adm5120_eth_vlans[priv->port] & 0x3f; + ports = adm5120_eth_vlans[priv->port] & SWITCH_PORTS_NOCPU; + t = sw_read_reg(SWITCH_REG_CPUP_CONF); if (dev->flags & IFF_PROMISC) - adm5120_set_reg(ADM5120_CPUP_CONF, - adm5120_get_reg(ADM5120_CPUP_CONF) & - ~((portmask << ADM5120_DISUNSHIFT) & ADM5120_DISUNALL)); + /* enable unknown packets */ + t &= ~(ports << CPUP_CONF_DUNP_SHIFT); else - adm5120_set_reg(ADM5120_CPUP_CONF, - adm5120_get_reg(ADM5120_CPUP_CONF) | - (portmask << ADM5120_DISUNSHIFT)); + /* disable unknown packets */ + t |= (ports << CPUP_CONF_DUNP_SHIFT); if (dev->flags & IFF_PROMISC || dev->flags & IFF_ALLMULTI || - dev->mc_count) - adm5120_set_reg(ADM5120_CPUP_CONF, - adm5120_get_reg(ADM5120_CPUP_CONF) & - ~((portmask << ADM5120_DISMCSHIFT) & ADM5120_DISMCALL)); + dev->mc_count) + /* enable multicast packets */ + t &= ~(ports << CPUP_CONF_DMCP_SHIFT); else - adm5120_set_reg(ADM5120_CPUP_CONF, - adm5120_get_reg(ADM5120_CPUP_CONF) | - (portmask << ADM5120_DISMCSHIFT)); + /* disable multicast packets */ + t |= (ports << CPUP_CONF_DMCP_SHIFT); + + /* If there is any port configured to be in promiscuous mode, then the */ + /* Bridge Test Mode has to be activated. This will result in */ + /* transporting also packets learned in another VLAN to be forwarded */ + /* to the CPU. */ + /* The difficult scenario is when we want to build a bridge on the CPU.*/ + /* Assume we have port0 and the CPU port in VLAN0 and port1 and the */ + /* CPU port in VLAN1. Now we build a bridge on the CPU between */ + /* VLAN0 and VLAN1. Both ports of the VLANs are set in promisc mode. */ + /* Now assume a packet with ethernet source address 99 enters port 0 */ + /* It will be forwarded to the CPU because it is unknown. Then the */ + /* bridge in the CPU will send it to VLAN1 and it goes out at port 1. */ + /* When now a packet with ethernet destination address 99 comes in at */ + /* port 1 in VLAN1, then the switch has learned that this address is */ + /* located at port 0 in VLAN0. Therefore the switch will drop */ + /* this packet. In order to avoid this and to send the packet still */ + /* to the CPU, the Bridge Test Mode has to be activated. */ + + /* Check if there is any vlan in promisc mode. */ + if (t & (SWITCH_PORTS_NOCPU << CPUP_CONF_DUNP_SHIFT)) + t &= ~CPUP_CONF_BTM; /* Disable Bridge Testing Mode */ + else + t |= CPUP_CONF_BTM; /* Enable Bridge Testing Mode */ + + sw_write_reg(SWITCH_REG_CPUP_CONF, t); + } static void adm5120_write_mac(struct net_device *dev) { struct adm5120_sw *priv = netdev_priv(dev); unsigned char *mac = dev->dev_addr; + u32 t; - adm5120_set_reg(ADM5120_MAC_WT1, - mac[2] | (mac[3]<<8) | (mac[4]<<16) | (mac[5]<<24)); - adm5120_set_reg(ADM5120_MAC_WT0, (priv->port<<3) | - (mac[0]<<16) | (mac[1]<<24) | ADM5120_MAC_WRITE | ADM5120_VLAN_EN); + t = mac[2] | (mac[3] << MAC_WT1_MAC3_SHIFT) | + (mac[4] << MAC_WT1_MAC4_SHIFT) | (mac[5] << MAC_WT1_MAC4_SHIFT); + sw_write_reg(SWITCH_REG_MAC_WT1, t); - while (!(adm5120_get_reg(ADM5120_MAC_WT0) & ADM5120_MAC_WRITE_DONE)); + t = (mac[0] << MAC_WT0_MAC0_SHIFT) | (mac[1] << MAC_WT0_MAC1_SHIFT) | + MAC_WT0_MAWC | MAC_WT0_WVE | (priv->port<<3); + + sw_write_reg(SWITCH_REG_MAC_WT0, t); + + while (!(sw_read_reg(SWITCH_REG_MAC_WT0) & MAC_WT0_MWD)); } static int adm5120_sw_set_mac_address(struct net_device *dev, void *p) @@ -436,115 +653,133 @@ static int adm5120_do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) struct adm5120_sw *priv = netdev_priv(dev); switch(cmd) { - case SIOCGADMINFO: - info.magic = 0x5120; - info.ports = adm5120_nrdevs; - info.vlan = priv->port; - err = copy_to_user(rq->ifr_data, &info, sizeof(info)); - if (err) - return -EFAULT; - break; - case SIOCSMATRIX: - if (!capable(CAP_NET_ADMIN)) - return -EPERM; - err = copy_from_user(adm5120_eth_vlans, rq->ifr_data, - sizeof(adm5120_eth_vlans)); - if (err) - return -EFAULT; - adm5120_set_vlan(adm5120_eth_vlans); - break; - case SIOCGMATRIX: - err = copy_to_user(rq->ifr_data, adm5120_eth_vlans, - sizeof(adm5120_eth_vlans)); - if (err) - return -EFAULT; - break; - case SIOCGETBW: - err = copy_to_user(rq->ifr_data, bw_matrix, sizeof(bw_matrix)); - if (err) - return -EFAULT; - break; - case SIOCSETBW: - if (!capable(CAP_NET_ADMIN)) - return -EPERM; - err = copy_from_user(bw_matrix, rq->ifr_data, sizeof(bw_matrix)); - if (err) - return -EFAULT; - adm5120_set_bw(bw_matrix); - break; - default: - return -EOPNOTSUPP; + case SIOCGADMINFO: + info.magic = 0x5120; + info.ports = adm5120_nrdevs; + info.vlan = priv->port; + err = copy_to_user(rq->ifr_data, &info, sizeof(info)); + if (err) + return -EFAULT; + break; + case SIOCSMATRIX: + if (!capable(CAP_NET_ADMIN)) + return -EPERM; + err = copy_from_user(adm5120_eth_vlans, rq->ifr_data, + sizeof(adm5120_eth_vlans)); + if (err) + return -EFAULT; + adm5120_set_vlan(adm5120_eth_vlans); + break; + case SIOCGMATRIX: + err = copy_to_user(rq->ifr_data, adm5120_eth_vlans, + sizeof(adm5120_eth_vlans)); + if (err) + return -EFAULT; + break; + case SIOCGETBW: + err = copy_to_user(rq->ifr_data, bw_matrix, sizeof(bw_matrix)); + if (err) + return -EFAULT; + break; + case SIOCSETBW: + if (!capable(CAP_NET_ADMIN)) + return -EPERM; + err = copy_from_user(bw_matrix, rq->ifr_data, sizeof(bw_matrix)); + if (err) + return -EFAULT; + adm5120_set_bw(bw_matrix); + break; + default: + return -EOPNOTSUPP; } return 0; } -static void adm5120_dma_tx_init(struct adm5120_dma *dma, struct sk_buff **skbl, - int num) +static void adm5120_dma_tx_init(struct dma_desc *desc, struct sk_buff **skbl, + int num) { - memset(dma, 0, sizeof(struct adm5120_dma)*num); - dma[num-1].data |= ADM5120_DMA_RINGEND; + memset(desc, 0, num * sizeof(*desc)); + desc[num-1].buf1 |= DESC_EOR; memset(skbl, 0, sizeof(struct skb*)*num); } -static void adm5120_dma_rx_init(struct adm5120_dma *dma, struct sk_buff **skbl, - int num) +static void adm5120_dma_rx_init(struct dma_desc *desc, struct sk_buff **skbl, + int num) { int i; - memset(dma, 0, sizeof(struct adm5120_dma)*num); + memset(desc, 0, num * sizeof(*desc)); for (i=0; iname, dev); + if (err) { + SW_ERR("unable to get irq for %s\n", dev->name); + goto out_int; + } + SET_MODULE_OWNER(dev); memset(netdev_priv(dev), 0, sizeof(struct adm5120_sw)); ((struct adm5120_sw*)netdev_priv(dev))->port = i; - dev->base_addr = SW_BASE; + dev->base_addr = ADM5120_SWITCH_BASE; dev->irq = ADM5120_IRQ_SWITCH; - dev->open = adm5120_sw_open; - dev->hard_start_xmit = adm5120_sw_tx; - dev->stop = adm5120_sw_stop; - dev->get_stats = adm5120_sw_stats; + dev->open = adm5120_switch_open; + dev->hard_start_xmit = adm5120_sw_start_xmit; + dev->stop = adm5120_switch_stop; dev->set_multicast_list = adm5120_set_multicast_list; dev->do_ioctl = adm5120_do_ioctl; dev->tx_timeout = adm5120_tx_timeout; - dev->watchdog_timeo = ETH_TX_TIMEOUT; + dev->watchdog_timeo = TX_TIMEOUT; dev->set_mac_address = adm5120_sw_set_mac_address; - dev->poll = adm5120_rx; + dev->poll = adm5120_switch_rx; dev->weight = 64; memcpy(dev->dev_addr, adm5120_eth_macs[i], 6); adm5120_write_mac(dev); if ((err = register_netdev(dev))) { + free_irq(ADM5120_IRQ_SWITCH, dev); free_netdev(dev); goto out_int; } - printk(KERN_INFO "%s: ADM5120 switch port%d\n", dev->name, i); + SW_INFO("%s created for switch port%d\n", dev->name, i); } /* setup vlan/port mapping after devs are filled up */ adm5120_set_vlan(adm5120_eth_vlans); - adm5120_set_reg(ADM5120_CPUP_CONF, - ADM5120_CRC_PADDING | ADM5120_DISUNALL | ADM5120_DISMCALL); + /* enable CPU port */ + t = sw_read_reg(SWITCH_REG_CPUP_CONF); + t &= ~CPUP_CONF_DCPUP; + sw_write_reg(SWITCH_REG_CPUP_CONF, t); return 0; @@ -594,7 +838,7 @@ out_int: } free_irq(ADM5120_IRQ_SWITCH, NULL); out: - printk(KERN_ERR "ADM5120 Ethernet switch init failed\n"); + SW_ERR("init failed\n"); return err; } @@ -602,22 +846,25 @@ static void __exit adm5120_sw_exit(void) { int i; - for (i = 0; i < SW_DEVS; i++) { - unregister_netdev(adm5120_devs[i]); - free_netdev(adm5120_devs[i]); + for (i = 0; i < SWITCH_NUM_PORTS; i++) { + struct net_device *dev = adm5120_devs[i]; + unregister_netdev(dev); + free_irq(ADM5120_IRQ_SWITCH, dev); + free_netdev(dev); } - free_irq(ADM5120_IRQ_SWITCH, NULL); + free_irq(ADM5120_IRQ_SWITCH, &sw_dev); - for (i = 0; i < ADM5120_DMA_RXH; i++) { - if (!adm5120_skb_rxh[i]) + for (i = 0; i < RX_RING_SIZE; i++) { + if (!rxh_skbuff[i]) break; - kfree_skb(adm5120_skb_rxh[i]); + kfree_skb(rxh_skbuff[i]); } - for (i = 0; i < ADM5120_DMA_RXL; i++) { - if (!adm5120_skb_rxl[i]) + + for (i = 0; i < RX_RING_SIZE; i++) { + if (!rxl_skbuff[i]) break; - kfree_skb(adm5120_skb_rxl[i]); + kfree_skb(rxl_skbuff[i]); } } diff --git a/target/linux/adm5120/files/drivers/net/adm5120sw.h b/target/linux/adm5120/files/drivers/net/adm5120sw.h index a3b4ec0179..34f1213749 100644 --- a/target/linux/adm5120/files/drivers/net/adm5120sw.h +++ b/target/linux/adm5120/files/drivers/net/adm5120sw.h @@ -10,93 +10,8 @@ #ifndef _INCLUDE_ADM5120SW_H_ #define _INCLUDE_ADM5120SW_H_ -#define SW_BASE KSEG1ADDR(0x12000000) -#define SW_DEVS 6 - -#define ETH_TX_TIMEOUT HZ*400 -#define ETH_FCS 4; - -#define ADM5120_CODE 0x00 /* CPU description */ -#define ADM5120_CODE_PQFP 0x20000000 /* package type */ -#define ADM5120_SW_CONF 0x20 /* Switch configuration register */ -#define ADM5120_SW_CONF_BPM 0x00300000 /* Mask for backpressure mode */ -#define ADM5120_CPUP_CONF 0x24 /* CPU port config */ -#define ADM5120_DISCCPUPORT 0x00000001 /* disable cpu port */ -#define ADM5120_CRC_PADDING 0x00000002 /* software crc */ -#define ADM5120_BTM 0x00000004 /* bridge test mode */ -#define ADM5120_DISUNSHIFT 9 -#define ADM5120_DISUNALL 0x00007e00 /* disable unknown from all */ -#define ADM5120_DISMCSHIFT 16 -#define ADM5120_DISMCALL 0x003f0000 /* disable multicast from all */ -#define ADM5120_PORT_CONF0 0x28 -#define ADM5120_ENMC 0x00003f00 /* Enable MC routing (ex cpu) */ -#define ADM5120_ENBP 0x003f0000 /* Enable Back Pressure */ -#define ADM5120_PORTDISALL 0x0000003F -#define ADM5120_VLAN_GI 0x40 /* VLAN settings */ -#define ADM5120_VLAN_GII 0x44 -#define ADM5120_SEND_TRIG 0x48 -#define ADM5120_SEND_TRIG_L 0x00000001 -#define ADM5120_SEND_TRIG_H 0x00000002 -#define ADM5120_MAC_WT0 0x58 -#define ADM5120_MAC_WRITE 0x00000001 -#define ADM5120_MAC_WRITE_DONE 0x00000002 -#define ADM5120_VLAN_EN 0x00000040 -#define ADM5120_MAC_WT1 0x5c -#define ADM5120_BW_CTL0 0x60 /* Bandwidth control 0 */ -#define ADM5120_BW_CTL1 0x64 /* Bandwidth control 1 */ -#define ADM5120_PHY_CNTL2 0x7c -#define ADM5120_AUTONEG 0x0000001f /* Auto negotiate */ -#define ADM5120_NORMAL 0x01f00000 /* PHY normal mode */ -#define ADM5120_AUTOMDIX 0x3e000000 /* Auto MDIX */ -#define ADM5120_PHY_CNTL3 0x80 -#define ADM5120_PHY_NTH 0x00000400 -#define ADM5120_PRI_CNTL 0x84 -#define ADM5120_INT_ST 0xb0 -#define ADM5120_INT_RXH 0x0000004 -#define ADM5120_INT_RXL 0x0000008 -#define ADM5120_INT_HFULL 0x0000010 -#define ADM5120_INT_LFULL 0x0000020 -#define ADM5120_INT_TXH 0x0000001 -#define ADM5120_INT_TXL 0x0000002 -#define ADM5120_INT_MASK 0xb4 -#define ADM5120_INTMASKALL 0x1FDEFFF /* All interrupts */ -#define ADM5120_INTHANDLE (ADM5120_INT_RXH | ADM5120_INT_RXL | \ - ADM5120_INT_HFULL | ADM5120_INT_LFULL | \ - ADM5120_INT_TXH | ADM5120_INT_TXL) -#define ADM5120_SEND_HBADDR 0xd0 -#define ADM5120_SEND_LBADDR 0xd4 -#define ADM5120_RECEIVE_HBADDR 0xd8 -#define ADM5120_RECEIVE_LBADDR 0xdc - -struct adm5120_dma { - u32 data; - u32 cntl; - u32 len; - u32 status; -} __attribute__ ((packed)); - -#define ADM5120_DMA_MASK 0x01ffffff -#define ADM5120_DMA_OWN 0x80000000 /* buffer owner */ -#define ADM5120_DMA_RINGEND 0x10000000 /* Last in DMA ring */ - -#define ADM5120_DMA_ADDR(ptr) ((u32)(ptr) & ADM5120_DMA_MASK) -#define ADM5120_DMA_PORTID 0x00007000 -#define ADM5120_DMA_PORTSHIFT 12 -#define ADM5120_DMA_LEN 0x07ff0000 -#define ADM5120_DMA_LENSHIFT 16 -#define ADM5120_DMA_FCSERR 0x00000008 - -#define ADM5120_DMA_TXH 2 -#define ADM5120_DMA_TXL 64 -#define ADM5120_DMA_RXH 2 -#define ADM5120_DMA_RXL 64 - -#define ADM5120_DMA_RXSIZE 1550 -#define ADM5120_DMA_EXTRA 20 - struct adm5120_sw { - int port; - struct net_device_stats stats; + int port; }; #define SIOCSMATRIX SIOCDEVPRIVATE diff --git a/target/linux/adm5120/files/include/asm-mips/mach-adm5120/adm5120_switch.h b/target/linux/adm5120/files/include/asm-mips/mach-adm5120/adm5120_switch.h index c3af94ba74..d185ce5316 100644 --- a/target/linux/adm5120/files/include/asm-mips/mach-adm5120/adm5120_switch.h +++ b/target/linux/adm5120/files/include/asm-mips/mach-adm5120/adm5120_switch.h @@ -141,6 +141,9 @@ #define CPUP_CONF_DCPUP BIT(0) /* Disable CPU port */ #define CPUP_CONF_CRCP BIT(1) /* CRC padding from CPU */ #define CPUP_CONF_BTM BIT(2) /* Bridge Testing Mode */ +#define CPUP_CONF_DUNP_SHIFT 9 /* Disable Unknown Packets for portX */ +#define CPUP_CONF_DMCP_SHIFT 16 /* Disable Mcast Packets form portX */ +#define CPUP_CONF_DBCP_SHIFT 24 /* Disable Bcast Packets form portX */ /* PORT_CONF0 register bits */ #define PORT_CONF0_DP_SHIFT 0 /* Disable Port */ @@ -156,6 +159,26 @@ #define SEND_TRIG_STL BIT(0) /* Send Trigger Low */ #define SEND_TRIG_STH BIT(1) /* Send Trigger High */ +/* MAC_WT0 register bits */ +#define MAC_WT0_MAWC BIT(0) /* MAC address write command */ +#define MAC_WT0_MWD_SHIFT 1 +#define MAC_WT0_MWD BIT(1) /* MAC write done */ +#define MAC_WT0_WFB BIT(2) /* Write Filter Bit */ +#define MAC_WT0_WVN_SHIFT 3 +#define MAC_WT0_WVE BIT(6) /* Write VLAN enable */ +#define MAC_WT0_WPMN_SHIFT 7 +#define MAC_WT0_WAF_SHIFT 13 /* Write Age Field shift */ +#define MAC_WT0_WAF_EMPTY 0 +#define MAC_WT0_WAF_STATIC 7 +#define MAC_WT0_MAC0_SHIFT 16 +#define MAC_WT0_MAC1_SHIFT 24 + +/* MAC_WT1 register bits */ +#define MAC_WT1_MAC2_SHIFT 0 +#define MAC_WT1_MAC3_SHIFT 8 +#define MAC_WT1_MAC4_SHIFT 16 +#define MAC_WT1_MAC5_SHIFT 24 + /* BW_CNTL0/BW_CNTL1 register bits */ #define BW_CNTL_DISABLE 0x00 #define BW_CNTL_64K 0x01 @@ -198,6 +221,9 @@ /* PHY_CNTL2_RMAE is bad in datasheet */ #define PHY_CNTL2_RMAE BIT(31) /* Recommended MCC Average enable */ +/* PHY_CNTL3 register bits */ +#define PHY_CNTL3_RNT BIT(10) /* Recommend Normal Threshold */ + /* PORT_TH register bits */ #define PORT_TH_PPT_MASK BITMASK(8) /* Per Port Threshold */ #define PORT_TH_CPUT_SHIFT 8 /* CPU Port Buffer Threshold */