|  | /* | 
|  | * Copyright (C) 2006-2007 Freescale Semicondutor, Inc. All rights reserved. | 
|  | * | 
|  | * Author: Shlomi Gridish <gridish@freescale.com> | 
|  | *	   Li Yang <leoli@freescale.com> | 
|  | * | 
|  | * Description: | 
|  | * QE UCC Gigabit Ethernet Driver | 
|  | * | 
|  | * This program is free software; you can redistribute  it and/or modify it | 
|  | * under  the terms of  the GNU General  Public License as published by the | 
|  | * Free Software Foundation;  either version 2 of the  License, or (at your | 
|  | * option) any later version. | 
|  | */ | 
|  | #include <linux/kernel.h> | 
|  | #include <linux/init.h> | 
|  | #include <linux/errno.h> | 
|  | #include <linux/slab.h> | 
|  | #include <linux/stddef.h> | 
|  | #include <linux/interrupt.h> | 
|  | #include <linux/netdevice.h> | 
|  | #include <linux/etherdevice.h> | 
|  | #include <linux/skbuff.h> | 
|  | #include <linux/spinlock.h> | 
|  | #include <linux/mm.h> | 
|  | #include <linux/dma-mapping.h> | 
|  | #include <linux/fsl_devices.h> | 
|  | #include <linux/mii.h> | 
|  | #include <linux/phy.h> | 
|  | #include <linux/workqueue.h> | 
|  |  | 
|  | #include <asm/of_platform.h> | 
|  | #include <asm/uaccess.h> | 
|  | #include <asm/irq.h> | 
|  | #include <asm/io.h> | 
|  | #include <asm/immap_qe.h> | 
|  | #include <asm/qe.h> | 
|  | #include <asm/ucc.h> | 
|  | #include <asm/ucc_fast.h> | 
|  |  | 
|  | #include "ucc_geth.h" | 
|  | #include "ucc_geth_mii.h" | 
|  |  | 
|  | #undef DEBUG | 
|  |  | 
|  | #define ugeth_printk(level, format, arg...)  \ | 
|  | printk(level format "\n", ## arg) | 
|  |  | 
|  | #define ugeth_dbg(format, arg...)            \ | 
|  | ugeth_printk(KERN_DEBUG , format , ## arg) | 
|  | #define ugeth_err(format, arg...)            \ | 
|  | ugeth_printk(KERN_ERR , format , ## arg) | 
|  | #define ugeth_info(format, arg...)           \ | 
|  | ugeth_printk(KERN_INFO , format , ## arg) | 
|  | #define ugeth_warn(format, arg...)           \ | 
|  | ugeth_printk(KERN_WARNING , format , ## arg) | 
|  |  | 
|  | #ifdef UGETH_VERBOSE_DEBUG | 
|  | #define ugeth_vdbg ugeth_dbg | 
|  | #else | 
|  | #define ugeth_vdbg(fmt, args...) do { } while (0) | 
|  | #endif				/* UGETH_VERBOSE_DEBUG */ | 
|  | #define UGETH_MSG_DEFAULT	(NETIF_MSG_IFUP << 1 ) - 1 | 
|  |  | 
|  | void uec_set_ethtool_ops(struct net_device *netdev); | 
|  |  | 
|  | static DEFINE_SPINLOCK(ugeth_lock); | 
|  |  | 
|  | static struct { | 
|  | u32 msg_enable; | 
|  | } debug = { -1 }; | 
|  |  | 
|  | module_param_named(debug, debug.msg_enable, int, 0); | 
|  | MODULE_PARM_DESC(debug, "Debug verbosity level (0=none, ..., 0xffff=all)"); | 
|  |  | 
|  | static struct ucc_geth_info ugeth_primary_info = { | 
|  | .uf_info = { | 
|  | .bd_mem_part = MEM_PART_SYSTEM, | 
|  | .rtsm = UCC_FAST_SEND_IDLES_BETWEEN_FRAMES, | 
|  | .max_rx_buf_length = 1536, | 
|  | /* adjusted at startup if max-speed 1000 */ | 
|  | .urfs = UCC_GETH_URFS_INIT, | 
|  | .urfet = UCC_GETH_URFET_INIT, | 
|  | .urfset = UCC_GETH_URFSET_INIT, | 
|  | .utfs = UCC_GETH_UTFS_INIT, | 
|  | .utfet = UCC_GETH_UTFET_INIT, | 
|  | .utftt = UCC_GETH_UTFTT_INIT, | 
|  | .ufpt = 256, | 
|  | .mode = UCC_FAST_PROTOCOL_MODE_ETHERNET, | 
|  | .ttx_trx = UCC_FAST_GUMR_TRANSPARENT_TTX_TRX_NORMAL, | 
|  | .tenc = UCC_FAST_TX_ENCODING_NRZ, | 
|  | .renc = UCC_FAST_RX_ENCODING_NRZ, | 
|  | .tcrc = UCC_FAST_16_BIT_CRC, | 
|  | .synl = UCC_FAST_SYNC_LEN_NOT_USED, | 
|  | }, | 
|  | .numQueuesTx = 1, | 
|  | .numQueuesRx = 1, | 
|  | .extendedFilteringChainPointer = ((uint32_t) NULL), | 
|  | .typeorlen = 3072 /*1536 */ , | 
|  | .nonBackToBackIfgPart1 = 0x40, | 
|  | .nonBackToBackIfgPart2 = 0x60, | 
|  | .miminumInterFrameGapEnforcement = 0x50, | 
|  | .backToBackInterFrameGap = 0x60, | 
|  | .mblinterval = 128, | 
|  | .nortsrbytetime = 5, | 
|  | .fracsiz = 1, | 
|  | .strictpriorityq = 0xff, | 
|  | .altBebTruncation = 0xa, | 
|  | .excessDefer = 1, | 
|  | .maxRetransmission = 0xf, | 
|  | .collisionWindow = 0x37, | 
|  | .receiveFlowControl = 1, | 
|  | .transmitFlowControl = 1, | 
|  | .maxGroupAddrInHash = 4, | 
|  | .maxIndAddrInHash = 4, | 
|  | .prel = 7, | 
|  | .maxFrameLength = 1518, | 
|  | .minFrameLength = 64, | 
|  | .maxD1Length = 1520, | 
|  | .maxD2Length = 1520, | 
|  | .vlantype = 0x8100, | 
|  | .ecamptr = ((uint32_t) NULL), | 
|  | .eventRegMask = UCCE_OTHER, | 
|  | .pausePeriod = 0xf000, | 
|  | .interruptcoalescingmaxvalue = {1, 1, 1, 1, 1, 1, 1, 1}, | 
|  | .bdRingLenTx = { | 
|  | TX_BD_RING_LEN, | 
|  | TX_BD_RING_LEN, | 
|  | TX_BD_RING_LEN, | 
|  | TX_BD_RING_LEN, | 
|  | TX_BD_RING_LEN, | 
|  | TX_BD_RING_LEN, | 
|  | TX_BD_RING_LEN, | 
|  | TX_BD_RING_LEN}, | 
|  |  | 
|  | .bdRingLenRx = { | 
|  | RX_BD_RING_LEN, | 
|  | RX_BD_RING_LEN, | 
|  | RX_BD_RING_LEN, | 
|  | RX_BD_RING_LEN, | 
|  | RX_BD_RING_LEN, | 
|  | RX_BD_RING_LEN, | 
|  | RX_BD_RING_LEN, | 
|  | RX_BD_RING_LEN}, | 
|  |  | 
|  | .numStationAddresses = UCC_GETH_NUM_OF_STATION_ADDRESSES_1, | 
|  | .largestexternallookupkeysize = | 
|  | QE_FLTR_LARGEST_EXTERNAL_TABLE_LOOKUP_KEY_SIZE_NONE, | 
|  | .statisticsMode = UCC_GETH_STATISTICS_GATHERING_MODE_HARDWARE | | 
|  | UCC_GETH_STATISTICS_GATHERING_MODE_FIRMWARE_TX | | 
|  | UCC_GETH_STATISTICS_GATHERING_MODE_FIRMWARE_RX, | 
|  | .vlanOperationTagged = UCC_GETH_VLAN_OPERATION_TAGGED_NOP, | 
|  | .vlanOperationNonTagged = UCC_GETH_VLAN_OPERATION_NON_TAGGED_NOP, | 
|  | .rxQoSMode = UCC_GETH_QOS_MODE_DEFAULT, | 
|  | .aufc = UPSMR_AUTOMATIC_FLOW_CONTROL_MODE_NONE, | 
|  | .padAndCrc = MACCFG2_PAD_AND_CRC_MODE_PAD_AND_CRC, | 
|  | .numThreadsTx = UCC_GETH_NUM_OF_THREADS_4, | 
|  | .numThreadsRx = UCC_GETH_NUM_OF_THREADS_4, | 
|  | .riscTx = QE_RISC_ALLOCATION_RISC1_AND_RISC2, | 
|  | .riscRx = QE_RISC_ALLOCATION_RISC1_AND_RISC2, | 
|  | }; | 
|  |  | 
|  | static struct ucc_geth_info ugeth_info[8]; | 
|  |  | 
|  | #ifdef DEBUG | 
|  | static void mem_disp(u8 *addr, int size) | 
|  | { | 
|  | u8 *i; | 
|  | int size16Aling = (size >> 4) << 4; | 
|  | int size4Aling = (size >> 2) << 2; | 
|  | int notAlign = 0; | 
|  | if (size % 16) | 
|  | notAlign = 1; | 
|  |  | 
|  | for (i = addr; (u32) i < (u32) addr + size16Aling; i += 16) | 
|  | printk("0x%08x: %08x %08x %08x %08x\r\n", | 
|  | (u32) i, | 
|  | *((u32 *) (i)), | 
|  | *((u32 *) (i + 4)), | 
|  | *((u32 *) (i + 8)), *((u32 *) (i + 12))); | 
|  | if (notAlign == 1) | 
|  | printk("0x%08x: ", (u32) i); | 
|  | for (; (u32) i < (u32) addr + size4Aling; i += 4) | 
|  | printk("%08x ", *((u32 *) (i))); | 
|  | for (; (u32) i < (u32) addr + size; i++) | 
|  | printk("%02x", *((u8 *) (i))); | 
|  | if (notAlign == 1) | 
|  | printk("\r\n"); | 
|  | } | 
|  | #endif /* DEBUG */ | 
|  |  | 
|  | #ifdef CONFIG_UGETH_FILTERING | 
|  | static void enqueue(struct list_head *node, struct list_head *lh) | 
|  | { | 
|  | unsigned long flags; | 
|  |  | 
|  | spin_lock_irqsave(&ugeth_lock, flags); | 
|  | list_add_tail(node, lh); | 
|  | spin_unlock_irqrestore(&ugeth_lock, flags); | 
|  | } | 
|  | #endif /* CONFIG_UGETH_FILTERING */ | 
|  |  | 
|  | static struct list_head *dequeue(struct list_head *lh) | 
|  | { | 
|  | unsigned long flags; | 
|  |  | 
|  | spin_lock_irqsave(&ugeth_lock, flags); | 
|  | if (!list_empty(lh)) { | 
|  | struct list_head *node = lh->next; | 
|  | list_del(node); | 
|  | spin_unlock_irqrestore(&ugeth_lock, flags); | 
|  | return node; | 
|  | } else { | 
|  | spin_unlock_irqrestore(&ugeth_lock, flags); | 
|  | return NULL; | 
|  | } | 
|  | } | 
|  |  | 
|  | static struct sk_buff *get_new_skb(struct ucc_geth_private *ugeth, u8 *bd) | 
|  | { | 
|  | struct sk_buff *skb = NULL; | 
|  |  | 
|  | skb = dev_alloc_skb(ugeth->ug_info->uf_info.max_rx_buf_length + | 
|  | UCC_GETH_RX_DATA_BUF_ALIGNMENT); | 
|  |  | 
|  | if (skb == NULL) | 
|  | return NULL; | 
|  |  | 
|  | /* We need the data buffer to be aligned properly.  We will reserve | 
|  | * as many bytes as needed to align the data properly | 
|  | */ | 
|  | skb_reserve(skb, | 
|  | UCC_GETH_RX_DATA_BUF_ALIGNMENT - | 
|  | (((unsigned)skb->data) & (UCC_GETH_RX_DATA_BUF_ALIGNMENT - | 
|  | 1))); | 
|  |  | 
|  | skb->dev = ugeth->dev; | 
|  |  | 
|  | out_be32(&((struct qe_bd *)bd)->buf, | 
|  | dma_map_single(NULL, | 
|  | skb->data, | 
|  | ugeth->ug_info->uf_info.max_rx_buf_length + | 
|  | UCC_GETH_RX_DATA_BUF_ALIGNMENT, | 
|  | DMA_FROM_DEVICE)); | 
|  |  | 
|  | out_be32((u32 *)bd, (R_E | R_I | (in_be32((u32 *)bd) & R_W))); | 
|  |  | 
|  | return skb; | 
|  | } | 
|  |  | 
|  | static int rx_bd_buffer_set(struct ucc_geth_private *ugeth, u8 rxQ) | 
|  | { | 
|  | u8 *bd; | 
|  | u32 bd_status; | 
|  | struct sk_buff *skb; | 
|  | int i; | 
|  |  | 
|  | bd = ugeth->p_rx_bd_ring[rxQ]; | 
|  | i = 0; | 
|  |  | 
|  | do { | 
|  | bd_status = in_be32((u32*)bd); | 
|  | skb = get_new_skb(ugeth, bd); | 
|  |  | 
|  | if (!skb)	/* If can not allocate data buffer, | 
|  | abort. Cleanup will be elsewhere */ | 
|  | return -ENOMEM; | 
|  |  | 
|  | ugeth->rx_skbuff[rxQ][i] = skb; | 
|  |  | 
|  | /* advance the BD pointer */ | 
|  | bd += sizeof(struct qe_bd); | 
|  | i++; | 
|  | } while (!(bd_status & R_W)); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int fill_init_enet_entries(struct ucc_geth_private *ugeth, | 
|  | volatile u32 *p_start, | 
|  | u8 num_entries, | 
|  | u32 thread_size, | 
|  | u32 thread_alignment, | 
|  | enum qe_risc_allocation risc, | 
|  | int skip_page_for_first_entry) | 
|  | { | 
|  | u32 init_enet_offset; | 
|  | u8 i; | 
|  | int snum; | 
|  |  | 
|  | for (i = 0; i < num_entries; i++) { | 
|  | if ((snum = qe_get_snum()) < 0) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("fill_init_enet_entries: Can not get SNUM."); | 
|  | return snum; | 
|  | } | 
|  | if ((i == 0) && skip_page_for_first_entry) | 
|  | /* First entry of Rx does not have page */ | 
|  | init_enet_offset = 0; | 
|  | else { | 
|  | init_enet_offset = | 
|  | qe_muram_alloc(thread_size, thread_alignment); | 
|  | if (IS_ERR_VALUE(init_enet_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("fill_init_enet_entries: Can not allocate DPRAM memory."); | 
|  | qe_put_snum((u8) snum); | 
|  | return -ENOMEM; | 
|  | } | 
|  | } | 
|  | *(p_start++) = | 
|  | ((u8) snum << ENET_INIT_PARAM_SNUM_SHIFT) | init_enet_offset | 
|  | | risc; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int return_init_enet_entries(struct ucc_geth_private *ugeth, | 
|  | volatile u32 *p_start, | 
|  | u8 num_entries, | 
|  | enum qe_risc_allocation risc, | 
|  | int skip_page_for_first_entry) | 
|  | { | 
|  | u32 init_enet_offset; | 
|  | u8 i; | 
|  | int snum; | 
|  |  | 
|  | for (i = 0; i < num_entries; i++) { | 
|  | /* Check that this entry was actually valid -- | 
|  | needed in case failed in allocations */ | 
|  | if ((*p_start & ENET_INIT_PARAM_RISC_MASK) == risc) { | 
|  | snum = | 
|  | (u32) (*p_start & ENET_INIT_PARAM_SNUM_MASK) >> | 
|  | ENET_INIT_PARAM_SNUM_SHIFT; | 
|  | qe_put_snum((u8) snum); | 
|  | if (!((i == 0) && skip_page_for_first_entry)) { | 
|  | /* First entry of Rx does not have page */ | 
|  | init_enet_offset = | 
|  | (in_be32(p_start) & | 
|  | ENET_INIT_PARAM_PTR_MASK); | 
|  | qe_muram_free(init_enet_offset); | 
|  | } | 
|  | *(p_start++) = 0;	/* Just for cosmetics */ | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | #ifdef DEBUG | 
|  | static int dump_init_enet_entries(struct ucc_geth_private *ugeth, | 
|  | volatile u32 *p_start, | 
|  | u8 num_entries, | 
|  | u32 thread_size, | 
|  | enum qe_risc_allocation risc, | 
|  | int skip_page_for_first_entry) | 
|  | { | 
|  | u32 init_enet_offset; | 
|  | u8 i; | 
|  | int snum; | 
|  |  | 
|  | for (i = 0; i < num_entries; i++) { | 
|  | /* Check that this entry was actually valid -- | 
|  | needed in case failed in allocations */ | 
|  | if ((*p_start & ENET_INIT_PARAM_RISC_MASK) == risc) { | 
|  | snum = | 
|  | (u32) (*p_start & ENET_INIT_PARAM_SNUM_MASK) >> | 
|  | ENET_INIT_PARAM_SNUM_SHIFT; | 
|  | qe_put_snum((u8) snum); | 
|  | if (!((i == 0) && skip_page_for_first_entry)) { | 
|  | /* First entry of Rx does not have page */ | 
|  | init_enet_offset = | 
|  | (in_be32(p_start) & | 
|  | ENET_INIT_PARAM_PTR_MASK); | 
|  | ugeth_info("Init enet entry %d:", i); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) | 
|  | qe_muram_addr(init_enet_offset)); | 
|  | mem_disp(qe_muram_addr(init_enet_offset), | 
|  | thread_size); | 
|  | } | 
|  | p_start++; | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  | #endif | 
|  |  | 
|  | #ifdef CONFIG_UGETH_FILTERING | 
|  | static struct enet_addr_container *get_enet_addr_container(void) | 
|  | { | 
|  | struct enet_addr_container *enet_addr_cont; | 
|  |  | 
|  | /* allocate memory */ | 
|  | enet_addr_cont = kmalloc(sizeof(struct enet_addr_container), GFP_KERNEL); | 
|  | if (!enet_addr_cont) { | 
|  | ugeth_err("%s: No memory for enet_addr_container object.", | 
|  | __FUNCTION__); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | return enet_addr_cont; | 
|  | } | 
|  | #endif /* CONFIG_UGETH_FILTERING */ | 
|  |  | 
|  | static void put_enet_addr_container(struct enet_addr_container *enet_addr_cont) | 
|  | { | 
|  | kfree(enet_addr_cont); | 
|  | } | 
|  |  | 
|  | static void set_mac_addr(__be16 __iomem *reg, u8 *mac) | 
|  | { | 
|  | out_be16(®[0], ((u16)mac[5] << 8) | mac[4]); | 
|  | out_be16(®[1], ((u16)mac[3] << 8) | mac[2]); | 
|  | out_be16(®[2], ((u16)mac[1] << 8) | mac[0]); | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_UGETH_FILTERING | 
|  | static int hw_add_addr_in_paddr(struct ucc_geth_private *ugeth, | 
|  | u8 *p_enet_addr, u8 paddr_num) | 
|  | { | 
|  | struct ucc_geth_82xx_address_filtering_pram *p_82xx_addr_filt; | 
|  |  | 
|  | if (!(paddr_num < NUM_OF_PADDRS)) { | 
|  | ugeth_warn("%s: Illegal paddr_num.", __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | p_82xx_addr_filt = | 
|  | (struct ucc_geth_82xx_address_filtering_pram *) ugeth->p_rx_glbl_pram-> | 
|  | addressfiltering; | 
|  |  | 
|  | /* Ethernet frames are defined in Little Endian mode,    */ | 
|  | /* therefore to insert the address we reverse the bytes. */ | 
|  | set_mac_addr(&p_82xx_addr_filt->paddr[paddr_num].h, p_enet_addr); | 
|  | return 0; | 
|  | } | 
|  | #endif /* CONFIG_UGETH_FILTERING */ | 
|  |  | 
|  | static int hw_clear_addr_in_paddr(struct ucc_geth_private *ugeth, u8 paddr_num) | 
|  | { | 
|  | struct ucc_geth_82xx_address_filtering_pram *p_82xx_addr_filt; | 
|  |  | 
|  | if (!(paddr_num < NUM_OF_PADDRS)) { | 
|  | ugeth_warn("%s: Illagel paddr_num.", __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | p_82xx_addr_filt = | 
|  | (struct ucc_geth_82xx_address_filtering_pram *) ugeth->p_rx_glbl_pram-> | 
|  | addressfiltering; | 
|  |  | 
|  | /* Writing address ff.ff.ff.ff.ff.ff disables address | 
|  | recognition for this register */ | 
|  | out_be16(&p_82xx_addr_filt->paddr[paddr_num].h, 0xffff); | 
|  | out_be16(&p_82xx_addr_filt->paddr[paddr_num].m, 0xffff); | 
|  | out_be16(&p_82xx_addr_filt->paddr[paddr_num].l, 0xffff); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void hw_add_addr_in_hash(struct ucc_geth_private *ugeth, | 
|  | u8 *p_enet_addr) | 
|  | { | 
|  | struct ucc_geth_82xx_address_filtering_pram *p_82xx_addr_filt; | 
|  | u32 cecr_subblock; | 
|  |  | 
|  | p_82xx_addr_filt = | 
|  | (struct ucc_geth_82xx_address_filtering_pram *) ugeth->p_rx_glbl_pram-> | 
|  | addressfiltering; | 
|  |  | 
|  | cecr_subblock = | 
|  | ucc_fast_get_qe_cr_subblock(ugeth->ug_info->uf_info.ucc_num); | 
|  |  | 
|  | /* Ethernet frames are defined in Little Endian mode, | 
|  | therefor to insert */ | 
|  | /* the address to the hash (Big Endian mode), we reverse the bytes.*/ | 
|  |  | 
|  | set_mac_addr(&p_82xx_addr_filt->taddr.h, p_enet_addr); | 
|  |  | 
|  | qe_issue_cmd(QE_SET_GROUP_ADDRESS, cecr_subblock, | 
|  | QE_CR_PROTOCOL_ETHERNET, 0); | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_UGETH_MAGIC_PACKET | 
|  | static void magic_packet_detection_enable(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | struct ucc_fast_private *uccf; | 
|  | struct ucc_geth *ug_regs; | 
|  | u32 maccfg2, uccm; | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  | ug_regs = ugeth->ug_regs; | 
|  |  | 
|  | /* Enable interrupts for magic packet detection */ | 
|  | uccm = in_be32(uccf->p_uccm); | 
|  | uccm |= UCCE_MPD; | 
|  | out_be32(uccf->p_uccm, uccm); | 
|  |  | 
|  | /* Enable magic packet detection */ | 
|  | maccfg2 = in_be32(&ug_regs->maccfg2); | 
|  | maccfg2 |= MACCFG2_MPE; | 
|  | out_be32(&ug_regs->maccfg2, maccfg2); | 
|  | } | 
|  |  | 
|  | static void magic_packet_detection_disable(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | struct ucc_fast_private *uccf; | 
|  | struct ucc_geth *ug_regs; | 
|  | u32 maccfg2, uccm; | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  | ug_regs = ugeth->ug_regs; | 
|  |  | 
|  | /* Disable interrupts for magic packet detection */ | 
|  | uccm = in_be32(uccf->p_uccm); | 
|  | uccm &= ~UCCE_MPD; | 
|  | out_be32(uccf->p_uccm, uccm); | 
|  |  | 
|  | /* Disable magic packet detection */ | 
|  | maccfg2 = in_be32(&ug_regs->maccfg2); | 
|  | maccfg2 &= ~MACCFG2_MPE; | 
|  | out_be32(&ug_regs->maccfg2, maccfg2); | 
|  | } | 
|  | #endif /* MAGIC_PACKET */ | 
|  |  | 
|  | static inline int compare_addr(u8 **addr1, u8 **addr2) | 
|  | { | 
|  | return memcmp(addr1, addr2, ENET_NUM_OCTETS_PER_ADDRESS); | 
|  | } | 
|  |  | 
|  | #ifdef DEBUG | 
|  | static void get_statistics(struct ucc_geth_private *ugeth, | 
|  | struct ucc_geth_tx_firmware_statistics * | 
|  | tx_firmware_statistics, | 
|  | struct ucc_geth_rx_firmware_statistics * | 
|  | rx_firmware_statistics, | 
|  | struct ucc_geth_hardware_statistics *hardware_statistics) | 
|  | { | 
|  | struct ucc_fast *uf_regs; | 
|  | struct ucc_geth *ug_regs; | 
|  | struct ucc_geth_tx_firmware_statistics_pram *p_tx_fw_statistics_pram; | 
|  | struct ucc_geth_rx_firmware_statistics_pram *p_rx_fw_statistics_pram; | 
|  |  | 
|  | ug_regs = ugeth->ug_regs; | 
|  | uf_regs = (struct ucc_fast *) ug_regs; | 
|  | p_tx_fw_statistics_pram = ugeth->p_tx_fw_statistics_pram; | 
|  | p_rx_fw_statistics_pram = ugeth->p_rx_fw_statistics_pram; | 
|  |  | 
|  | /* Tx firmware only if user handed pointer and driver actually | 
|  | gathers Tx firmware statistics */ | 
|  | if (tx_firmware_statistics && p_tx_fw_statistics_pram) { | 
|  | tx_firmware_statistics->sicoltx = | 
|  | in_be32(&p_tx_fw_statistics_pram->sicoltx); | 
|  | tx_firmware_statistics->mulcoltx = | 
|  | in_be32(&p_tx_fw_statistics_pram->mulcoltx); | 
|  | tx_firmware_statistics->latecoltxfr = | 
|  | in_be32(&p_tx_fw_statistics_pram->latecoltxfr); | 
|  | tx_firmware_statistics->frabortduecol = | 
|  | in_be32(&p_tx_fw_statistics_pram->frabortduecol); | 
|  | tx_firmware_statistics->frlostinmactxer = | 
|  | in_be32(&p_tx_fw_statistics_pram->frlostinmactxer); | 
|  | tx_firmware_statistics->carriersenseertx = | 
|  | in_be32(&p_tx_fw_statistics_pram->carriersenseertx); | 
|  | tx_firmware_statistics->frtxok = | 
|  | in_be32(&p_tx_fw_statistics_pram->frtxok); | 
|  | tx_firmware_statistics->txfrexcessivedefer = | 
|  | in_be32(&p_tx_fw_statistics_pram->txfrexcessivedefer); | 
|  | tx_firmware_statistics->txpkts256 = | 
|  | in_be32(&p_tx_fw_statistics_pram->txpkts256); | 
|  | tx_firmware_statistics->txpkts512 = | 
|  | in_be32(&p_tx_fw_statistics_pram->txpkts512); | 
|  | tx_firmware_statistics->txpkts1024 = | 
|  | in_be32(&p_tx_fw_statistics_pram->txpkts1024); | 
|  | tx_firmware_statistics->txpktsjumbo = | 
|  | in_be32(&p_tx_fw_statistics_pram->txpktsjumbo); | 
|  | } | 
|  |  | 
|  | /* Rx firmware only if user handed pointer and driver actually | 
|  | * gathers Rx firmware statistics */ | 
|  | if (rx_firmware_statistics && p_rx_fw_statistics_pram) { | 
|  | int i; | 
|  | rx_firmware_statistics->frrxfcser = | 
|  | in_be32(&p_rx_fw_statistics_pram->frrxfcser); | 
|  | rx_firmware_statistics->fraligner = | 
|  | in_be32(&p_rx_fw_statistics_pram->fraligner); | 
|  | rx_firmware_statistics->inrangelenrxer = | 
|  | in_be32(&p_rx_fw_statistics_pram->inrangelenrxer); | 
|  | rx_firmware_statistics->outrangelenrxer = | 
|  | in_be32(&p_rx_fw_statistics_pram->outrangelenrxer); | 
|  | rx_firmware_statistics->frtoolong = | 
|  | in_be32(&p_rx_fw_statistics_pram->frtoolong); | 
|  | rx_firmware_statistics->runt = | 
|  | in_be32(&p_rx_fw_statistics_pram->runt); | 
|  | rx_firmware_statistics->verylongevent = | 
|  | in_be32(&p_rx_fw_statistics_pram->verylongevent); | 
|  | rx_firmware_statistics->symbolerror = | 
|  | in_be32(&p_rx_fw_statistics_pram->symbolerror); | 
|  | rx_firmware_statistics->dropbsy = | 
|  | in_be32(&p_rx_fw_statistics_pram->dropbsy); | 
|  | for (i = 0; i < 0x8; i++) | 
|  | rx_firmware_statistics->res0[i] = | 
|  | p_rx_fw_statistics_pram->res0[i]; | 
|  | rx_firmware_statistics->mismatchdrop = | 
|  | in_be32(&p_rx_fw_statistics_pram->mismatchdrop); | 
|  | rx_firmware_statistics->underpkts = | 
|  | in_be32(&p_rx_fw_statistics_pram->underpkts); | 
|  | rx_firmware_statistics->pkts256 = | 
|  | in_be32(&p_rx_fw_statistics_pram->pkts256); | 
|  | rx_firmware_statistics->pkts512 = | 
|  | in_be32(&p_rx_fw_statistics_pram->pkts512); | 
|  | rx_firmware_statistics->pkts1024 = | 
|  | in_be32(&p_rx_fw_statistics_pram->pkts1024); | 
|  | rx_firmware_statistics->pktsjumbo = | 
|  | in_be32(&p_rx_fw_statistics_pram->pktsjumbo); | 
|  | rx_firmware_statistics->frlossinmacer = | 
|  | in_be32(&p_rx_fw_statistics_pram->frlossinmacer); | 
|  | rx_firmware_statistics->pausefr = | 
|  | in_be32(&p_rx_fw_statistics_pram->pausefr); | 
|  | for (i = 0; i < 0x4; i++) | 
|  | rx_firmware_statistics->res1[i] = | 
|  | p_rx_fw_statistics_pram->res1[i]; | 
|  | rx_firmware_statistics->removevlan = | 
|  | in_be32(&p_rx_fw_statistics_pram->removevlan); | 
|  | rx_firmware_statistics->replacevlan = | 
|  | in_be32(&p_rx_fw_statistics_pram->replacevlan); | 
|  | rx_firmware_statistics->insertvlan = | 
|  | in_be32(&p_rx_fw_statistics_pram->insertvlan); | 
|  | } | 
|  |  | 
|  | /* Hardware only if user handed pointer and driver actually | 
|  | gathers hardware statistics */ | 
|  | if (hardware_statistics && (in_be32(&uf_regs->upsmr) & UPSMR_HSE)) { | 
|  | hardware_statistics->tx64 = in_be32(&ug_regs->tx64); | 
|  | hardware_statistics->tx127 = in_be32(&ug_regs->tx127); | 
|  | hardware_statistics->tx255 = in_be32(&ug_regs->tx255); | 
|  | hardware_statistics->rx64 = in_be32(&ug_regs->rx64); | 
|  | hardware_statistics->rx127 = in_be32(&ug_regs->rx127); | 
|  | hardware_statistics->rx255 = in_be32(&ug_regs->rx255); | 
|  | hardware_statistics->txok = in_be32(&ug_regs->txok); | 
|  | hardware_statistics->txcf = in_be16(&ug_regs->txcf); | 
|  | hardware_statistics->tmca = in_be32(&ug_regs->tmca); | 
|  | hardware_statistics->tbca = in_be32(&ug_regs->tbca); | 
|  | hardware_statistics->rxfok = in_be32(&ug_regs->rxfok); | 
|  | hardware_statistics->rxbok = in_be32(&ug_regs->rxbok); | 
|  | hardware_statistics->rbyt = in_be32(&ug_regs->rbyt); | 
|  | hardware_statistics->rmca = in_be32(&ug_regs->rmca); | 
|  | hardware_statistics->rbca = in_be32(&ug_regs->rbca); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void dump_bds(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | int i; | 
|  | int length; | 
|  |  | 
|  | for (i = 0; i < ugeth->ug_info->numQueuesTx; i++) { | 
|  | if (ugeth->p_tx_bd_ring[i]) { | 
|  | length = | 
|  | (ugeth->ug_info->bdRingLenTx[i] * | 
|  | sizeof(struct qe_bd)); | 
|  | ugeth_info("TX BDs[%d]", i); | 
|  | mem_disp(ugeth->p_tx_bd_ring[i], length); | 
|  | } | 
|  | } | 
|  | for (i = 0; i < ugeth->ug_info->numQueuesRx; i++) { | 
|  | if (ugeth->p_rx_bd_ring[i]) { | 
|  | length = | 
|  | (ugeth->ug_info->bdRingLenRx[i] * | 
|  | sizeof(struct qe_bd)); | 
|  | ugeth_info("RX BDs[%d]", i); | 
|  | mem_disp(ugeth->p_rx_bd_ring[i], length); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | static void dump_regs(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | int i; | 
|  |  | 
|  | ugeth_info("UCC%d Geth registers:", ugeth->ug_info->uf_info.ucc_num); | 
|  | ugeth_info("Base address: 0x%08x", (u32) ugeth->ug_regs); | 
|  |  | 
|  | ugeth_info("maccfg1    : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->maccfg1, | 
|  | in_be32(&ugeth->ug_regs->maccfg1)); | 
|  | ugeth_info("maccfg2    : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->maccfg2, | 
|  | in_be32(&ugeth->ug_regs->maccfg2)); | 
|  | ugeth_info("ipgifg     : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->ipgifg, | 
|  | in_be32(&ugeth->ug_regs->ipgifg)); | 
|  | ugeth_info("hafdup     : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->hafdup, | 
|  | in_be32(&ugeth->ug_regs->hafdup)); | 
|  | ugeth_info("ifctl      : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->ifctl, | 
|  | in_be32(&ugeth->ug_regs->ifctl)); | 
|  | ugeth_info("ifstat     : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->ifstat, | 
|  | in_be32(&ugeth->ug_regs->ifstat)); | 
|  | ugeth_info("macstnaddr1: addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->macstnaddr1, | 
|  | in_be32(&ugeth->ug_regs->macstnaddr1)); | 
|  | ugeth_info("macstnaddr2: addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->macstnaddr2, | 
|  | in_be32(&ugeth->ug_regs->macstnaddr2)); | 
|  | ugeth_info("uempr      : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->uempr, | 
|  | in_be32(&ugeth->ug_regs->uempr)); | 
|  | ugeth_info("utbipar    : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->utbipar, | 
|  | in_be32(&ugeth->ug_regs->utbipar)); | 
|  | ugeth_info("uescr      : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->ug_regs->uescr, | 
|  | in_be16(&ugeth->ug_regs->uescr)); | 
|  | ugeth_info("tx64       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->tx64, | 
|  | in_be32(&ugeth->ug_regs->tx64)); | 
|  | ugeth_info("tx127      : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->tx127, | 
|  | in_be32(&ugeth->ug_regs->tx127)); | 
|  | ugeth_info("tx255      : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->tx255, | 
|  | in_be32(&ugeth->ug_regs->tx255)); | 
|  | ugeth_info("rx64       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->rx64, | 
|  | in_be32(&ugeth->ug_regs->rx64)); | 
|  | ugeth_info("rx127      : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->rx127, | 
|  | in_be32(&ugeth->ug_regs->rx127)); | 
|  | ugeth_info("rx255      : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->rx255, | 
|  | in_be32(&ugeth->ug_regs->rx255)); | 
|  | ugeth_info("txok       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->txok, | 
|  | in_be32(&ugeth->ug_regs->txok)); | 
|  | ugeth_info("txcf       : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->ug_regs->txcf, | 
|  | in_be16(&ugeth->ug_regs->txcf)); | 
|  | ugeth_info("tmca       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->tmca, | 
|  | in_be32(&ugeth->ug_regs->tmca)); | 
|  | ugeth_info("tbca       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->tbca, | 
|  | in_be32(&ugeth->ug_regs->tbca)); | 
|  | ugeth_info("rxfok      : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->rxfok, | 
|  | in_be32(&ugeth->ug_regs->rxfok)); | 
|  | ugeth_info("rxbok      : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->rxbok, | 
|  | in_be32(&ugeth->ug_regs->rxbok)); | 
|  | ugeth_info("rbyt       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->rbyt, | 
|  | in_be32(&ugeth->ug_regs->rbyt)); | 
|  | ugeth_info("rmca       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->rmca, | 
|  | in_be32(&ugeth->ug_regs->rmca)); | 
|  | ugeth_info("rbca       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->rbca, | 
|  | in_be32(&ugeth->ug_regs->rbca)); | 
|  | ugeth_info("scar       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->scar, | 
|  | in_be32(&ugeth->ug_regs->scar)); | 
|  | ugeth_info("scam       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->ug_regs->scam, | 
|  | in_be32(&ugeth->ug_regs->scam)); | 
|  |  | 
|  | if (ugeth->p_thread_data_tx) { | 
|  | int numThreadsTxNumerical; | 
|  | switch (ugeth->ug_info->numThreadsTx) { | 
|  | case UCC_GETH_NUM_OF_THREADS_1: | 
|  | numThreadsTxNumerical = 1; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_2: | 
|  | numThreadsTxNumerical = 2; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_4: | 
|  | numThreadsTxNumerical = 4; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_6: | 
|  | numThreadsTxNumerical = 6; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_8: | 
|  | numThreadsTxNumerical = 8; | 
|  | break; | 
|  | default: | 
|  | numThreadsTxNumerical = 0; | 
|  | break; | 
|  | } | 
|  |  | 
|  | ugeth_info("Thread data TXs:"); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) ugeth->p_thread_data_tx); | 
|  | for (i = 0; i < numThreadsTxNumerical; i++) { | 
|  | ugeth_info("Thread data TX[%d]:", i); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) & ugeth->p_thread_data_tx[i]); | 
|  | mem_disp((u8 *) & ugeth->p_thread_data_tx[i], | 
|  | sizeof(struct ucc_geth_thread_data_tx)); | 
|  | } | 
|  | } | 
|  | if (ugeth->p_thread_data_rx) { | 
|  | int numThreadsRxNumerical; | 
|  | switch (ugeth->ug_info->numThreadsRx) { | 
|  | case UCC_GETH_NUM_OF_THREADS_1: | 
|  | numThreadsRxNumerical = 1; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_2: | 
|  | numThreadsRxNumerical = 2; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_4: | 
|  | numThreadsRxNumerical = 4; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_6: | 
|  | numThreadsRxNumerical = 6; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_8: | 
|  | numThreadsRxNumerical = 8; | 
|  | break; | 
|  | default: | 
|  | numThreadsRxNumerical = 0; | 
|  | break; | 
|  | } | 
|  |  | 
|  | ugeth_info("Thread data RX:"); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) ugeth->p_thread_data_rx); | 
|  | for (i = 0; i < numThreadsRxNumerical; i++) { | 
|  | ugeth_info("Thread data RX[%d]:", i); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) & ugeth->p_thread_data_rx[i]); | 
|  | mem_disp((u8 *) & ugeth->p_thread_data_rx[i], | 
|  | sizeof(struct ucc_geth_thread_data_rx)); | 
|  | } | 
|  | } | 
|  | if (ugeth->p_exf_glbl_param) { | 
|  | ugeth_info("EXF global param:"); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) ugeth->p_exf_glbl_param); | 
|  | mem_disp((u8 *) ugeth->p_exf_glbl_param, | 
|  | sizeof(*ugeth->p_exf_glbl_param)); | 
|  | } | 
|  | if (ugeth->p_tx_glbl_pram) { | 
|  | ugeth_info("TX global param:"); | 
|  | ugeth_info("Base address: 0x%08x", (u32) ugeth->p_tx_glbl_pram); | 
|  | ugeth_info("temoder      : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->temoder, | 
|  | in_be16(&ugeth->p_tx_glbl_pram->temoder)); | 
|  | ugeth_info("sqptr        : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->sqptr, | 
|  | in_be32(&ugeth->p_tx_glbl_pram->sqptr)); | 
|  | ugeth_info("schedulerbasepointer: addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->schedulerbasepointer, | 
|  | in_be32(&ugeth->p_tx_glbl_pram-> | 
|  | schedulerbasepointer)); | 
|  | ugeth_info("txrmonbaseptr: addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->txrmonbaseptr, | 
|  | in_be32(&ugeth->p_tx_glbl_pram->txrmonbaseptr)); | 
|  | ugeth_info("tstate       : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->tstate, | 
|  | in_be32(&ugeth->p_tx_glbl_pram->tstate)); | 
|  | ugeth_info("iphoffset[0] : addr - 0x%08x, val - 0x%02x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->iphoffset[0], | 
|  | ugeth->p_tx_glbl_pram->iphoffset[0]); | 
|  | ugeth_info("iphoffset[1] : addr - 0x%08x, val - 0x%02x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->iphoffset[1], | 
|  | ugeth->p_tx_glbl_pram->iphoffset[1]); | 
|  | ugeth_info("iphoffset[2] : addr - 0x%08x, val - 0x%02x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->iphoffset[2], | 
|  | ugeth->p_tx_glbl_pram->iphoffset[2]); | 
|  | ugeth_info("iphoffset[3] : addr - 0x%08x, val - 0x%02x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->iphoffset[3], | 
|  | ugeth->p_tx_glbl_pram->iphoffset[3]); | 
|  | ugeth_info("iphoffset[4] : addr - 0x%08x, val - 0x%02x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->iphoffset[4], | 
|  | ugeth->p_tx_glbl_pram->iphoffset[4]); | 
|  | ugeth_info("iphoffset[5] : addr - 0x%08x, val - 0x%02x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->iphoffset[5], | 
|  | ugeth->p_tx_glbl_pram->iphoffset[5]); | 
|  | ugeth_info("iphoffset[6] : addr - 0x%08x, val - 0x%02x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->iphoffset[6], | 
|  | ugeth->p_tx_glbl_pram->iphoffset[6]); | 
|  | ugeth_info("iphoffset[7] : addr - 0x%08x, val - 0x%02x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->iphoffset[7], | 
|  | ugeth->p_tx_glbl_pram->iphoffset[7]); | 
|  | ugeth_info("vtagtable[0] : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->vtagtable[0], | 
|  | in_be32(&ugeth->p_tx_glbl_pram->vtagtable[0])); | 
|  | ugeth_info("vtagtable[1] : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->vtagtable[1], | 
|  | in_be32(&ugeth->p_tx_glbl_pram->vtagtable[1])); | 
|  | ugeth_info("vtagtable[2] : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->vtagtable[2], | 
|  | in_be32(&ugeth->p_tx_glbl_pram->vtagtable[2])); | 
|  | ugeth_info("vtagtable[3] : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->vtagtable[3], | 
|  | in_be32(&ugeth->p_tx_glbl_pram->vtagtable[3])); | 
|  | ugeth_info("vtagtable[4] : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->vtagtable[4], | 
|  | in_be32(&ugeth->p_tx_glbl_pram->vtagtable[4])); | 
|  | ugeth_info("vtagtable[5] : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->vtagtable[5], | 
|  | in_be32(&ugeth->p_tx_glbl_pram->vtagtable[5])); | 
|  | ugeth_info("vtagtable[6] : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->vtagtable[6], | 
|  | in_be32(&ugeth->p_tx_glbl_pram->vtagtable[6])); | 
|  | ugeth_info("vtagtable[7] : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->vtagtable[7], | 
|  | in_be32(&ugeth->p_tx_glbl_pram->vtagtable[7])); | 
|  | ugeth_info("tqptr        : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_tx_glbl_pram->tqptr, | 
|  | in_be32(&ugeth->p_tx_glbl_pram->tqptr)); | 
|  | } | 
|  | if (ugeth->p_rx_glbl_pram) { | 
|  | ugeth_info("RX global param:"); | 
|  | ugeth_info("Base address: 0x%08x", (u32) ugeth->p_rx_glbl_pram); | 
|  | ugeth_info("remoder         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->remoder, | 
|  | in_be32(&ugeth->p_rx_glbl_pram->remoder)); | 
|  | ugeth_info("rqptr           : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->rqptr, | 
|  | in_be32(&ugeth->p_rx_glbl_pram->rqptr)); | 
|  | ugeth_info("typeorlen       : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->typeorlen, | 
|  | in_be16(&ugeth->p_rx_glbl_pram->typeorlen)); | 
|  | ugeth_info("rxgstpack       : addr - 0x%08x, val - 0x%02x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->rxgstpack, | 
|  | ugeth->p_rx_glbl_pram->rxgstpack); | 
|  | ugeth_info("rxrmonbaseptr   : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->rxrmonbaseptr, | 
|  | in_be32(&ugeth->p_rx_glbl_pram->rxrmonbaseptr)); | 
|  | ugeth_info("intcoalescingptr: addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->intcoalescingptr, | 
|  | in_be32(&ugeth->p_rx_glbl_pram->intcoalescingptr)); | 
|  | ugeth_info("rstate          : addr - 0x%08x, val - 0x%02x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->rstate, | 
|  | ugeth->p_rx_glbl_pram->rstate); | 
|  | ugeth_info("mrblr           : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->mrblr, | 
|  | in_be16(&ugeth->p_rx_glbl_pram->mrblr)); | 
|  | ugeth_info("rbdqptr         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->rbdqptr, | 
|  | in_be32(&ugeth->p_rx_glbl_pram->rbdqptr)); | 
|  | ugeth_info("mflr            : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->mflr, | 
|  | in_be16(&ugeth->p_rx_glbl_pram->mflr)); | 
|  | ugeth_info("minflr          : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->minflr, | 
|  | in_be16(&ugeth->p_rx_glbl_pram->minflr)); | 
|  | ugeth_info("maxd1           : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->maxd1, | 
|  | in_be16(&ugeth->p_rx_glbl_pram->maxd1)); | 
|  | ugeth_info("maxd2           : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->maxd2, | 
|  | in_be16(&ugeth->p_rx_glbl_pram->maxd2)); | 
|  | ugeth_info("ecamptr         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->ecamptr, | 
|  | in_be32(&ugeth->p_rx_glbl_pram->ecamptr)); | 
|  | ugeth_info("l2qt            : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->l2qt, | 
|  | in_be32(&ugeth->p_rx_glbl_pram->l2qt)); | 
|  | ugeth_info("l3qt[0]         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->l3qt[0], | 
|  | in_be32(&ugeth->p_rx_glbl_pram->l3qt[0])); | 
|  | ugeth_info("l3qt[1]         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->l3qt[1], | 
|  | in_be32(&ugeth->p_rx_glbl_pram->l3qt[1])); | 
|  | ugeth_info("l3qt[2]         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->l3qt[2], | 
|  | in_be32(&ugeth->p_rx_glbl_pram->l3qt[2])); | 
|  | ugeth_info("l3qt[3]         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->l3qt[3], | 
|  | in_be32(&ugeth->p_rx_glbl_pram->l3qt[3])); | 
|  | ugeth_info("l3qt[4]         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->l3qt[4], | 
|  | in_be32(&ugeth->p_rx_glbl_pram->l3qt[4])); | 
|  | ugeth_info("l3qt[5]         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->l3qt[5], | 
|  | in_be32(&ugeth->p_rx_glbl_pram->l3qt[5])); | 
|  | ugeth_info("l3qt[6]         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->l3qt[6], | 
|  | in_be32(&ugeth->p_rx_glbl_pram->l3qt[6])); | 
|  | ugeth_info("l3qt[7]         : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->l3qt[7], | 
|  | in_be32(&ugeth->p_rx_glbl_pram->l3qt[7])); | 
|  | ugeth_info("vlantype        : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->vlantype, | 
|  | in_be16(&ugeth->p_rx_glbl_pram->vlantype)); | 
|  | ugeth_info("vlantci         : addr - 0x%08x, val - 0x%04x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->vlantci, | 
|  | in_be16(&ugeth->p_rx_glbl_pram->vlantci)); | 
|  | for (i = 0; i < 64; i++) | 
|  | ugeth_info | 
|  | ("addressfiltering[%d]: addr - 0x%08x, val - 0x%02x", | 
|  | i, | 
|  | (u32) & ugeth->p_rx_glbl_pram->addressfiltering[i], | 
|  | ugeth->p_rx_glbl_pram->addressfiltering[i]); | 
|  | ugeth_info("exfGlobalParam  : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_glbl_pram->exfGlobalParam, | 
|  | in_be32(&ugeth->p_rx_glbl_pram->exfGlobalParam)); | 
|  | } | 
|  | if (ugeth->p_send_q_mem_reg) { | 
|  | ugeth_info("Send Q memory registers:"); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) ugeth->p_send_q_mem_reg); | 
|  | for (i = 0; i < ugeth->ug_info->numQueuesTx; i++) { | 
|  | ugeth_info("SQQD[%d]:", i); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) & ugeth->p_send_q_mem_reg->sqqd[i]); | 
|  | mem_disp((u8 *) & ugeth->p_send_q_mem_reg->sqqd[i], | 
|  | sizeof(struct ucc_geth_send_queue_qd)); | 
|  | } | 
|  | } | 
|  | if (ugeth->p_scheduler) { | 
|  | ugeth_info("Scheduler:"); | 
|  | ugeth_info("Base address: 0x%08x", (u32) ugeth->p_scheduler); | 
|  | mem_disp((u8 *) ugeth->p_scheduler, | 
|  | sizeof(*ugeth->p_scheduler)); | 
|  | } | 
|  | if (ugeth->p_tx_fw_statistics_pram) { | 
|  | ugeth_info("TX FW statistics pram:"); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) ugeth->p_tx_fw_statistics_pram); | 
|  | mem_disp((u8 *) ugeth->p_tx_fw_statistics_pram, | 
|  | sizeof(*ugeth->p_tx_fw_statistics_pram)); | 
|  | } | 
|  | if (ugeth->p_rx_fw_statistics_pram) { | 
|  | ugeth_info("RX FW statistics pram:"); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) ugeth->p_rx_fw_statistics_pram); | 
|  | mem_disp((u8 *) ugeth->p_rx_fw_statistics_pram, | 
|  | sizeof(*ugeth->p_rx_fw_statistics_pram)); | 
|  | } | 
|  | if (ugeth->p_rx_irq_coalescing_tbl) { | 
|  | ugeth_info("RX IRQ coalescing tables:"); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) ugeth->p_rx_irq_coalescing_tbl); | 
|  | for (i = 0; i < ugeth->ug_info->numQueuesRx; i++) { | 
|  | ugeth_info("RX IRQ coalescing table entry[%d]:", i); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) & ugeth->p_rx_irq_coalescing_tbl-> | 
|  | coalescingentry[i]); | 
|  | ugeth_info | 
|  | ("interruptcoalescingmaxvalue: addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_irq_coalescing_tbl-> | 
|  | coalescingentry[i].interruptcoalescingmaxvalue, | 
|  | in_be32(&ugeth->p_rx_irq_coalescing_tbl-> | 
|  | coalescingentry[i]. | 
|  | interruptcoalescingmaxvalue)); | 
|  | ugeth_info | 
|  | ("interruptcoalescingcounter : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_irq_coalescing_tbl-> | 
|  | coalescingentry[i].interruptcoalescingcounter, | 
|  | in_be32(&ugeth->p_rx_irq_coalescing_tbl-> | 
|  | coalescingentry[i]. | 
|  | interruptcoalescingcounter)); | 
|  | } | 
|  | } | 
|  | if (ugeth->p_rx_bd_qs_tbl) { | 
|  | ugeth_info("RX BD QS tables:"); | 
|  | ugeth_info("Base address: 0x%08x", (u32) ugeth->p_rx_bd_qs_tbl); | 
|  | for (i = 0; i < ugeth->ug_info->numQueuesRx; i++) { | 
|  | ugeth_info("RX BD QS table[%d]:", i); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) & ugeth->p_rx_bd_qs_tbl[i]); | 
|  | ugeth_info | 
|  | ("bdbaseptr        : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_bd_qs_tbl[i].bdbaseptr, | 
|  | in_be32(&ugeth->p_rx_bd_qs_tbl[i].bdbaseptr)); | 
|  | ugeth_info | 
|  | ("bdptr            : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_bd_qs_tbl[i].bdptr, | 
|  | in_be32(&ugeth->p_rx_bd_qs_tbl[i].bdptr)); | 
|  | ugeth_info | 
|  | ("externalbdbaseptr: addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_bd_qs_tbl[i].externalbdbaseptr, | 
|  | in_be32(&ugeth->p_rx_bd_qs_tbl[i]. | 
|  | externalbdbaseptr)); | 
|  | ugeth_info | 
|  | ("externalbdptr    : addr - 0x%08x, val - 0x%08x", | 
|  | (u32) & ugeth->p_rx_bd_qs_tbl[i].externalbdptr, | 
|  | in_be32(&ugeth->p_rx_bd_qs_tbl[i].externalbdptr)); | 
|  | ugeth_info("ucode RX Prefetched BDs:"); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) | 
|  | qe_muram_addr(in_be32 | 
|  | (&ugeth->p_rx_bd_qs_tbl[i]. | 
|  | bdbaseptr))); | 
|  | mem_disp((u8 *) | 
|  | qe_muram_addr(in_be32 | 
|  | (&ugeth->p_rx_bd_qs_tbl[i]. | 
|  | bdbaseptr)), | 
|  | sizeof(struct ucc_geth_rx_prefetched_bds)); | 
|  | } | 
|  | } | 
|  | if (ugeth->p_init_enet_param_shadow) { | 
|  | int size; | 
|  | ugeth_info("Init enet param shadow:"); | 
|  | ugeth_info("Base address: 0x%08x", | 
|  | (u32) ugeth->p_init_enet_param_shadow); | 
|  | mem_disp((u8 *) ugeth->p_init_enet_param_shadow, | 
|  | sizeof(*ugeth->p_init_enet_param_shadow)); | 
|  |  | 
|  | size = sizeof(struct ucc_geth_thread_rx_pram); | 
|  | if (ugeth->ug_info->rxExtendedFiltering) { | 
|  | size += | 
|  | THREAD_RX_PRAM_ADDITIONAL_FOR_EXTENDED_FILTERING; | 
|  | if (ugeth->ug_info->largestexternallookupkeysize == | 
|  | QE_FLTR_TABLE_LOOKUP_KEY_SIZE_8_BYTES) | 
|  | size += | 
|  | THREAD_RX_PRAM_ADDITIONAL_FOR_EXTENDED_FILTERING_8; | 
|  | if (ugeth->ug_info->largestexternallookupkeysize == | 
|  | QE_FLTR_TABLE_LOOKUP_KEY_SIZE_16_BYTES) | 
|  | size += | 
|  | THREAD_RX_PRAM_ADDITIONAL_FOR_EXTENDED_FILTERING_16; | 
|  | } | 
|  |  | 
|  | dump_init_enet_entries(ugeth, | 
|  | &(ugeth->p_init_enet_param_shadow-> | 
|  | txthread[0]), | 
|  | ENET_INIT_PARAM_MAX_ENTRIES_TX, | 
|  | sizeof(struct ucc_geth_thread_tx_pram), | 
|  | ugeth->ug_info->riscTx, 0); | 
|  | dump_init_enet_entries(ugeth, | 
|  | &(ugeth->p_init_enet_param_shadow-> | 
|  | rxthread[0]), | 
|  | ENET_INIT_PARAM_MAX_ENTRIES_RX, size, | 
|  | ugeth->ug_info->riscRx, 1); | 
|  | } | 
|  | } | 
|  | #endif /* DEBUG */ | 
|  |  | 
|  | static void init_default_reg_vals(volatile u32 *upsmr_register, | 
|  | volatile u32 *maccfg1_register, | 
|  | volatile u32 *maccfg2_register) | 
|  | { | 
|  | out_be32(upsmr_register, UCC_GETH_UPSMR_INIT); | 
|  | out_be32(maccfg1_register, UCC_GETH_MACCFG1_INIT); | 
|  | out_be32(maccfg2_register, UCC_GETH_MACCFG2_INIT); | 
|  | } | 
|  |  | 
|  | static int init_half_duplex_params(int alt_beb, | 
|  | int back_pressure_no_backoff, | 
|  | int no_backoff, | 
|  | int excess_defer, | 
|  | u8 alt_beb_truncation, | 
|  | u8 max_retransmissions, | 
|  | u8 collision_window, | 
|  | volatile u32 *hafdup_register) | 
|  | { | 
|  | u32 value = 0; | 
|  |  | 
|  | if ((alt_beb_truncation > HALFDUP_ALT_BEB_TRUNCATION_MAX) || | 
|  | (max_retransmissions > HALFDUP_MAX_RETRANSMISSION_MAX) || | 
|  | (collision_window > HALFDUP_COLLISION_WINDOW_MAX)) | 
|  | return -EINVAL; | 
|  |  | 
|  | value = (u32) (alt_beb_truncation << HALFDUP_ALT_BEB_TRUNCATION_SHIFT); | 
|  |  | 
|  | if (alt_beb) | 
|  | value |= HALFDUP_ALT_BEB; | 
|  | if (back_pressure_no_backoff) | 
|  | value |= HALFDUP_BACK_PRESSURE_NO_BACKOFF; | 
|  | if (no_backoff) | 
|  | value |= HALFDUP_NO_BACKOFF; | 
|  | if (excess_defer) | 
|  | value |= HALFDUP_EXCESSIVE_DEFER; | 
|  |  | 
|  | value |= (max_retransmissions << HALFDUP_MAX_RETRANSMISSION_SHIFT); | 
|  |  | 
|  | value |= collision_window; | 
|  |  | 
|  | out_be32(hafdup_register, value); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int init_inter_frame_gap_params(u8 non_btb_cs_ipg, | 
|  | u8 non_btb_ipg, | 
|  | u8 min_ifg, | 
|  | u8 btb_ipg, | 
|  | volatile u32 *ipgifg_register) | 
|  | { | 
|  | u32 value = 0; | 
|  |  | 
|  | /* Non-Back-to-back IPG part 1 should be <= Non-Back-to-back | 
|  | IPG part 2 */ | 
|  | if (non_btb_cs_ipg > non_btb_ipg) | 
|  | return -EINVAL; | 
|  |  | 
|  | if ((non_btb_cs_ipg > IPGIFG_NON_BACK_TO_BACK_IFG_PART1_MAX) || | 
|  | (non_btb_ipg > IPGIFG_NON_BACK_TO_BACK_IFG_PART2_MAX) || | 
|  | /*(min_ifg        > IPGIFG_MINIMUM_IFG_ENFORCEMENT_MAX) || */ | 
|  | (btb_ipg > IPGIFG_BACK_TO_BACK_IFG_MAX)) | 
|  | return -EINVAL; | 
|  |  | 
|  | value |= | 
|  | ((non_btb_cs_ipg << IPGIFG_NON_BACK_TO_BACK_IFG_PART1_SHIFT) & | 
|  | IPGIFG_NBTB_CS_IPG_MASK); | 
|  | value |= | 
|  | ((non_btb_ipg << IPGIFG_NON_BACK_TO_BACK_IFG_PART2_SHIFT) & | 
|  | IPGIFG_NBTB_IPG_MASK); | 
|  | value |= | 
|  | ((min_ifg << IPGIFG_MINIMUM_IFG_ENFORCEMENT_SHIFT) & | 
|  | IPGIFG_MIN_IFG_MASK); | 
|  | value |= (btb_ipg & IPGIFG_BTB_IPG_MASK); | 
|  |  | 
|  | out_be32(ipgifg_register, value); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int init_flow_control_params(u32 automatic_flow_control_mode, | 
|  | int rx_flow_control_enable, | 
|  | int tx_flow_control_enable, | 
|  | u16 pause_period, | 
|  | u16 extension_field, | 
|  | volatile u32 *upsmr_register, | 
|  | volatile u32 *uempr_register, | 
|  | volatile u32 *maccfg1_register) | 
|  | { | 
|  | u32 value = 0; | 
|  |  | 
|  | /* Set UEMPR register */ | 
|  | value = (u32) pause_period << UEMPR_PAUSE_TIME_VALUE_SHIFT; | 
|  | value |= (u32) extension_field << UEMPR_EXTENDED_PAUSE_TIME_VALUE_SHIFT; | 
|  | out_be32(uempr_register, value); | 
|  |  | 
|  | /* Set UPSMR register */ | 
|  | value = in_be32(upsmr_register); | 
|  | value |= automatic_flow_control_mode; | 
|  | out_be32(upsmr_register, value); | 
|  |  | 
|  | value = in_be32(maccfg1_register); | 
|  | if (rx_flow_control_enable) | 
|  | value |= MACCFG1_FLOW_RX; | 
|  | if (tx_flow_control_enable) | 
|  | value |= MACCFG1_FLOW_TX; | 
|  | out_be32(maccfg1_register, value); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int init_hw_statistics_gathering_mode(int enable_hardware_statistics, | 
|  | int auto_zero_hardware_statistics, | 
|  | volatile u32 *upsmr_register, | 
|  | volatile u16 *uescr_register) | 
|  | { | 
|  | u32 upsmr_value = 0; | 
|  | u16 uescr_value = 0; | 
|  | /* Enable hardware statistics gathering if requested */ | 
|  | if (enable_hardware_statistics) { | 
|  | upsmr_value = in_be32(upsmr_register); | 
|  | upsmr_value |= UPSMR_HSE; | 
|  | out_be32(upsmr_register, upsmr_value); | 
|  | } | 
|  |  | 
|  | /* Clear hardware statistics counters */ | 
|  | uescr_value = in_be16(uescr_register); | 
|  | uescr_value |= UESCR_CLRCNT; | 
|  | /* Automatically zero hardware statistics counters on read, | 
|  | if requested */ | 
|  | if (auto_zero_hardware_statistics) | 
|  | uescr_value |= UESCR_AUTOZ; | 
|  | out_be16(uescr_register, uescr_value); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int init_firmware_statistics_gathering_mode(int | 
|  | enable_tx_firmware_statistics, | 
|  | int enable_rx_firmware_statistics, | 
|  | volatile u32 *tx_rmon_base_ptr, | 
|  | u32 tx_firmware_statistics_structure_address, | 
|  | volatile u32 *rx_rmon_base_ptr, | 
|  | u32 rx_firmware_statistics_structure_address, | 
|  | volatile u16 *temoder_register, | 
|  | volatile u32 *remoder_register) | 
|  | { | 
|  | /* Note: this function does not check if */ | 
|  | /* the parameters it receives are NULL   */ | 
|  | u16 temoder_value; | 
|  | u32 remoder_value; | 
|  |  | 
|  | if (enable_tx_firmware_statistics) { | 
|  | out_be32(tx_rmon_base_ptr, | 
|  | tx_firmware_statistics_structure_address); | 
|  | temoder_value = in_be16(temoder_register); | 
|  | temoder_value |= TEMODER_TX_RMON_STATISTICS_ENABLE; | 
|  | out_be16(temoder_register, temoder_value); | 
|  | } | 
|  |  | 
|  | if (enable_rx_firmware_statistics) { | 
|  | out_be32(rx_rmon_base_ptr, | 
|  | rx_firmware_statistics_structure_address); | 
|  | remoder_value = in_be32(remoder_register); | 
|  | remoder_value |= REMODER_RX_RMON_STATISTICS_ENABLE; | 
|  | out_be32(remoder_register, remoder_value); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int init_mac_station_addr_regs(u8 address_byte_0, | 
|  | u8 address_byte_1, | 
|  | u8 address_byte_2, | 
|  | u8 address_byte_3, | 
|  | u8 address_byte_4, | 
|  | u8 address_byte_5, | 
|  | volatile u32 *macstnaddr1_register, | 
|  | volatile u32 *macstnaddr2_register) | 
|  | { | 
|  | u32 value = 0; | 
|  |  | 
|  | /* Example: for a station address of 0x12345678ABCD, */ | 
|  | /* 0x12 is byte 0, 0x34 is byte 1 and so on and 0xCD is byte 5 */ | 
|  |  | 
|  | /* MACSTNADDR1 Register: */ | 
|  |  | 
|  | /* 0                      7   8                      15  */ | 
|  | /* station address byte 5     station address byte 4     */ | 
|  | /* 16                     23  24                     31  */ | 
|  | /* station address byte 3     station address byte 2     */ | 
|  | value |= (u32) ((address_byte_2 << 0) & 0x000000FF); | 
|  | value |= (u32) ((address_byte_3 << 8) & 0x0000FF00); | 
|  | value |= (u32) ((address_byte_4 << 16) & 0x00FF0000); | 
|  | value |= (u32) ((address_byte_5 << 24) & 0xFF000000); | 
|  |  | 
|  | out_be32(macstnaddr1_register, value); | 
|  |  | 
|  | /* MACSTNADDR2 Register: */ | 
|  |  | 
|  | /* 0                      7   8                      15  */ | 
|  | /* station address byte 1     station address byte 0     */ | 
|  | /* 16                     23  24                     31  */ | 
|  | /*         reserved                   reserved           */ | 
|  | value = 0; | 
|  | value |= (u32) ((address_byte_0 << 16) & 0x00FF0000); | 
|  | value |= (u32) ((address_byte_1 << 24) & 0xFF000000); | 
|  |  | 
|  | out_be32(macstnaddr2_register, value); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int init_check_frame_length_mode(int length_check, | 
|  | volatile u32 *maccfg2_register) | 
|  | { | 
|  | u32 value = 0; | 
|  |  | 
|  | value = in_be32(maccfg2_register); | 
|  |  | 
|  | if (length_check) | 
|  | value |= MACCFG2_LC; | 
|  | else | 
|  | value &= ~MACCFG2_LC; | 
|  |  | 
|  | out_be32(maccfg2_register, value); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int init_preamble_length(u8 preamble_length, | 
|  | volatile u32 *maccfg2_register) | 
|  | { | 
|  | u32 value = 0; | 
|  |  | 
|  | if ((preamble_length < 3) || (preamble_length > 7)) | 
|  | return -EINVAL; | 
|  |  | 
|  | value = in_be32(maccfg2_register); | 
|  | value &= ~MACCFG2_PREL_MASK; | 
|  | value |= (preamble_length << MACCFG2_PREL_SHIFT); | 
|  | out_be32(maccfg2_register, value); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int init_rx_parameters(int reject_broadcast, | 
|  | int receive_short_frames, | 
|  | int promiscuous, volatile u32 *upsmr_register) | 
|  | { | 
|  | u32 value = 0; | 
|  |  | 
|  | value = in_be32(upsmr_register); | 
|  |  | 
|  | if (reject_broadcast) | 
|  | value |= UPSMR_BRO; | 
|  | else | 
|  | value &= ~UPSMR_BRO; | 
|  |  | 
|  | if (receive_short_frames) | 
|  | value |= UPSMR_RSH; | 
|  | else | 
|  | value &= ~UPSMR_RSH; | 
|  |  | 
|  | if (promiscuous) | 
|  | value |= UPSMR_PRO; | 
|  | else | 
|  | value &= ~UPSMR_PRO; | 
|  |  | 
|  | out_be32(upsmr_register, value); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int init_max_rx_buff_len(u16 max_rx_buf_len, | 
|  | volatile u16 *mrblr_register) | 
|  | { | 
|  | /* max_rx_buf_len value must be a multiple of 128 */ | 
|  | if ((max_rx_buf_len == 0) | 
|  | || (max_rx_buf_len % UCC_GETH_MRBLR_ALIGNMENT)) | 
|  | return -EINVAL; | 
|  |  | 
|  | out_be16(mrblr_register, max_rx_buf_len); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int init_min_frame_len(u16 min_frame_length, | 
|  | volatile u16 *minflr_register, | 
|  | volatile u16 *mrblr_register) | 
|  | { | 
|  | u16 mrblr_value = 0; | 
|  |  | 
|  | mrblr_value = in_be16(mrblr_register); | 
|  | if (min_frame_length >= (mrblr_value - 4)) | 
|  | return -EINVAL; | 
|  |  | 
|  | out_be16(minflr_register, min_frame_length); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int adjust_enet_interface(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | struct ucc_geth_info *ug_info; | 
|  | struct ucc_geth *ug_regs; | 
|  | struct ucc_fast *uf_regs; | 
|  | int ret_val; | 
|  | u32 upsmr, maccfg2, tbiBaseAddress; | 
|  | u16 value; | 
|  |  | 
|  | ugeth_vdbg("%s: IN", __FUNCTION__); | 
|  |  | 
|  | ug_info = ugeth->ug_info; | 
|  | ug_regs = ugeth->ug_regs; | 
|  | uf_regs = ugeth->uccf->uf_regs; | 
|  |  | 
|  | /*                    Set MACCFG2                    */ | 
|  | maccfg2 = in_be32(&ug_regs->maccfg2); | 
|  | maccfg2 &= ~MACCFG2_INTERFACE_MODE_MASK; | 
|  | if ((ugeth->max_speed == SPEED_10) || | 
|  | (ugeth->max_speed == SPEED_100)) | 
|  | maccfg2 |= MACCFG2_INTERFACE_MODE_NIBBLE; | 
|  | else if (ugeth->max_speed == SPEED_1000) | 
|  | maccfg2 |= MACCFG2_INTERFACE_MODE_BYTE; | 
|  | maccfg2 |= ug_info->padAndCrc; | 
|  | out_be32(&ug_regs->maccfg2, maccfg2); | 
|  |  | 
|  | /*                    Set UPSMR                      */ | 
|  | upsmr = in_be32(&uf_regs->upsmr); | 
|  | upsmr &= ~(UPSMR_RPM | UPSMR_R10M | UPSMR_TBIM | UPSMR_RMM); | 
|  | if ((ugeth->phy_interface == PHY_INTERFACE_MODE_RMII) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RGMII) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RGMII_ID) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RGMII_RXID) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RGMII_TXID) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RTBI)) { | 
|  | upsmr |= UPSMR_RPM; | 
|  | switch (ugeth->max_speed) { | 
|  | case SPEED_10: | 
|  | upsmr |= UPSMR_R10M; | 
|  | /* FALLTHROUGH */ | 
|  | case SPEED_100: | 
|  | if (ugeth->phy_interface != PHY_INTERFACE_MODE_RTBI) | 
|  | upsmr |= UPSMR_RMM; | 
|  | } | 
|  | } | 
|  | if ((ugeth->phy_interface == PHY_INTERFACE_MODE_TBI) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RTBI)) { | 
|  | upsmr |= UPSMR_TBIM; | 
|  | } | 
|  | out_be32(&uf_regs->upsmr, upsmr); | 
|  |  | 
|  | /* Disable autonegotiation in tbi mode, because by default it | 
|  | comes up in autonegotiation mode. */ | 
|  | /* Note that this depends on proper setting in utbipar register. */ | 
|  | if ((ugeth->phy_interface == PHY_INTERFACE_MODE_TBI) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RTBI)) { | 
|  | tbiBaseAddress = in_be32(&ug_regs->utbipar); | 
|  | tbiBaseAddress &= UTBIPAR_PHY_ADDRESS_MASK; | 
|  | tbiBaseAddress >>= UTBIPAR_PHY_ADDRESS_SHIFT; | 
|  | value = ugeth->phydev->bus->read(ugeth->phydev->bus, | 
|  | (u8) tbiBaseAddress, ENET_TBI_MII_CR); | 
|  | value &= ~0x1000;	/* Turn off autonegotiation */ | 
|  | ugeth->phydev->bus->write(ugeth->phydev->bus, | 
|  | (u8) tbiBaseAddress, ENET_TBI_MII_CR, value); | 
|  | } | 
|  |  | 
|  | init_check_frame_length_mode(ug_info->lengthCheckRx, &ug_regs->maccfg2); | 
|  |  | 
|  | ret_val = init_preamble_length(ug_info->prel, &ug_regs->maccfg2); | 
|  | if (ret_val != 0) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err("%s: Preamble length must be between 3 and 7 inclusive.", | 
|  | __FUNCTION__); | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Called every time the controller might need to be made | 
|  | * aware of new link state.  The PHY code conveys this | 
|  | * information through variables in the ugeth structure, and this | 
|  | * function converts those variables into the appropriate | 
|  | * register values, and can bring down the device if needed. | 
|  | */ | 
|  |  | 
|  | static void adjust_link(struct net_device *dev) | 
|  | { | 
|  | struct ucc_geth_private *ugeth = netdev_priv(dev); | 
|  | struct ucc_geth *ug_regs; | 
|  | struct ucc_fast *uf_regs; | 
|  | struct phy_device *phydev = ugeth->phydev; | 
|  | unsigned long flags; | 
|  | int new_state = 0; | 
|  |  | 
|  | ug_regs = ugeth->ug_regs; | 
|  | uf_regs = ugeth->uccf->uf_regs; | 
|  |  | 
|  | spin_lock_irqsave(&ugeth->lock, flags); | 
|  |  | 
|  | if (phydev->link) { | 
|  | u32 tempval = in_be32(&ug_regs->maccfg2); | 
|  | u32 upsmr = in_be32(&uf_regs->upsmr); | 
|  | /* Now we make sure that we can be in full duplex mode. | 
|  | * If not, we operate in half-duplex mode. */ | 
|  | if (phydev->duplex != ugeth->oldduplex) { | 
|  | new_state = 1; | 
|  | if (!(phydev->duplex)) | 
|  | tempval &= ~(MACCFG2_FDX); | 
|  | else | 
|  | tempval |= MACCFG2_FDX; | 
|  | ugeth->oldduplex = phydev->duplex; | 
|  | } | 
|  |  | 
|  | if (phydev->speed != ugeth->oldspeed) { | 
|  | new_state = 1; | 
|  | switch (phydev->speed) { | 
|  | case SPEED_1000: | 
|  | tempval = ((tempval & | 
|  | ~(MACCFG2_INTERFACE_MODE_MASK)) | | 
|  | MACCFG2_INTERFACE_MODE_BYTE); | 
|  | break; | 
|  | case SPEED_100: | 
|  | case SPEED_10: | 
|  | tempval = ((tempval & | 
|  | ~(MACCFG2_INTERFACE_MODE_MASK)) | | 
|  | MACCFG2_INTERFACE_MODE_NIBBLE); | 
|  | /* if reduced mode, re-set UPSMR.R10M */ | 
|  | if ((ugeth->phy_interface == PHY_INTERFACE_MODE_RMII) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RGMII) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RGMII_ID) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RGMII_RXID) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RGMII_TXID) || | 
|  | (ugeth->phy_interface == PHY_INTERFACE_MODE_RTBI)) { | 
|  | if (phydev->speed == SPEED_10) | 
|  | upsmr |= UPSMR_R10M; | 
|  | else | 
|  | upsmr &= ~(UPSMR_R10M); | 
|  | } | 
|  | break; | 
|  | default: | 
|  | if (netif_msg_link(ugeth)) | 
|  | ugeth_warn( | 
|  | "%s: Ack!  Speed (%d) is not 10/100/1000!", | 
|  | dev->name, phydev->speed); | 
|  | break; | 
|  | } | 
|  | ugeth->oldspeed = phydev->speed; | 
|  | } | 
|  |  | 
|  | out_be32(&ug_regs->maccfg2, tempval); | 
|  | out_be32(&uf_regs->upsmr, upsmr); | 
|  |  | 
|  | if (!ugeth->oldlink) { | 
|  | new_state = 1; | 
|  | ugeth->oldlink = 1; | 
|  | netif_schedule(dev); | 
|  | } | 
|  | } else if (ugeth->oldlink) { | 
|  | new_state = 1; | 
|  | ugeth->oldlink = 0; | 
|  | ugeth->oldspeed = 0; | 
|  | ugeth->oldduplex = -1; | 
|  | } | 
|  |  | 
|  | if (new_state && netif_msg_link(ugeth)) | 
|  | phy_print_status(phydev); | 
|  |  | 
|  | spin_unlock_irqrestore(&ugeth->lock, flags); | 
|  | } | 
|  |  | 
|  | /* Configure the PHY for dev. | 
|  | * returns 0 if success.  -1 if failure | 
|  | */ | 
|  | static int init_phy(struct net_device *dev) | 
|  | { | 
|  | struct ucc_geth_private *priv = netdev_priv(dev); | 
|  | struct phy_device *phydev; | 
|  | char phy_id[BUS_ID_SIZE]; | 
|  |  | 
|  | priv->oldlink = 0; | 
|  | priv->oldspeed = 0; | 
|  | priv->oldduplex = -1; | 
|  |  | 
|  | snprintf(phy_id, BUS_ID_SIZE, PHY_ID_FMT, priv->ug_info->mdio_bus, | 
|  | priv->ug_info->phy_address); | 
|  |  | 
|  | phydev = phy_connect(dev, phy_id, &adjust_link, 0, priv->phy_interface); | 
|  |  | 
|  | if (IS_ERR(phydev)) { | 
|  | printk("%s: Could not attach to PHY\n", dev->name); | 
|  | return PTR_ERR(phydev); | 
|  | } | 
|  |  | 
|  | phydev->supported &= (ADVERTISED_10baseT_Half | | 
|  | ADVERTISED_10baseT_Full | | 
|  | ADVERTISED_100baseT_Half | | 
|  | ADVERTISED_100baseT_Full); | 
|  |  | 
|  | if (priv->max_speed == SPEED_1000) | 
|  | phydev->supported |= ADVERTISED_1000baseT_Full; | 
|  |  | 
|  | phydev->advertising = phydev->supported; | 
|  |  | 
|  | priv->phydev = phydev; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  |  | 
|  |  | 
|  | static int ugeth_graceful_stop_tx(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | struct ucc_fast_private *uccf; | 
|  | u32 cecr_subblock; | 
|  | u32 temp; | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  |  | 
|  | /* Mask GRACEFUL STOP TX interrupt bit and clear it */ | 
|  | temp = in_be32(uccf->p_uccm); | 
|  | temp &= ~UCCE_GRA; | 
|  | out_be32(uccf->p_uccm, temp); | 
|  | out_be32(uccf->p_ucce, UCCE_GRA);	/* clear by writing 1 */ | 
|  |  | 
|  | /* Issue host command */ | 
|  | cecr_subblock = | 
|  | ucc_fast_get_qe_cr_subblock(ugeth->ug_info->uf_info.ucc_num); | 
|  | qe_issue_cmd(QE_GRACEFUL_STOP_TX, cecr_subblock, | 
|  | QE_CR_PROTOCOL_ETHERNET, 0); | 
|  |  | 
|  | /* Wait for command to complete */ | 
|  | do { | 
|  | temp = in_be32(uccf->p_ucce); | 
|  | } while (!(temp & UCCE_GRA)); | 
|  |  | 
|  | uccf->stopped_tx = 1; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ugeth_graceful_stop_rx(struct ucc_geth_private * ugeth) | 
|  | { | 
|  | struct ucc_fast_private *uccf; | 
|  | u32 cecr_subblock; | 
|  | u8 temp; | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  |  | 
|  | /* Clear acknowledge bit */ | 
|  | temp = ugeth->p_rx_glbl_pram->rxgstpack; | 
|  | temp &= ~GRACEFUL_STOP_ACKNOWLEDGE_RX; | 
|  | ugeth->p_rx_glbl_pram->rxgstpack = temp; | 
|  |  | 
|  | /* Keep issuing command and checking acknowledge bit until | 
|  | it is asserted, according to spec */ | 
|  | do { | 
|  | /* Issue host command */ | 
|  | cecr_subblock = | 
|  | ucc_fast_get_qe_cr_subblock(ugeth->ug_info->uf_info. | 
|  | ucc_num); | 
|  | qe_issue_cmd(QE_GRACEFUL_STOP_RX, cecr_subblock, | 
|  | QE_CR_PROTOCOL_ETHERNET, 0); | 
|  |  | 
|  | temp = ugeth->p_rx_glbl_pram->rxgstpack; | 
|  | } while (!(temp & GRACEFUL_STOP_ACKNOWLEDGE_RX)); | 
|  |  | 
|  | uccf->stopped_rx = 1; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ugeth_restart_tx(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | struct ucc_fast_private *uccf; | 
|  | u32 cecr_subblock; | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  |  | 
|  | cecr_subblock = | 
|  | ucc_fast_get_qe_cr_subblock(ugeth->ug_info->uf_info.ucc_num); | 
|  | qe_issue_cmd(QE_RESTART_TX, cecr_subblock, QE_CR_PROTOCOL_ETHERNET, 0); | 
|  | uccf->stopped_tx = 0; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ugeth_restart_rx(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | struct ucc_fast_private *uccf; | 
|  | u32 cecr_subblock; | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  |  | 
|  | cecr_subblock = | 
|  | ucc_fast_get_qe_cr_subblock(ugeth->ug_info->uf_info.ucc_num); | 
|  | qe_issue_cmd(QE_RESTART_RX, cecr_subblock, QE_CR_PROTOCOL_ETHERNET, | 
|  | 0); | 
|  | uccf->stopped_rx = 0; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ugeth_enable(struct ucc_geth_private *ugeth, enum comm_dir mode) | 
|  | { | 
|  | struct ucc_fast_private *uccf; | 
|  | int enabled_tx, enabled_rx; | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  |  | 
|  | /* check if the UCC number is in range. */ | 
|  | if (ugeth->ug_info->uf_info.ucc_num >= UCC_MAX_NUM) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err("%s: ucc_num out of range.", __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | enabled_tx = uccf->enabled_tx; | 
|  | enabled_rx = uccf->enabled_rx; | 
|  |  | 
|  | /* Get Tx and Rx going again, in case this channel was actively | 
|  | disabled. */ | 
|  | if ((mode & COMM_DIR_TX) && (!enabled_tx) && uccf->stopped_tx) | 
|  | ugeth_restart_tx(ugeth); | 
|  | if ((mode & COMM_DIR_RX) && (!enabled_rx) && uccf->stopped_rx) | 
|  | ugeth_restart_rx(ugeth); | 
|  |  | 
|  | ucc_fast_enable(uccf, mode);	/* OK to do even if not disabled */ | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | } | 
|  |  | 
|  | static int ugeth_disable(struct ucc_geth_private * ugeth, enum comm_dir mode) | 
|  | { | 
|  | struct ucc_fast_private *uccf; | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  |  | 
|  | /* check if the UCC number is in range. */ | 
|  | if (ugeth->ug_info->uf_info.ucc_num >= UCC_MAX_NUM) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err("%s: ucc_num out of range.", __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* Stop any transmissions */ | 
|  | if ((mode & COMM_DIR_TX) && uccf->enabled_tx && !uccf->stopped_tx) | 
|  | ugeth_graceful_stop_tx(ugeth); | 
|  |  | 
|  | /* Stop any receptions */ | 
|  | if ((mode & COMM_DIR_RX) && uccf->enabled_rx && !uccf->stopped_rx) | 
|  | ugeth_graceful_stop_rx(ugeth); | 
|  |  | 
|  | ucc_fast_disable(ugeth->uccf, mode); /* OK to do even if not enabled */ | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void ugeth_dump_regs(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | #ifdef DEBUG | 
|  | ucc_fast_dump_regs(ugeth->uccf); | 
|  | dump_regs(ugeth); | 
|  | dump_bds(ugeth); | 
|  | #endif | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_UGETH_FILTERING | 
|  | static int ugeth_ext_filtering_serialize_tad(struct ucc_geth_tad_params * | 
|  | p_UccGethTadParams, | 
|  | struct qe_fltr_tad *qe_fltr_tad) | 
|  | { | 
|  | u16 temp; | 
|  |  | 
|  | /* Zero serialized TAD */ | 
|  | memset(qe_fltr_tad, 0, QE_FLTR_TAD_SIZE); | 
|  |  | 
|  | qe_fltr_tad->serialized[0] |= UCC_GETH_TAD_V;	/* Must have this */ | 
|  | if (p_UccGethTadParams->rx_non_dynamic_extended_features_mode || | 
|  | (p_UccGethTadParams->vtag_op != UCC_GETH_VLAN_OPERATION_TAGGED_NOP) | 
|  | || (p_UccGethTadParams->vnontag_op != | 
|  | UCC_GETH_VLAN_OPERATION_NON_TAGGED_NOP) | 
|  | ) | 
|  | qe_fltr_tad->serialized[0] |= UCC_GETH_TAD_EF; | 
|  | if (p_UccGethTadParams->reject_frame) | 
|  | qe_fltr_tad->serialized[0] |= UCC_GETH_TAD_REJ; | 
|  | temp = | 
|  | (u16) (((u16) p_UccGethTadParams-> | 
|  | vtag_op) << UCC_GETH_TAD_VTAG_OP_SHIFT); | 
|  | qe_fltr_tad->serialized[0] |= (u8) (temp >> 8);	/* upper bits */ | 
|  |  | 
|  | qe_fltr_tad->serialized[1] |= (u8) (temp & 0x00ff);	/* lower bits */ | 
|  | if (p_UccGethTadParams->vnontag_op == | 
|  | UCC_GETH_VLAN_OPERATION_NON_TAGGED_Q_TAG_INSERT) | 
|  | qe_fltr_tad->serialized[1] |= UCC_GETH_TAD_V_NON_VTAG_OP; | 
|  | qe_fltr_tad->serialized[1] |= | 
|  | p_UccGethTadParams->rqos << UCC_GETH_TAD_RQOS_SHIFT; | 
|  |  | 
|  | qe_fltr_tad->serialized[2] |= | 
|  | p_UccGethTadParams->vpri << UCC_GETH_TAD_V_PRIORITY_SHIFT; | 
|  | /* upper bits */ | 
|  | qe_fltr_tad->serialized[2] |= (u8) (p_UccGethTadParams->vid >> 8); | 
|  | /* lower bits */ | 
|  | qe_fltr_tad->serialized[3] |= (u8) (p_UccGethTadParams->vid & 0x00ff); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static struct enet_addr_container_t | 
|  | *ugeth_82xx_filtering_get_match_addr_in_hash(struct ucc_geth_private *ugeth, | 
|  | struct enet_addr *p_enet_addr) | 
|  | { | 
|  | struct enet_addr_container *enet_addr_cont; | 
|  | struct list_head *p_lh; | 
|  | u16 i, num; | 
|  | int32_t j; | 
|  | u8 *p_counter; | 
|  |  | 
|  | if ((*p_enet_addr)[0] & ENET_GROUP_ADDR) { | 
|  | p_lh = &ugeth->group_hash_q; | 
|  | p_counter = &(ugeth->numGroupAddrInHash); | 
|  | } else { | 
|  | p_lh = &ugeth->ind_hash_q; | 
|  | p_counter = &(ugeth->numIndAddrInHash); | 
|  | } | 
|  |  | 
|  | if (!p_lh) | 
|  | return NULL; | 
|  |  | 
|  | num = *p_counter; | 
|  |  | 
|  | for (i = 0; i < num; i++) { | 
|  | enet_addr_cont = | 
|  | (struct enet_addr_container *) | 
|  | ENET_ADDR_CONT_ENTRY(dequeue(p_lh)); | 
|  | for (j = ENET_NUM_OCTETS_PER_ADDRESS - 1; j >= 0; j--) { | 
|  | if ((*p_enet_addr)[j] != (enet_addr_cont->address)[j]) | 
|  | break; | 
|  | if (j == 0) | 
|  | return enet_addr_cont;	/* Found */ | 
|  | } | 
|  | enqueue(p_lh, &enet_addr_cont->node);	/* Put it back */ | 
|  | } | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static int ugeth_82xx_filtering_add_addr_in_hash(struct ucc_geth_private *ugeth, | 
|  | struct enet_addr *p_enet_addr) | 
|  | { | 
|  | enum ucc_geth_enet_address_recognition_location location; | 
|  | struct enet_addr_container *enet_addr_cont; | 
|  | struct list_head *p_lh; | 
|  | u8 i; | 
|  | u32 limit; | 
|  | u8 *p_counter; | 
|  |  | 
|  | if ((*p_enet_addr)[0] & ENET_GROUP_ADDR) { | 
|  | p_lh = &ugeth->group_hash_q; | 
|  | limit = ugeth->ug_info->maxGroupAddrInHash; | 
|  | location = | 
|  | UCC_GETH_ENET_ADDRESS_RECOGNITION_LOCATION_GROUP_HASH; | 
|  | p_counter = &(ugeth->numGroupAddrInHash); | 
|  | } else { | 
|  | p_lh = &ugeth->ind_hash_q; | 
|  | limit = ugeth->ug_info->maxIndAddrInHash; | 
|  | location = | 
|  | UCC_GETH_ENET_ADDRESS_RECOGNITION_LOCATION_INDIVIDUAL_HASH; | 
|  | p_counter = &(ugeth->numIndAddrInHash); | 
|  | } | 
|  |  | 
|  | if ((enet_addr_cont = | 
|  | ugeth_82xx_filtering_get_match_addr_in_hash(ugeth, p_enet_addr))) { | 
|  | list_add(p_lh, &enet_addr_cont->node);	/* Put it back */ | 
|  | return 0; | 
|  | } | 
|  | if ((!p_lh) || (!(*p_counter < limit))) | 
|  | return -EBUSY; | 
|  | if (!(enet_addr_cont = get_enet_addr_container())) | 
|  | return -ENOMEM; | 
|  | for (i = 0; i < ENET_NUM_OCTETS_PER_ADDRESS; i++) | 
|  | (enet_addr_cont->address)[i] = (*p_enet_addr)[i]; | 
|  | enet_addr_cont->location = location; | 
|  | enqueue(p_lh, &enet_addr_cont->node);	/* Put it back */ | 
|  | ++(*p_counter); | 
|  |  | 
|  | hw_add_addr_in_hash(ugeth, enet_addr_cont->address); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ugeth_82xx_filtering_clear_addr_in_hash(struct ucc_geth_private *ugeth, | 
|  | struct enet_addr *p_enet_addr) | 
|  | { | 
|  | struct ucc_geth_82xx_address_filtering_pram *p_82xx_addr_filt; | 
|  | struct enet_addr_container *enet_addr_cont; | 
|  | struct ucc_fast_private *uccf; | 
|  | enum comm_dir comm_dir; | 
|  | u16 i, num; | 
|  | struct list_head *p_lh; | 
|  | u32 *addr_h, *addr_l; | 
|  | u8 *p_counter; | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  |  | 
|  | p_82xx_addr_filt = | 
|  | (struct ucc_geth_82xx_address_filtering_pram *) ugeth->p_rx_glbl_pram-> | 
|  | addressfiltering; | 
|  |  | 
|  | if (! | 
|  | (enet_addr_cont = | 
|  | ugeth_82xx_filtering_get_match_addr_in_hash(ugeth, p_enet_addr))) | 
|  | return -ENOENT; | 
|  |  | 
|  | /* It's been found and removed from the CQ. */ | 
|  | /* Now destroy its container */ | 
|  | put_enet_addr_container(enet_addr_cont); | 
|  |  | 
|  | if ((*p_enet_addr)[0] & ENET_GROUP_ADDR) { | 
|  | addr_h = &(p_82xx_addr_filt->gaddr_h); | 
|  | addr_l = &(p_82xx_addr_filt->gaddr_l); | 
|  | p_lh = &ugeth->group_hash_q; | 
|  | p_counter = &(ugeth->numGroupAddrInHash); | 
|  | } else { | 
|  | addr_h = &(p_82xx_addr_filt->iaddr_h); | 
|  | addr_l = &(p_82xx_addr_filt->iaddr_l); | 
|  | p_lh = &ugeth->ind_hash_q; | 
|  | p_counter = &(ugeth->numIndAddrInHash); | 
|  | } | 
|  |  | 
|  | comm_dir = 0; | 
|  | if (uccf->enabled_tx) | 
|  | comm_dir |= COMM_DIR_TX; | 
|  | if (uccf->enabled_rx) | 
|  | comm_dir |= COMM_DIR_RX; | 
|  | if (comm_dir) | 
|  | ugeth_disable(ugeth, comm_dir); | 
|  |  | 
|  | /* Clear the hash table. */ | 
|  | out_be32(addr_h, 0x00000000); | 
|  | out_be32(addr_l, 0x00000000); | 
|  |  | 
|  | /* Add all remaining CQ elements back into hash */ | 
|  | num = --(*p_counter); | 
|  | for (i = 0; i < num; i++) { | 
|  | enet_addr_cont = | 
|  | (struct enet_addr_container *) | 
|  | ENET_ADDR_CONT_ENTRY(dequeue(p_lh)); | 
|  | hw_add_addr_in_hash(ugeth, enet_addr_cont->address); | 
|  | enqueue(p_lh, &enet_addr_cont->node);	/* Put it back */ | 
|  | } | 
|  |  | 
|  | if (comm_dir) | 
|  | ugeth_enable(ugeth, comm_dir); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  | #endif /* CONFIG_UGETH_FILTERING */ | 
|  |  | 
|  | static int ugeth_82xx_filtering_clear_all_addr_in_hash(struct ucc_geth_private * | 
|  | ugeth, | 
|  | enum enet_addr_type | 
|  | enet_addr_type) | 
|  | { | 
|  | struct ucc_geth_82xx_address_filtering_pram *p_82xx_addr_filt; | 
|  | struct ucc_fast_private *uccf; | 
|  | enum comm_dir comm_dir; | 
|  | struct list_head *p_lh; | 
|  | u16 i, num; | 
|  | u32 *addr_h, *addr_l; | 
|  | u8 *p_counter; | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  |  | 
|  | p_82xx_addr_filt = | 
|  | (struct ucc_geth_82xx_address_filtering_pram *) ugeth->p_rx_glbl_pram-> | 
|  | addressfiltering; | 
|  |  | 
|  | if (enet_addr_type == ENET_ADDR_TYPE_GROUP) { | 
|  | addr_h = &(p_82xx_addr_filt->gaddr_h); | 
|  | addr_l = &(p_82xx_addr_filt->gaddr_l); | 
|  | p_lh = &ugeth->group_hash_q; | 
|  | p_counter = &(ugeth->numGroupAddrInHash); | 
|  | } else if (enet_addr_type == ENET_ADDR_TYPE_INDIVIDUAL) { | 
|  | addr_h = &(p_82xx_addr_filt->iaddr_h); | 
|  | addr_l = &(p_82xx_addr_filt->iaddr_l); | 
|  | p_lh = &ugeth->ind_hash_q; | 
|  | p_counter = &(ugeth->numIndAddrInHash); | 
|  | } else | 
|  | return -EINVAL; | 
|  |  | 
|  | comm_dir = 0; | 
|  | if (uccf->enabled_tx) | 
|  | comm_dir |= COMM_DIR_TX; | 
|  | if (uccf->enabled_rx) | 
|  | comm_dir |= COMM_DIR_RX; | 
|  | if (comm_dir) | 
|  | ugeth_disable(ugeth, comm_dir); | 
|  |  | 
|  | /* Clear the hash table. */ | 
|  | out_be32(addr_h, 0x00000000); | 
|  | out_be32(addr_l, 0x00000000); | 
|  |  | 
|  | if (!p_lh) | 
|  | return 0; | 
|  |  | 
|  | num = *p_counter; | 
|  |  | 
|  | /* Delete all remaining CQ elements */ | 
|  | for (i = 0; i < num; i++) | 
|  | put_enet_addr_container(ENET_ADDR_CONT_ENTRY(dequeue(p_lh))); | 
|  |  | 
|  | *p_counter = 0; | 
|  |  | 
|  | if (comm_dir) | 
|  | ugeth_enable(ugeth, comm_dir); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_UGETH_FILTERING | 
|  | static int ugeth_82xx_filtering_add_addr_in_paddr(struct ucc_geth_private *ugeth, | 
|  | struct enet_addr *p_enet_addr, | 
|  | u8 paddr_num) | 
|  | { | 
|  | int i; | 
|  |  | 
|  | if ((*p_enet_addr)[0] & ENET_GROUP_ADDR) | 
|  | ugeth_warn | 
|  | ("%s: multicast address added to paddr will have no " | 
|  | "effect - is this what you wanted?", | 
|  | __FUNCTION__); | 
|  |  | 
|  | ugeth->indAddrRegUsed[paddr_num] = 1;	/* mark this paddr as used */ | 
|  | /* store address in our database */ | 
|  | for (i = 0; i < ENET_NUM_OCTETS_PER_ADDRESS; i++) | 
|  | ugeth->paddr[paddr_num][i] = (*p_enet_addr)[i]; | 
|  | /* put in hardware */ | 
|  | return hw_add_addr_in_paddr(ugeth, p_enet_addr, paddr_num); | 
|  | } | 
|  | #endif /* CONFIG_UGETH_FILTERING */ | 
|  |  | 
|  | static int ugeth_82xx_filtering_clear_addr_in_paddr(struct ucc_geth_private *ugeth, | 
|  | u8 paddr_num) | 
|  | { | 
|  | ugeth->indAddrRegUsed[paddr_num] = 0; /* mark this paddr as not used */ | 
|  | return hw_clear_addr_in_paddr(ugeth, paddr_num);/* clear in hardware */ | 
|  | } | 
|  |  | 
|  | static void ucc_geth_memclean(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | u16 i, j; | 
|  | u8 *bd; | 
|  |  | 
|  | if (!ugeth) | 
|  | return; | 
|  |  | 
|  | if (ugeth->uccf) { | 
|  | ucc_fast_free(ugeth->uccf); | 
|  | ugeth->uccf = NULL; | 
|  | } | 
|  |  | 
|  | if (ugeth->p_thread_data_tx) { | 
|  | qe_muram_free(ugeth->thread_dat_tx_offset); | 
|  | ugeth->p_thread_data_tx = NULL; | 
|  | } | 
|  | if (ugeth->p_thread_data_rx) { | 
|  | qe_muram_free(ugeth->thread_dat_rx_offset); | 
|  | ugeth->p_thread_data_rx = NULL; | 
|  | } | 
|  | if (ugeth->p_exf_glbl_param) { | 
|  | qe_muram_free(ugeth->exf_glbl_param_offset); | 
|  | ugeth->p_exf_glbl_param = NULL; | 
|  | } | 
|  | if (ugeth->p_rx_glbl_pram) { | 
|  | qe_muram_free(ugeth->rx_glbl_pram_offset); | 
|  | ugeth->p_rx_glbl_pram = NULL; | 
|  | } | 
|  | if (ugeth->p_tx_glbl_pram) { | 
|  | qe_muram_free(ugeth->tx_glbl_pram_offset); | 
|  | ugeth->p_tx_glbl_pram = NULL; | 
|  | } | 
|  | if (ugeth->p_send_q_mem_reg) { | 
|  | qe_muram_free(ugeth->send_q_mem_reg_offset); | 
|  | ugeth->p_send_q_mem_reg = NULL; | 
|  | } | 
|  | if (ugeth->p_scheduler) { | 
|  | qe_muram_free(ugeth->scheduler_offset); | 
|  | ugeth->p_scheduler = NULL; | 
|  | } | 
|  | if (ugeth->p_tx_fw_statistics_pram) { | 
|  | qe_muram_free(ugeth->tx_fw_statistics_pram_offset); | 
|  | ugeth->p_tx_fw_statistics_pram = NULL; | 
|  | } | 
|  | if (ugeth->p_rx_fw_statistics_pram) { | 
|  | qe_muram_free(ugeth->rx_fw_statistics_pram_offset); | 
|  | ugeth->p_rx_fw_statistics_pram = NULL; | 
|  | } | 
|  | if (ugeth->p_rx_irq_coalescing_tbl) { | 
|  | qe_muram_free(ugeth->rx_irq_coalescing_tbl_offset); | 
|  | ugeth->p_rx_irq_coalescing_tbl = NULL; | 
|  | } | 
|  | if (ugeth->p_rx_bd_qs_tbl) { | 
|  | qe_muram_free(ugeth->rx_bd_qs_tbl_offset); | 
|  | ugeth->p_rx_bd_qs_tbl = NULL; | 
|  | } | 
|  | if (ugeth->p_init_enet_param_shadow) { | 
|  | return_init_enet_entries(ugeth, | 
|  | &(ugeth->p_init_enet_param_shadow-> | 
|  | rxthread[0]), | 
|  | ENET_INIT_PARAM_MAX_ENTRIES_RX, | 
|  | ugeth->ug_info->riscRx, 1); | 
|  | return_init_enet_entries(ugeth, | 
|  | &(ugeth->p_init_enet_param_shadow-> | 
|  | txthread[0]), | 
|  | ENET_INIT_PARAM_MAX_ENTRIES_TX, | 
|  | ugeth->ug_info->riscTx, 0); | 
|  | kfree(ugeth->p_init_enet_param_shadow); | 
|  | ugeth->p_init_enet_param_shadow = NULL; | 
|  | } | 
|  | for (i = 0; i < ugeth->ug_info->numQueuesTx; i++) { | 
|  | bd = ugeth->p_tx_bd_ring[i]; | 
|  | if (!bd) | 
|  | continue; | 
|  | for (j = 0; j < ugeth->ug_info->bdRingLenTx[i]; j++) { | 
|  | if (ugeth->tx_skbuff[i][j]) { | 
|  | dma_unmap_single(NULL, | 
|  | ((struct qe_bd *)bd)->buf, | 
|  | (in_be32((u32 *)bd) & | 
|  | BD_LENGTH_MASK), | 
|  | DMA_TO_DEVICE); | 
|  | dev_kfree_skb_any(ugeth->tx_skbuff[i][j]); | 
|  | ugeth->tx_skbuff[i][j] = NULL; | 
|  | } | 
|  | } | 
|  |  | 
|  | kfree(ugeth->tx_skbuff[i]); | 
|  |  | 
|  | if (ugeth->p_tx_bd_ring[i]) { | 
|  | if (ugeth->ug_info->uf_info.bd_mem_part == | 
|  | MEM_PART_SYSTEM) | 
|  | kfree((void *)ugeth->tx_bd_ring_offset[i]); | 
|  | else if (ugeth->ug_info->uf_info.bd_mem_part == | 
|  | MEM_PART_MURAM) | 
|  | qe_muram_free(ugeth->tx_bd_ring_offset[i]); | 
|  | ugeth->p_tx_bd_ring[i] = NULL; | 
|  | } | 
|  | } | 
|  | for (i = 0; i < ugeth->ug_info->numQueuesRx; i++) { | 
|  | if (ugeth->p_rx_bd_ring[i]) { | 
|  | /* Return existing data buffers in ring */ | 
|  | bd = ugeth->p_rx_bd_ring[i]; | 
|  | for (j = 0; j < ugeth->ug_info->bdRingLenRx[i]; j++) { | 
|  | if (ugeth->rx_skbuff[i][j]) { | 
|  | dma_unmap_single(NULL, | 
|  | ((struct qe_bd *)bd)->buf, | 
|  | ugeth->ug_info-> | 
|  | uf_info.max_rx_buf_length + | 
|  | UCC_GETH_RX_DATA_BUF_ALIGNMENT, | 
|  | DMA_FROM_DEVICE); | 
|  | dev_kfree_skb_any( | 
|  | ugeth->rx_skbuff[i][j]); | 
|  | ugeth->rx_skbuff[i][j] = NULL; | 
|  | } | 
|  | bd += sizeof(struct qe_bd); | 
|  | } | 
|  |  | 
|  | kfree(ugeth->rx_skbuff[i]); | 
|  |  | 
|  | if (ugeth->ug_info->uf_info.bd_mem_part == | 
|  | MEM_PART_SYSTEM) | 
|  | kfree((void *)ugeth->rx_bd_ring_offset[i]); | 
|  | else if (ugeth->ug_info->uf_info.bd_mem_part == | 
|  | MEM_PART_MURAM) | 
|  | qe_muram_free(ugeth->rx_bd_ring_offset[i]); | 
|  | ugeth->p_rx_bd_ring[i] = NULL; | 
|  | } | 
|  | } | 
|  | while (!list_empty(&ugeth->group_hash_q)) | 
|  | put_enet_addr_container(ENET_ADDR_CONT_ENTRY | 
|  | (dequeue(&ugeth->group_hash_q))); | 
|  | while (!list_empty(&ugeth->ind_hash_q)) | 
|  | put_enet_addr_container(ENET_ADDR_CONT_ENTRY | 
|  | (dequeue(&ugeth->ind_hash_q))); | 
|  |  | 
|  | } | 
|  |  | 
|  | static void ucc_geth_set_multi(struct net_device *dev) | 
|  | { | 
|  | struct ucc_geth_private *ugeth; | 
|  | struct dev_mc_list *dmi; | 
|  | struct ucc_fast *uf_regs; | 
|  | struct ucc_geth_82xx_address_filtering_pram *p_82xx_addr_filt; | 
|  | int i; | 
|  |  | 
|  | ugeth = netdev_priv(dev); | 
|  |  | 
|  | uf_regs = ugeth->uccf->uf_regs; | 
|  |  | 
|  | if (dev->flags & IFF_PROMISC) { | 
|  |  | 
|  | uf_regs->upsmr |= UPSMR_PRO; | 
|  |  | 
|  | } else { | 
|  |  | 
|  | uf_regs->upsmr &= ~UPSMR_PRO; | 
|  |  | 
|  | p_82xx_addr_filt = | 
|  | (struct ucc_geth_82xx_address_filtering_pram *) ugeth-> | 
|  | p_rx_glbl_pram->addressfiltering; | 
|  |  | 
|  | if (dev->flags & IFF_ALLMULTI) { | 
|  | /* Catch all multicast addresses, so set the | 
|  | * filter to all 1's. | 
|  | */ | 
|  | out_be32(&p_82xx_addr_filt->gaddr_h, 0xffffffff); | 
|  | out_be32(&p_82xx_addr_filt->gaddr_l, 0xffffffff); | 
|  | } else { | 
|  | /* Clear filter and add the addresses in the list. | 
|  | */ | 
|  | out_be32(&p_82xx_addr_filt->gaddr_h, 0x0); | 
|  | out_be32(&p_82xx_addr_filt->gaddr_l, 0x0); | 
|  |  | 
|  | dmi = dev->mc_list; | 
|  |  | 
|  | for (i = 0; i < dev->mc_count; i++, dmi = dmi->next) { | 
|  |  | 
|  | /* Only support group multicast for now. | 
|  | */ | 
|  | if (!(dmi->dmi_addr[0] & 1)) | 
|  | continue; | 
|  |  | 
|  | /* Ask CPM to run CRC and set bit in | 
|  | * filter mask. | 
|  | */ | 
|  | hw_add_addr_in_hash(ugeth, dmi->dmi_addr); | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | static void ucc_geth_stop(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | struct ucc_geth *ug_regs = ugeth->ug_regs; | 
|  | struct phy_device *phydev = ugeth->phydev; | 
|  | u32 tempval; | 
|  |  | 
|  | ugeth_vdbg("%s: IN", __FUNCTION__); | 
|  |  | 
|  | /* Disable the controller */ | 
|  | ugeth_disable(ugeth, COMM_DIR_RX_AND_TX); | 
|  |  | 
|  | /* Tell the kernel the link is down */ | 
|  | phy_stop(phydev); | 
|  |  | 
|  | /* Mask all interrupts */ | 
|  | out_be32(ugeth->uccf->p_uccm, 0x00000000); | 
|  |  | 
|  | /* Clear all interrupts */ | 
|  | out_be32(ugeth->uccf->p_ucce, 0xffffffff); | 
|  |  | 
|  | /* Disable Rx and Tx */ | 
|  | tempval = in_be32(&ug_regs->maccfg1); | 
|  | tempval &= ~(MACCFG1_ENABLE_RX | MACCFG1_ENABLE_TX); | 
|  | out_be32(&ug_regs->maccfg1, tempval); | 
|  |  | 
|  | free_irq(ugeth->ug_info->uf_info.irq, ugeth->dev); | 
|  |  | 
|  | ucc_geth_memclean(ugeth); | 
|  | } | 
|  |  | 
|  | static int ucc_struct_init(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | struct ucc_geth_info *ug_info; | 
|  | struct ucc_fast_info *uf_info; | 
|  | int i; | 
|  |  | 
|  | ug_info = ugeth->ug_info; | 
|  | uf_info = &ug_info->uf_info; | 
|  |  | 
|  | if (!((uf_info->bd_mem_part == MEM_PART_SYSTEM) || | 
|  | (uf_info->bd_mem_part == MEM_PART_MURAM))) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err("%s: Bad memory partition value.", | 
|  | __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* Rx BD lengths */ | 
|  | for (i = 0; i < ug_info->numQueuesRx; i++) { | 
|  | if ((ug_info->bdRingLenRx[i] < UCC_GETH_RX_BD_RING_SIZE_MIN) || | 
|  | (ug_info->bdRingLenRx[i] % | 
|  | UCC_GETH_RX_BD_RING_SIZE_ALIGNMENT)) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Rx BD ring length must be multiple of 4, no smaller than 8.", | 
|  | __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Tx BD lengths */ | 
|  | for (i = 0; i < ug_info->numQueuesTx; i++) { | 
|  | if (ug_info->bdRingLenTx[i] < UCC_GETH_TX_BD_RING_SIZE_MIN) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Tx BD ring length must be no smaller than 2.", | 
|  | __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* mrblr */ | 
|  | if ((uf_info->max_rx_buf_length == 0) || | 
|  | (uf_info->max_rx_buf_length % UCC_GETH_MRBLR_ALIGNMENT)) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: max_rx_buf_length must be non-zero multiple of 128.", | 
|  | __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* num Tx queues */ | 
|  | if (ug_info->numQueuesTx > NUM_TX_QUEUES) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err("%s: number of tx queues too large.", __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* num Rx queues */ | 
|  | if (ug_info->numQueuesRx > NUM_RX_QUEUES) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err("%s: number of rx queues too large.", __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* l2qt */ | 
|  | for (i = 0; i < UCC_GETH_VLAN_PRIORITY_MAX; i++) { | 
|  | if (ug_info->l2qt[i] >= ug_info->numQueuesRx) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: VLAN priority table entry must not be" | 
|  | " larger than number of Rx queues.", | 
|  | __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* l3qt */ | 
|  | for (i = 0; i < UCC_GETH_IP_PRIORITY_MAX; i++) { | 
|  | if (ug_info->l3qt[i] >= ug_info->numQueuesRx) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: IP priority table entry must not be" | 
|  | " larger than number of Rx queues.", | 
|  | __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  | } | 
|  |  | 
|  | if (ug_info->cam && !ug_info->ecamptr) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err("%s: If cam mode is chosen, must supply cam ptr.", | 
|  | __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | if ((ug_info->numStationAddresses != | 
|  | UCC_GETH_NUM_OF_STATION_ADDRESSES_1) | 
|  | && ug_info->rxExtendedFiltering) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err("%s: Number of station addresses greater than 1 " | 
|  | "not allowed in extended parsing mode.", | 
|  | __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* Generate uccm_mask for receive */ | 
|  | uf_info->uccm_mask = ug_info->eventRegMask & UCCE_OTHER;/* Errors */ | 
|  | for (i = 0; i < ug_info->numQueuesRx; i++) | 
|  | uf_info->uccm_mask |= (UCCE_RXBF_SINGLE_MASK << i); | 
|  |  | 
|  | for (i = 0; i < ug_info->numQueuesTx; i++) | 
|  | uf_info->uccm_mask |= (UCCE_TXBF_SINGLE_MASK << i); | 
|  | /* Initialize the general fast UCC block. */ | 
|  | if (ucc_fast_init(uf_info, &ugeth->uccf)) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err("%s: Failed to init uccf.", __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | ugeth->ug_regs = (struct ucc_geth *) ioremap(uf_info->regs, sizeof(struct ucc_geth)); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ucc_geth_startup(struct ucc_geth_private *ugeth) | 
|  | { | 
|  | struct ucc_geth_82xx_address_filtering_pram *p_82xx_addr_filt; | 
|  | struct ucc_geth_init_pram *p_init_enet_pram; | 
|  | struct ucc_fast_private *uccf; | 
|  | struct ucc_geth_info *ug_info; | 
|  | struct ucc_fast_info *uf_info; | 
|  | struct ucc_fast *uf_regs; | 
|  | struct ucc_geth *ug_regs; | 
|  | int ret_val = -EINVAL; | 
|  | u32 remoder = UCC_GETH_REMODER_INIT; | 
|  | u32 init_enet_pram_offset, cecr_subblock, command, maccfg1; | 
|  | u32 ifstat, i, j, size, l2qt, l3qt, length; | 
|  | u16 temoder = UCC_GETH_TEMODER_INIT; | 
|  | u16 test; | 
|  | u8 function_code = 0; | 
|  | u8 *bd, *endOfRing; | 
|  | u8 numThreadsRxNumerical, numThreadsTxNumerical; | 
|  |  | 
|  | ugeth_vdbg("%s: IN", __FUNCTION__); | 
|  | uccf = ugeth->uccf; | 
|  | ug_info = ugeth->ug_info; | 
|  | uf_info = &ug_info->uf_info; | 
|  | uf_regs = uccf->uf_regs; | 
|  | ug_regs = ugeth->ug_regs; | 
|  |  | 
|  | switch (ug_info->numThreadsRx) { | 
|  | case UCC_GETH_NUM_OF_THREADS_1: | 
|  | numThreadsRxNumerical = 1; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_2: | 
|  | numThreadsRxNumerical = 2; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_4: | 
|  | numThreadsRxNumerical = 4; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_6: | 
|  | numThreadsRxNumerical = 6; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_8: | 
|  | numThreadsRxNumerical = 8; | 
|  | break; | 
|  | default: | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Bad number of Rx threads value.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -EINVAL; | 
|  | break; | 
|  | } | 
|  |  | 
|  | switch (ug_info->numThreadsTx) { | 
|  | case UCC_GETH_NUM_OF_THREADS_1: | 
|  | numThreadsTxNumerical = 1; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_2: | 
|  | numThreadsTxNumerical = 2; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_4: | 
|  | numThreadsTxNumerical = 4; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_6: | 
|  | numThreadsTxNumerical = 6; | 
|  | break; | 
|  | case UCC_GETH_NUM_OF_THREADS_8: | 
|  | numThreadsTxNumerical = 8; | 
|  | break; | 
|  | default: | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Bad number of Tx threads value.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -EINVAL; | 
|  | break; | 
|  | } | 
|  |  | 
|  | /* Calculate rx_extended_features */ | 
|  | ugeth->rx_non_dynamic_extended_features = ug_info->ipCheckSumCheck || | 
|  | ug_info->ipAddressAlignment || | 
|  | (ug_info->numStationAddresses != | 
|  | UCC_GETH_NUM_OF_STATION_ADDRESSES_1); | 
|  |  | 
|  | ugeth->rx_extended_features = ugeth->rx_non_dynamic_extended_features || | 
|  | (ug_info->vlanOperationTagged != UCC_GETH_VLAN_OPERATION_TAGGED_NOP) | 
|  | || (ug_info->vlanOperationNonTagged != | 
|  | UCC_GETH_VLAN_OPERATION_NON_TAGGED_NOP); | 
|  |  | 
|  | init_default_reg_vals(&uf_regs->upsmr, | 
|  | &ug_regs->maccfg1, &ug_regs->maccfg2); | 
|  |  | 
|  | /*                    Set UPSMR                      */ | 
|  | /* For more details see the hardware spec.           */ | 
|  | init_rx_parameters(ug_info->bro, | 
|  | ug_info->rsh, ug_info->pro, &uf_regs->upsmr); | 
|  |  | 
|  | /* We're going to ignore other registers for now, */ | 
|  | /* except as needed to get up and running         */ | 
|  |  | 
|  | /*                    Set MACCFG1                    */ | 
|  | /* For more details see the hardware spec.           */ | 
|  | init_flow_control_params(ug_info->aufc, | 
|  | ug_info->receiveFlowControl, | 
|  | ug_info->transmitFlowControl, | 
|  | ug_info->pausePeriod, | 
|  | ug_info->extensionField, | 
|  | &uf_regs->upsmr, | 
|  | &ug_regs->uempr, &ug_regs->maccfg1); | 
|  |  | 
|  | maccfg1 = in_be32(&ug_regs->maccfg1); | 
|  | maccfg1 |= MACCFG1_ENABLE_RX; | 
|  | maccfg1 |= MACCFG1_ENABLE_TX; | 
|  | out_be32(&ug_regs->maccfg1, maccfg1); | 
|  |  | 
|  | /*                    Set IPGIFG                     */ | 
|  | /* For more details see the hardware spec.           */ | 
|  | ret_val = init_inter_frame_gap_params(ug_info->nonBackToBackIfgPart1, | 
|  | ug_info->nonBackToBackIfgPart2, | 
|  | ug_info-> | 
|  | miminumInterFrameGapEnforcement, | 
|  | ug_info->backToBackInterFrameGap, | 
|  | &ug_regs->ipgifg); | 
|  | if (ret_val != 0) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: IPGIFG initialization parameter too large.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  | /*                    Set HAFDUP                     */ | 
|  | /* For more details see the hardware spec.           */ | 
|  | ret_val = init_half_duplex_params(ug_info->altBeb, | 
|  | ug_info->backPressureNoBackoff, | 
|  | ug_info->noBackoff, | 
|  | ug_info->excessDefer, | 
|  | ug_info->altBebTruncation, | 
|  | ug_info->maxRetransmission, | 
|  | ug_info->collisionWindow, | 
|  | &ug_regs->hafdup); | 
|  | if (ret_val != 0) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Half Duplex initialization parameter too large.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  | /*                    Set IFSTAT                     */ | 
|  | /* For more details see the hardware spec.           */ | 
|  | /* Read only - resets upon read                      */ | 
|  | ifstat = in_be32(&ug_regs->ifstat); | 
|  |  | 
|  | /*                    Clear UEMPR                    */ | 
|  | /* For more details see the hardware spec.           */ | 
|  | out_be32(&ug_regs->uempr, 0); | 
|  |  | 
|  | /*                    Set UESCR                      */ | 
|  | /* For more details see the hardware spec.           */ | 
|  | init_hw_statistics_gathering_mode((ug_info->statisticsMode & | 
|  | UCC_GETH_STATISTICS_GATHERING_MODE_HARDWARE), | 
|  | 0, &uf_regs->upsmr, &ug_regs->uescr); | 
|  |  | 
|  | /* Allocate Tx bds */ | 
|  | for (j = 0; j < ug_info->numQueuesTx; j++) { | 
|  | /* Allocate in multiple of | 
|  | UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT, | 
|  | according to spec */ | 
|  | length = ((ug_info->bdRingLenTx[j] * sizeof(struct qe_bd)) | 
|  | / UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT) | 
|  | * UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT; | 
|  | if ((ug_info->bdRingLenTx[j] * sizeof(struct qe_bd)) % | 
|  | UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT) | 
|  | length += UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT; | 
|  | if (uf_info->bd_mem_part == MEM_PART_SYSTEM) { | 
|  | u32 align = 4; | 
|  | if (UCC_GETH_TX_BD_RING_ALIGNMENT > 4) | 
|  | align = UCC_GETH_TX_BD_RING_ALIGNMENT; | 
|  | ugeth->tx_bd_ring_offset[j] = | 
|  | kmalloc((u32) (length + align), GFP_KERNEL); | 
|  |  | 
|  | if (ugeth->tx_bd_ring_offset[j] != 0) | 
|  | ugeth->p_tx_bd_ring[j] = | 
|  | (void*)((ugeth->tx_bd_ring_offset[j] + | 
|  | align) & ~(align - 1)); | 
|  | } else if (uf_info->bd_mem_part == MEM_PART_MURAM) { | 
|  | ugeth->tx_bd_ring_offset[j] = | 
|  | qe_muram_alloc(length, | 
|  | UCC_GETH_TX_BD_RING_ALIGNMENT); | 
|  | if (!IS_ERR_VALUE(ugeth->tx_bd_ring_offset[j])) | 
|  | ugeth->p_tx_bd_ring[j] = | 
|  | (u8 *) qe_muram_addr(ugeth-> | 
|  | tx_bd_ring_offset[j]); | 
|  | } | 
|  | if (!ugeth->p_tx_bd_ring[j]) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate memory for Tx bd rings.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  | /* Zero unused end of bd ring, according to spec */ | 
|  | memset(ugeth->p_tx_bd_ring[j] + | 
|  | ug_info->bdRingLenTx[j] * sizeof(struct qe_bd), 0, | 
|  | length - ug_info->bdRingLenTx[j] * sizeof(struct qe_bd)); | 
|  | } | 
|  |  | 
|  | /* Allocate Rx bds */ | 
|  | for (j = 0; j < ug_info->numQueuesRx; j++) { | 
|  | length = ug_info->bdRingLenRx[j] * sizeof(struct qe_bd); | 
|  | if (uf_info->bd_mem_part == MEM_PART_SYSTEM) { | 
|  | u32 align = 4; | 
|  | if (UCC_GETH_RX_BD_RING_ALIGNMENT > 4) | 
|  | align = UCC_GETH_RX_BD_RING_ALIGNMENT; | 
|  | ugeth->rx_bd_ring_offset[j] = | 
|  | kmalloc((u32) (length + align), GFP_KERNEL); | 
|  | if (ugeth->rx_bd_ring_offset[j] != 0) | 
|  | ugeth->p_rx_bd_ring[j] = | 
|  | (void*)((ugeth->rx_bd_ring_offset[j] + | 
|  | align) & ~(align - 1)); | 
|  | } else if (uf_info->bd_mem_part == MEM_PART_MURAM) { | 
|  | ugeth->rx_bd_ring_offset[j] = | 
|  | qe_muram_alloc(length, | 
|  | UCC_GETH_RX_BD_RING_ALIGNMENT); | 
|  | if (!IS_ERR_VALUE(ugeth->rx_bd_ring_offset[j])) | 
|  | ugeth->p_rx_bd_ring[j] = | 
|  | (u8 *) qe_muram_addr(ugeth-> | 
|  | rx_bd_ring_offset[j]); | 
|  | } | 
|  | if (!ugeth->p_rx_bd_ring[j]) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate memory for Rx bd rings.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Init Tx bds */ | 
|  | for (j = 0; j < ug_info->numQueuesTx; j++) { | 
|  | /* Setup the skbuff rings */ | 
|  | ugeth->tx_skbuff[j] = kmalloc(sizeof(struct sk_buff *) * | 
|  | ugeth->ug_info->bdRingLenTx[j], | 
|  | GFP_KERNEL); | 
|  |  | 
|  | if (ugeth->tx_skbuff[j] == NULL) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Could not allocate tx_skbuff", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | for (i = 0; i < ugeth->ug_info->bdRingLenTx[j]; i++) | 
|  | ugeth->tx_skbuff[j][i] = NULL; | 
|  |  | 
|  | ugeth->skb_curtx[j] = ugeth->skb_dirtytx[j] = 0; | 
|  | bd = ugeth->confBd[j] = ugeth->txBd[j] = ugeth->p_tx_bd_ring[j]; | 
|  | for (i = 0; i < ug_info->bdRingLenTx[j]; i++) { | 
|  | /* clear bd buffer */ | 
|  | out_be32(&((struct qe_bd *)bd)->buf, 0); | 
|  | /* set bd status and length */ | 
|  | out_be32((u32 *)bd, 0); | 
|  | bd += sizeof(struct qe_bd); | 
|  | } | 
|  | bd -= sizeof(struct qe_bd); | 
|  | /* set bd status and length */ | 
|  | out_be32((u32 *)bd, T_W);	/* for last BD set Wrap bit */ | 
|  | } | 
|  |  | 
|  | /* Init Rx bds */ | 
|  | for (j = 0; j < ug_info->numQueuesRx; j++) { | 
|  | /* Setup the skbuff rings */ | 
|  | ugeth->rx_skbuff[j] = kmalloc(sizeof(struct sk_buff *) * | 
|  | ugeth->ug_info->bdRingLenRx[j], | 
|  | GFP_KERNEL); | 
|  |  | 
|  | if (ugeth->rx_skbuff[j] == NULL) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Could not allocate rx_skbuff", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | for (i = 0; i < ugeth->ug_info->bdRingLenRx[j]; i++) | 
|  | ugeth->rx_skbuff[j][i] = NULL; | 
|  |  | 
|  | ugeth->skb_currx[j] = 0; | 
|  | bd = ugeth->rxBd[j] = ugeth->p_rx_bd_ring[j]; | 
|  | for (i = 0; i < ug_info->bdRingLenRx[j]; i++) { | 
|  | /* set bd status and length */ | 
|  | out_be32((u32 *)bd, R_I); | 
|  | /* clear bd buffer */ | 
|  | out_be32(&((struct qe_bd *)bd)->buf, 0); | 
|  | bd += sizeof(struct qe_bd); | 
|  | } | 
|  | bd -= sizeof(struct qe_bd); | 
|  | /* set bd status and length */ | 
|  | out_be32((u32 *)bd, R_W); /* for last BD set Wrap bit */ | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Global PRAM | 
|  | */ | 
|  | /* Tx global PRAM */ | 
|  | /* Allocate global tx parameter RAM page */ | 
|  | ugeth->tx_glbl_pram_offset = | 
|  | qe_muram_alloc(sizeof(struct ucc_geth_tx_global_pram), | 
|  | UCC_GETH_TX_GLOBAL_PRAM_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->tx_glbl_pram_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for p_tx_glbl_pram.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  | ugeth->p_tx_glbl_pram = | 
|  | (struct ucc_geth_tx_global_pram *) qe_muram_addr(ugeth-> | 
|  | tx_glbl_pram_offset); | 
|  | /* Zero out p_tx_glbl_pram */ | 
|  | memset(ugeth->p_tx_glbl_pram, 0, sizeof(struct ucc_geth_tx_global_pram)); | 
|  |  | 
|  | /* Fill global PRAM */ | 
|  |  | 
|  | /* TQPTR */ | 
|  | /* Size varies with number of Tx threads */ | 
|  | ugeth->thread_dat_tx_offset = | 
|  | qe_muram_alloc(numThreadsTxNumerical * | 
|  | sizeof(struct ucc_geth_thread_data_tx) + | 
|  | 32 * (numThreadsTxNumerical == 1), | 
|  | UCC_GETH_THREAD_DATA_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->thread_dat_tx_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for p_thread_data_tx.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | ugeth->p_thread_data_tx = | 
|  | (struct ucc_geth_thread_data_tx *) qe_muram_addr(ugeth-> | 
|  | thread_dat_tx_offset); | 
|  | out_be32(&ugeth->p_tx_glbl_pram->tqptr, ugeth->thread_dat_tx_offset); | 
|  |  | 
|  | /* vtagtable */ | 
|  | for (i = 0; i < UCC_GETH_TX_VTAG_TABLE_ENTRY_MAX; i++) | 
|  | out_be32(&ugeth->p_tx_glbl_pram->vtagtable[i], | 
|  | ug_info->vtagtable[i]); | 
|  |  | 
|  | /* iphoffset */ | 
|  | for (i = 0; i < TX_IP_OFFSET_ENTRY_MAX; i++) | 
|  | ugeth->p_tx_glbl_pram->iphoffset[i] = ug_info->iphoffset[i]; | 
|  |  | 
|  | /* SQPTR */ | 
|  | /* Size varies with number of Tx queues */ | 
|  | ugeth->send_q_mem_reg_offset = | 
|  | qe_muram_alloc(ug_info->numQueuesTx * | 
|  | sizeof(struct ucc_geth_send_queue_qd), | 
|  | UCC_GETH_SEND_QUEUE_QUEUE_DESCRIPTOR_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->send_q_mem_reg_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for p_send_q_mem_reg.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | ugeth->p_send_q_mem_reg = | 
|  | (struct ucc_geth_send_queue_mem_region *) qe_muram_addr(ugeth-> | 
|  | send_q_mem_reg_offset); | 
|  | out_be32(&ugeth->p_tx_glbl_pram->sqptr, ugeth->send_q_mem_reg_offset); | 
|  |  | 
|  | /* Setup the table */ | 
|  | /* Assume BD rings are already established */ | 
|  | for (i = 0; i < ug_info->numQueuesTx; i++) { | 
|  | endOfRing = | 
|  | ugeth->p_tx_bd_ring[i] + (ug_info->bdRingLenTx[i] - | 
|  | 1) * sizeof(struct qe_bd); | 
|  | if (ugeth->ug_info->uf_info.bd_mem_part == MEM_PART_SYSTEM) { | 
|  | out_be32(&ugeth->p_send_q_mem_reg->sqqd[i].bd_ring_base, | 
|  | (u32) virt_to_phys(ugeth->p_tx_bd_ring[i])); | 
|  | out_be32(&ugeth->p_send_q_mem_reg->sqqd[i]. | 
|  | last_bd_completed_address, | 
|  | (u32) virt_to_phys(endOfRing)); | 
|  | } else if (ugeth->ug_info->uf_info.bd_mem_part == | 
|  | MEM_PART_MURAM) { | 
|  | out_be32(&ugeth->p_send_q_mem_reg->sqqd[i].bd_ring_base, | 
|  | (u32) immrbar_virt_to_phys(ugeth-> | 
|  | p_tx_bd_ring[i])); | 
|  | out_be32(&ugeth->p_send_q_mem_reg->sqqd[i]. | 
|  | last_bd_completed_address, | 
|  | (u32) immrbar_virt_to_phys(endOfRing)); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* schedulerbasepointer */ | 
|  |  | 
|  | if (ug_info->numQueuesTx > 1) { | 
|  | /* scheduler exists only if more than 1 tx queue */ | 
|  | ugeth->scheduler_offset = | 
|  | qe_muram_alloc(sizeof(struct ucc_geth_scheduler), | 
|  | UCC_GETH_SCHEDULER_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->scheduler_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for p_scheduler.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | ugeth->p_scheduler = | 
|  | (struct ucc_geth_scheduler *) qe_muram_addr(ugeth-> | 
|  | scheduler_offset); | 
|  | out_be32(&ugeth->p_tx_glbl_pram->schedulerbasepointer, | 
|  | ugeth->scheduler_offset); | 
|  | /* Zero out p_scheduler */ | 
|  | memset(ugeth->p_scheduler, 0, sizeof(struct ucc_geth_scheduler)); | 
|  |  | 
|  | /* Set values in scheduler */ | 
|  | out_be32(&ugeth->p_scheduler->mblinterval, | 
|  | ug_info->mblinterval); | 
|  | out_be16(&ugeth->p_scheduler->nortsrbytetime, | 
|  | ug_info->nortsrbytetime); | 
|  | ugeth->p_scheduler->fracsiz = ug_info->fracsiz; | 
|  | ugeth->p_scheduler->strictpriorityq = ug_info->strictpriorityq; | 
|  | ugeth->p_scheduler->txasap = ug_info->txasap; | 
|  | ugeth->p_scheduler->extrabw = ug_info->extrabw; | 
|  | for (i = 0; i < NUM_TX_QUEUES; i++) | 
|  | ugeth->p_scheduler->weightfactor[i] = | 
|  | ug_info->weightfactor[i]; | 
|  |  | 
|  | /* Set pointers to cpucount registers in scheduler */ | 
|  | ugeth->p_cpucount[0] = &(ugeth->p_scheduler->cpucount0); | 
|  | ugeth->p_cpucount[1] = &(ugeth->p_scheduler->cpucount1); | 
|  | ugeth->p_cpucount[2] = &(ugeth->p_scheduler->cpucount2); | 
|  | ugeth->p_cpucount[3] = &(ugeth->p_scheduler->cpucount3); | 
|  | ugeth->p_cpucount[4] = &(ugeth->p_scheduler->cpucount4); | 
|  | ugeth->p_cpucount[5] = &(ugeth->p_scheduler->cpucount5); | 
|  | ugeth->p_cpucount[6] = &(ugeth->p_scheduler->cpucount6); | 
|  | ugeth->p_cpucount[7] = &(ugeth->p_scheduler->cpucount7); | 
|  | } | 
|  |  | 
|  | /* schedulerbasepointer */ | 
|  | /* TxRMON_PTR (statistics) */ | 
|  | if (ug_info-> | 
|  | statisticsMode & UCC_GETH_STATISTICS_GATHERING_MODE_FIRMWARE_TX) { | 
|  | ugeth->tx_fw_statistics_pram_offset = | 
|  | qe_muram_alloc(sizeof | 
|  | (struct ucc_geth_tx_firmware_statistics_pram), | 
|  | UCC_GETH_TX_STATISTICS_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->tx_fw_statistics_pram_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for" | 
|  | " p_tx_fw_statistics_pram.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  | ugeth->p_tx_fw_statistics_pram = | 
|  | (struct ucc_geth_tx_firmware_statistics_pram *) | 
|  | qe_muram_addr(ugeth->tx_fw_statistics_pram_offset); | 
|  | /* Zero out p_tx_fw_statistics_pram */ | 
|  | memset(ugeth->p_tx_fw_statistics_pram, | 
|  | 0, sizeof(struct ucc_geth_tx_firmware_statistics_pram)); | 
|  | } | 
|  |  | 
|  | /* temoder */ | 
|  | /* Already has speed set */ | 
|  |  | 
|  | if (ug_info->numQueuesTx > 1) | 
|  | temoder |= TEMODER_SCHEDULER_ENABLE; | 
|  | if (ug_info->ipCheckSumGenerate) | 
|  | temoder |= TEMODER_IP_CHECKSUM_GENERATE; | 
|  | temoder |= ((ug_info->numQueuesTx - 1) << TEMODER_NUM_OF_QUEUES_SHIFT); | 
|  | out_be16(&ugeth->p_tx_glbl_pram->temoder, temoder); | 
|  |  | 
|  | test = in_be16(&ugeth->p_tx_glbl_pram->temoder); | 
|  |  | 
|  | /* Function code register value to be used later */ | 
|  | function_code = UCC_BMR_BO_BE | UCC_BMR_GBL; | 
|  | /* Required for QE */ | 
|  |  | 
|  | /* function code register */ | 
|  | out_be32(&ugeth->p_tx_glbl_pram->tstate, ((u32) function_code) << 24); | 
|  |  | 
|  | /* Rx global PRAM */ | 
|  | /* Allocate global rx parameter RAM page */ | 
|  | ugeth->rx_glbl_pram_offset = | 
|  | qe_muram_alloc(sizeof(struct ucc_geth_rx_global_pram), | 
|  | UCC_GETH_RX_GLOBAL_PRAM_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->rx_glbl_pram_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for p_rx_glbl_pram.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  | ugeth->p_rx_glbl_pram = | 
|  | (struct ucc_geth_rx_global_pram *) qe_muram_addr(ugeth-> | 
|  | rx_glbl_pram_offset); | 
|  | /* Zero out p_rx_glbl_pram */ | 
|  | memset(ugeth->p_rx_glbl_pram, 0, sizeof(struct ucc_geth_rx_global_pram)); | 
|  |  | 
|  | /* Fill global PRAM */ | 
|  |  | 
|  | /* RQPTR */ | 
|  | /* Size varies with number of Rx threads */ | 
|  | ugeth->thread_dat_rx_offset = | 
|  | qe_muram_alloc(numThreadsRxNumerical * | 
|  | sizeof(struct ucc_geth_thread_data_rx), | 
|  | UCC_GETH_THREAD_DATA_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->thread_dat_rx_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for p_thread_data_rx.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | ugeth->p_thread_data_rx = | 
|  | (struct ucc_geth_thread_data_rx *) qe_muram_addr(ugeth-> | 
|  | thread_dat_rx_offset); | 
|  | out_be32(&ugeth->p_rx_glbl_pram->rqptr, ugeth->thread_dat_rx_offset); | 
|  |  | 
|  | /* typeorlen */ | 
|  | out_be16(&ugeth->p_rx_glbl_pram->typeorlen, ug_info->typeorlen); | 
|  |  | 
|  | /* rxrmonbaseptr (statistics) */ | 
|  | if (ug_info-> | 
|  | statisticsMode & UCC_GETH_STATISTICS_GATHERING_MODE_FIRMWARE_RX) { | 
|  | ugeth->rx_fw_statistics_pram_offset = | 
|  | qe_muram_alloc(sizeof | 
|  | (struct ucc_geth_rx_firmware_statistics_pram), | 
|  | UCC_GETH_RX_STATISTICS_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->rx_fw_statistics_pram_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for" | 
|  | " p_rx_fw_statistics_pram.", __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  | ugeth->p_rx_fw_statistics_pram = | 
|  | (struct ucc_geth_rx_firmware_statistics_pram *) | 
|  | qe_muram_addr(ugeth->rx_fw_statistics_pram_offset); | 
|  | /* Zero out p_rx_fw_statistics_pram */ | 
|  | memset(ugeth->p_rx_fw_statistics_pram, 0, | 
|  | sizeof(struct ucc_geth_rx_firmware_statistics_pram)); | 
|  | } | 
|  |  | 
|  | /* intCoalescingPtr */ | 
|  |  | 
|  | /* Size varies with number of Rx queues */ | 
|  | ugeth->rx_irq_coalescing_tbl_offset = | 
|  | qe_muram_alloc(ug_info->numQueuesRx * | 
|  | sizeof(struct ucc_geth_rx_interrupt_coalescing_entry) | 
|  | + 4, UCC_GETH_RX_INTERRUPT_COALESCING_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->rx_irq_coalescing_tbl_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for" | 
|  | " p_rx_irq_coalescing_tbl.", __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | ugeth->p_rx_irq_coalescing_tbl = | 
|  | (struct ucc_geth_rx_interrupt_coalescing_table *) | 
|  | qe_muram_addr(ugeth->rx_irq_coalescing_tbl_offset); | 
|  | out_be32(&ugeth->p_rx_glbl_pram->intcoalescingptr, | 
|  | ugeth->rx_irq_coalescing_tbl_offset); | 
|  |  | 
|  | /* Fill interrupt coalescing table */ | 
|  | for (i = 0; i < ug_info->numQueuesRx; i++) { | 
|  | out_be32(&ugeth->p_rx_irq_coalescing_tbl->coalescingentry[i]. | 
|  | interruptcoalescingmaxvalue, | 
|  | ug_info->interruptcoalescingmaxvalue[i]); | 
|  | out_be32(&ugeth->p_rx_irq_coalescing_tbl->coalescingentry[i]. | 
|  | interruptcoalescingcounter, | 
|  | ug_info->interruptcoalescingmaxvalue[i]); | 
|  | } | 
|  |  | 
|  | /* MRBLR */ | 
|  | init_max_rx_buff_len(uf_info->max_rx_buf_length, | 
|  | &ugeth->p_rx_glbl_pram->mrblr); | 
|  | /* MFLR */ | 
|  | out_be16(&ugeth->p_rx_glbl_pram->mflr, ug_info->maxFrameLength); | 
|  | /* MINFLR */ | 
|  | init_min_frame_len(ug_info->minFrameLength, | 
|  | &ugeth->p_rx_glbl_pram->minflr, | 
|  | &ugeth->p_rx_glbl_pram->mrblr); | 
|  | /* MAXD1 */ | 
|  | out_be16(&ugeth->p_rx_glbl_pram->maxd1, ug_info->maxD1Length); | 
|  | /* MAXD2 */ | 
|  | out_be16(&ugeth->p_rx_glbl_pram->maxd2, ug_info->maxD2Length); | 
|  |  | 
|  | /* l2qt */ | 
|  | l2qt = 0; | 
|  | for (i = 0; i < UCC_GETH_VLAN_PRIORITY_MAX; i++) | 
|  | l2qt |= (ug_info->l2qt[i] << (28 - 4 * i)); | 
|  | out_be32(&ugeth->p_rx_glbl_pram->l2qt, l2qt); | 
|  |  | 
|  | /* l3qt */ | 
|  | for (j = 0; j < UCC_GETH_IP_PRIORITY_MAX; j += 8) { | 
|  | l3qt = 0; | 
|  | for (i = 0; i < 8; i++) | 
|  | l3qt |= (ug_info->l3qt[j + i] << (28 - 4 * i)); | 
|  | out_be32(&ugeth->p_rx_glbl_pram->l3qt[j/8], l3qt); | 
|  | } | 
|  |  | 
|  | /* vlantype */ | 
|  | out_be16(&ugeth->p_rx_glbl_pram->vlantype, ug_info->vlantype); | 
|  |  | 
|  | /* vlantci */ | 
|  | out_be16(&ugeth->p_rx_glbl_pram->vlantci, ug_info->vlantci); | 
|  |  | 
|  | /* ecamptr */ | 
|  | out_be32(&ugeth->p_rx_glbl_pram->ecamptr, ug_info->ecamptr); | 
|  |  | 
|  | /* RBDQPTR */ | 
|  | /* Size varies with number of Rx queues */ | 
|  | ugeth->rx_bd_qs_tbl_offset = | 
|  | qe_muram_alloc(ug_info->numQueuesRx * | 
|  | (sizeof(struct ucc_geth_rx_bd_queues_entry) + | 
|  | sizeof(struct ucc_geth_rx_prefetched_bds)), | 
|  | UCC_GETH_RX_BD_QUEUES_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->rx_bd_qs_tbl_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for p_rx_bd_qs_tbl.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | ugeth->p_rx_bd_qs_tbl = | 
|  | (struct ucc_geth_rx_bd_queues_entry *) qe_muram_addr(ugeth-> | 
|  | rx_bd_qs_tbl_offset); | 
|  | out_be32(&ugeth->p_rx_glbl_pram->rbdqptr, ugeth->rx_bd_qs_tbl_offset); | 
|  | /* Zero out p_rx_bd_qs_tbl */ | 
|  | memset(ugeth->p_rx_bd_qs_tbl, | 
|  | 0, | 
|  | ug_info->numQueuesRx * (sizeof(struct ucc_geth_rx_bd_queues_entry) + | 
|  | sizeof(struct ucc_geth_rx_prefetched_bds))); | 
|  |  | 
|  | /* Setup the table */ | 
|  | /* Assume BD rings are already established */ | 
|  | for (i = 0; i < ug_info->numQueuesRx; i++) { | 
|  | if (ugeth->ug_info->uf_info.bd_mem_part == MEM_PART_SYSTEM) { | 
|  | out_be32(&ugeth->p_rx_bd_qs_tbl[i].externalbdbaseptr, | 
|  | (u32) virt_to_phys(ugeth->p_rx_bd_ring[i])); | 
|  | } else if (ugeth->ug_info->uf_info.bd_mem_part == | 
|  | MEM_PART_MURAM) { | 
|  | out_be32(&ugeth->p_rx_bd_qs_tbl[i].externalbdbaseptr, | 
|  | (u32) immrbar_virt_to_phys(ugeth-> | 
|  | p_rx_bd_ring[i])); | 
|  | } | 
|  | /* rest of fields handled by QE */ | 
|  | } | 
|  |  | 
|  | /* remoder */ | 
|  | /* Already has speed set */ | 
|  |  | 
|  | if (ugeth->rx_extended_features) | 
|  | remoder |= REMODER_RX_EXTENDED_FEATURES; | 
|  | if (ug_info->rxExtendedFiltering) | 
|  | remoder |= REMODER_RX_EXTENDED_FILTERING; | 
|  | if (ug_info->dynamicMaxFrameLength) | 
|  | remoder |= REMODER_DYNAMIC_MAX_FRAME_LENGTH; | 
|  | if (ug_info->dynamicMinFrameLength) | 
|  | remoder |= REMODER_DYNAMIC_MIN_FRAME_LENGTH; | 
|  | remoder |= | 
|  | ug_info->vlanOperationTagged << REMODER_VLAN_OPERATION_TAGGED_SHIFT; | 
|  | remoder |= | 
|  | ug_info-> | 
|  | vlanOperationNonTagged << REMODER_VLAN_OPERATION_NON_TAGGED_SHIFT; | 
|  | remoder |= ug_info->rxQoSMode << REMODER_RX_QOS_MODE_SHIFT; | 
|  | remoder |= ((ug_info->numQueuesRx - 1) << REMODER_NUM_OF_QUEUES_SHIFT); | 
|  | if (ug_info->ipCheckSumCheck) | 
|  | remoder |= REMODER_IP_CHECKSUM_CHECK; | 
|  | if (ug_info->ipAddressAlignment) | 
|  | remoder |= REMODER_IP_ADDRESS_ALIGNMENT; | 
|  | out_be32(&ugeth->p_rx_glbl_pram->remoder, remoder); | 
|  |  | 
|  | /* Note that this function must be called */ | 
|  | /* ONLY AFTER p_tx_fw_statistics_pram */ | 
|  | /* andp_UccGethRxFirmwareStatisticsPram are allocated ! */ | 
|  | init_firmware_statistics_gathering_mode((ug_info-> | 
|  | statisticsMode & | 
|  | UCC_GETH_STATISTICS_GATHERING_MODE_FIRMWARE_TX), | 
|  | (ug_info->statisticsMode & | 
|  | UCC_GETH_STATISTICS_GATHERING_MODE_FIRMWARE_RX), | 
|  | &ugeth->p_tx_glbl_pram->txrmonbaseptr, | 
|  | ugeth->tx_fw_statistics_pram_offset, | 
|  | &ugeth->p_rx_glbl_pram->rxrmonbaseptr, | 
|  | ugeth->rx_fw_statistics_pram_offset, | 
|  | &ugeth->p_tx_glbl_pram->temoder, | 
|  | &ugeth->p_rx_glbl_pram->remoder); | 
|  |  | 
|  | /* function code register */ | 
|  | ugeth->p_rx_glbl_pram->rstate = function_code; | 
|  |  | 
|  | /* initialize extended filtering */ | 
|  | if (ug_info->rxExtendedFiltering) { | 
|  | if (!ug_info->extendedFilteringChainPointer) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Null Extended Filtering Chain Pointer.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* Allocate memory for extended filtering Mode Global | 
|  | Parameters */ | 
|  | ugeth->exf_glbl_param_offset = | 
|  | qe_muram_alloc(sizeof(struct ucc_geth_exf_global_pram), | 
|  | UCC_GETH_RX_EXTENDED_FILTERING_GLOBAL_PARAMETERS_ALIGNMENT); | 
|  | if (IS_ERR_VALUE(ugeth->exf_glbl_param_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for" | 
|  | " p_exf_glbl_param.", __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | ugeth->p_exf_glbl_param = | 
|  | (struct ucc_geth_exf_global_pram *) qe_muram_addr(ugeth-> | 
|  | exf_glbl_param_offset); | 
|  | out_be32(&ugeth->p_rx_glbl_pram->exfGlobalParam, | 
|  | ugeth->exf_glbl_param_offset); | 
|  | out_be32(&ugeth->p_exf_glbl_param->l2pcdptr, | 
|  | (u32) ug_info->extendedFilteringChainPointer); | 
|  |  | 
|  | } else {		/* initialize 82xx style address filtering */ | 
|  |  | 
|  | /* Init individual address recognition registers to disabled */ | 
|  |  | 
|  | for (j = 0; j < NUM_OF_PADDRS; j++) | 
|  | ugeth_82xx_filtering_clear_addr_in_paddr(ugeth, (u8) j); | 
|  |  | 
|  | p_82xx_addr_filt = | 
|  | (struct ucc_geth_82xx_address_filtering_pram *) ugeth-> | 
|  | p_rx_glbl_pram->addressfiltering; | 
|  |  | 
|  | ugeth_82xx_filtering_clear_all_addr_in_hash(ugeth, | 
|  | ENET_ADDR_TYPE_GROUP); | 
|  | ugeth_82xx_filtering_clear_all_addr_in_hash(ugeth, | 
|  | ENET_ADDR_TYPE_INDIVIDUAL); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Initialize UCC at QE level | 
|  | */ | 
|  |  | 
|  | command = QE_INIT_TX_RX; | 
|  |  | 
|  | /* Allocate shadow InitEnet command parameter structure. | 
|  | * This is needed because after the InitEnet command is executed, | 
|  | * the structure in DPRAM is released, because DPRAM is a premium | 
|  | * resource. | 
|  | * This shadow structure keeps a copy of what was done so that the | 
|  | * allocated resources can be released when the channel is freed. | 
|  | */ | 
|  | if (!(ugeth->p_init_enet_param_shadow = | 
|  | kmalloc(sizeof(struct ucc_geth_init_pram), GFP_KERNEL))) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate memory for" | 
|  | " p_UccInitEnetParamShadows.", __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  | /* Zero out *p_init_enet_param_shadow */ | 
|  | memset((char *)ugeth->p_init_enet_param_shadow, | 
|  | 0, sizeof(struct ucc_geth_init_pram)); | 
|  |  | 
|  | /* Fill shadow InitEnet command parameter structure */ | 
|  |  | 
|  | ugeth->p_init_enet_param_shadow->resinit1 = | 
|  | ENET_INIT_PARAM_MAGIC_RES_INIT1; | 
|  | ugeth->p_init_enet_param_shadow->resinit2 = | 
|  | ENET_INIT_PARAM_MAGIC_RES_INIT2; | 
|  | ugeth->p_init_enet_param_shadow->resinit3 = | 
|  | ENET_INIT_PARAM_MAGIC_RES_INIT3; | 
|  | ugeth->p_init_enet_param_shadow->resinit4 = | 
|  | ENET_INIT_PARAM_MAGIC_RES_INIT4; | 
|  | ugeth->p_init_enet_param_shadow->resinit5 = | 
|  | ENET_INIT_PARAM_MAGIC_RES_INIT5; | 
|  | ugeth->p_init_enet_param_shadow->rgftgfrxglobal |= | 
|  | ((u32) ug_info->numThreadsRx) << ENET_INIT_PARAM_RGF_SHIFT; | 
|  | ugeth->p_init_enet_param_shadow->rgftgfrxglobal |= | 
|  | ((u32) ug_info->numThreadsTx) << ENET_INIT_PARAM_TGF_SHIFT; | 
|  |  | 
|  | ugeth->p_init_enet_param_shadow->rgftgfrxglobal |= | 
|  | ugeth->rx_glbl_pram_offset | ug_info->riscRx; | 
|  | if ((ug_info->largestexternallookupkeysize != | 
|  | QE_FLTR_LARGEST_EXTERNAL_TABLE_LOOKUP_KEY_SIZE_NONE) | 
|  | && (ug_info->largestexternallookupkeysize != | 
|  | QE_FLTR_LARGEST_EXTERNAL_TABLE_LOOKUP_KEY_SIZE_8_BYTES) | 
|  | && (ug_info->largestexternallookupkeysize != | 
|  | QE_FLTR_LARGEST_EXTERNAL_TABLE_LOOKUP_KEY_SIZE_16_BYTES)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Invalid largest External Lookup Key Size.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -EINVAL; | 
|  | } | 
|  | ugeth->p_init_enet_param_shadow->largestexternallookupkeysize = | 
|  | ug_info->largestexternallookupkeysize; | 
|  | size = sizeof(struct ucc_geth_thread_rx_pram); | 
|  | if (ug_info->rxExtendedFiltering) { | 
|  | size += THREAD_RX_PRAM_ADDITIONAL_FOR_EXTENDED_FILTERING; | 
|  | if (ug_info->largestexternallookupkeysize == | 
|  | QE_FLTR_TABLE_LOOKUP_KEY_SIZE_8_BYTES) | 
|  | size += | 
|  | THREAD_RX_PRAM_ADDITIONAL_FOR_EXTENDED_FILTERING_8; | 
|  | if (ug_info->largestexternallookupkeysize == | 
|  | QE_FLTR_TABLE_LOOKUP_KEY_SIZE_16_BYTES) | 
|  | size += | 
|  | THREAD_RX_PRAM_ADDITIONAL_FOR_EXTENDED_FILTERING_16; | 
|  | } | 
|  |  | 
|  | if ((ret_val = fill_init_enet_entries(ugeth, &(ugeth-> | 
|  | p_init_enet_param_shadow->rxthread[0]), | 
|  | (u8) (numThreadsRxNumerical + 1) | 
|  | /* Rx needs one extra for terminator */ | 
|  | , size, UCC_GETH_THREAD_RX_PRAM_ALIGNMENT, | 
|  | ug_info->riscRx, 1)) != 0) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Can not fill p_init_enet_param_shadow.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  | ugeth->p_init_enet_param_shadow->txglobal = | 
|  | ugeth->tx_glbl_pram_offset | ug_info->riscTx; | 
|  | if ((ret_val = | 
|  | fill_init_enet_entries(ugeth, | 
|  | &(ugeth->p_init_enet_param_shadow-> | 
|  | txthread[0]), numThreadsTxNumerical, | 
|  | sizeof(struct ucc_geth_thread_tx_pram), | 
|  | UCC_GETH_THREAD_TX_PRAM_ALIGNMENT, | 
|  | ug_info->riscTx, 0)) != 0) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Can not fill p_init_enet_param_shadow.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return ret_val; | 
|  | } | 
|  |  | 
|  | /* Load Rx bds with buffers */ | 
|  | for (i = 0; i < ug_info->numQueuesRx; i++) { | 
|  | if ((ret_val = rx_bd_buffer_set(ugeth, (u8) i)) != 0) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Can not fill Rx bds with buffers.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return ret_val; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Allocate InitEnet command parameter structure */ | 
|  | init_enet_pram_offset = qe_muram_alloc(sizeof(struct ucc_geth_init_pram), 4); | 
|  | if (IS_ERR_VALUE(init_enet_pram_offset)) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err | 
|  | ("%s: Can not allocate DPRAM memory for p_init_enet_pram.", | 
|  | __FUNCTION__); | 
|  | ucc_geth_memclean(ugeth); | 
|  | return -ENOMEM; | 
|  | } | 
|  | p_init_enet_pram = | 
|  | (struct ucc_geth_init_pram *) qe_muram_addr(init_enet_pram_offset); | 
|  |  | 
|  | /* Copy shadow InitEnet command parameter structure into PRAM */ | 
|  | p_init_enet_pram->resinit1 = ugeth->p_init_enet_param_shadow->resinit1; | 
|  | p_init_enet_pram->resinit2 = ugeth->p_init_enet_param_shadow->resinit2; | 
|  | p_init_enet_pram->resinit3 = ugeth->p_init_enet_param_shadow->resinit3; | 
|  | p_init_enet_pram->resinit4 = ugeth->p_init_enet_param_shadow->resinit4; | 
|  | out_be16(&p_init_enet_pram->resinit5, | 
|  | ugeth->p_init_enet_param_shadow->resinit5); | 
|  | p_init_enet_pram->largestexternallookupkeysize = | 
|  | ugeth->p_init_enet_param_shadow->largestexternallookupkeysize; | 
|  | out_be32(&p_init_enet_pram->rgftgfrxglobal, | 
|  | ugeth->p_init_enet_param_shadow->rgftgfrxglobal); | 
|  | for (i = 0; i < ENET_INIT_PARAM_MAX_ENTRIES_RX; i++) | 
|  | out_be32(&p_init_enet_pram->rxthread[i], | 
|  | ugeth->p_init_enet_param_shadow->rxthread[i]); | 
|  | out_be32(&p_init_enet_pram->txglobal, | 
|  | ugeth->p_init_enet_param_shadow->txglobal); | 
|  | for (i = 0; i < ENET_INIT_PARAM_MAX_ENTRIES_TX; i++) | 
|  | out_be32(&p_init_enet_pram->txthread[i], | 
|  | ugeth->p_init_enet_param_shadow->txthread[i]); | 
|  |  | 
|  | /* Issue QE command */ | 
|  | cecr_subblock = | 
|  | ucc_fast_get_qe_cr_subblock(ugeth->ug_info->uf_info.ucc_num); | 
|  | qe_issue_cmd(command, cecr_subblock, QE_CR_PROTOCOL_ETHERNET, | 
|  | init_enet_pram_offset); | 
|  |  | 
|  | /* Free InitEnet command parameter */ | 
|  | qe_muram_free(init_enet_pram_offset); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* ucc_geth_timeout gets called when a packet has not been | 
|  | * transmitted after a set amount of time. | 
|  | * For now, assume that clearing out all the structures, and | 
|  | * starting over will fix the problem. */ | 
|  | static void ucc_geth_timeout(struct net_device *dev) | 
|  | { | 
|  | struct ucc_geth_private *ugeth = netdev_priv(dev); | 
|  |  | 
|  | ugeth_vdbg("%s: IN", __FUNCTION__); | 
|  |  | 
|  | dev->stats.tx_errors++; | 
|  |  | 
|  | ugeth_dump_regs(ugeth); | 
|  |  | 
|  | if (dev->flags & IFF_UP) { | 
|  | ucc_geth_stop(ugeth); | 
|  | ucc_geth_startup(ugeth); | 
|  | } | 
|  |  | 
|  | netif_schedule(dev); | 
|  | } | 
|  |  | 
|  | /* This is called by the kernel when a frame is ready for transmission. */ | 
|  | /* It is pointed to by the dev->hard_start_xmit function pointer */ | 
|  | static int ucc_geth_start_xmit(struct sk_buff *skb, struct net_device *dev) | 
|  | { | 
|  | struct ucc_geth_private *ugeth = netdev_priv(dev); | 
|  | #ifdef CONFIG_UGETH_TX_ON_DEMAND | 
|  | struct ucc_fast_private *uccf; | 
|  | #endif | 
|  | u8 *bd;			/* BD pointer */ | 
|  | u32 bd_status; | 
|  | u8 txQ = 0; | 
|  |  | 
|  | ugeth_vdbg("%s: IN", __FUNCTION__); | 
|  |  | 
|  | spin_lock_irq(&ugeth->lock); | 
|  |  | 
|  | dev->stats.tx_bytes += skb->len; | 
|  |  | 
|  | /* Start from the next BD that should be filled */ | 
|  | bd = ugeth->txBd[txQ]; | 
|  | bd_status = in_be32((u32 *)bd); | 
|  | /* Save the skb pointer so we can free it later */ | 
|  | ugeth->tx_skbuff[txQ][ugeth->skb_curtx[txQ]] = skb; | 
|  |  | 
|  | /* Update the current skb pointer (wrapping if this was the last) */ | 
|  | ugeth->skb_curtx[txQ] = | 
|  | (ugeth->skb_curtx[txQ] + | 
|  | 1) & TX_RING_MOD_MASK(ugeth->ug_info->bdRingLenTx[txQ]); | 
|  |  | 
|  | /* set up the buffer descriptor */ | 
|  | out_be32(&((struct qe_bd *)bd)->buf, | 
|  | dma_map_single(NULL, skb->data, skb->len, DMA_TO_DEVICE)); | 
|  |  | 
|  | /* printk(KERN_DEBUG"skb->data is 0x%x\n",skb->data); */ | 
|  |  | 
|  | bd_status = (bd_status & T_W) | T_R | T_I | T_L | skb->len; | 
|  |  | 
|  | /* set bd status and length */ | 
|  | out_be32((u32 *)bd, bd_status); | 
|  |  | 
|  | dev->trans_start = jiffies; | 
|  |  | 
|  | /* Move to next BD in the ring */ | 
|  | if (!(bd_status & T_W)) | 
|  | bd += sizeof(struct qe_bd); | 
|  | else | 
|  | bd = ugeth->p_tx_bd_ring[txQ]; | 
|  |  | 
|  | /* If the next BD still needs to be cleaned up, then the bds | 
|  | are full.  We need to tell the kernel to stop sending us stuff. */ | 
|  | if (bd == ugeth->confBd[txQ]) { | 
|  | if (!netif_queue_stopped(dev)) | 
|  | netif_stop_queue(dev); | 
|  | } | 
|  |  | 
|  | ugeth->txBd[txQ] = bd; | 
|  |  | 
|  | if (ugeth->p_scheduler) { | 
|  | ugeth->cpucount[txQ]++; | 
|  | /* Indicate to QE that there are more Tx bds ready for | 
|  | transmission */ | 
|  | /* This is done by writing a running counter of the bd | 
|  | count to the scheduler PRAM. */ | 
|  | out_be16(ugeth->p_cpucount[txQ], ugeth->cpucount[txQ]); | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_UGETH_TX_ON_DEMAND | 
|  | uccf = ugeth->uccf; | 
|  | out_be16(uccf->p_utodr, UCC_FAST_TOD); | 
|  | #endif | 
|  | spin_unlock_irq(&ugeth->lock); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ucc_geth_rx(struct ucc_geth_private *ugeth, u8 rxQ, int rx_work_limit) | 
|  | { | 
|  | struct sk_buff *skb; | 
|  | u8 *bd; | 
|  | u16 length, howmany = 0; | 
|  | u32 bd_status; | 
|  | u8 *bdBuffer; | 
|  | struct net_device *dev; | 
|  |  | 
|  | ugeth_vdbg("%s: IN", __FUNCTION__); | 
|  |  | 
|  | dev = ugeth->dev; | 
|  |  | 
|  | /* collect received buffers */ | 
|  | bd = ugeth->rxBd[rxQ]; | 
|  |  | 
|  | bd_status = in_be32((u32 *)bd); | 
|  |  | 
|  | /* while there are received buffers and BD is full (~R_E) */ | 
|  | while (!((bd_status & (R_E)) || (--rx_work_limit < 0))) { | 
|  | bdBuffer = (u8 *) in_be32(&((struct qe_bd *)bd)->buf); | 
|  | length = (u16) ((bd_status & BD_LENGTH_MASK) - 4); | 
|  | skb = ugeth->rx_skbuff[rxQ][ugeth->skb_currx[rxQ]]; | 
|  |  | 
|  | /* determine whether buffer is first, last, first and last | 
|  | (single buffer frame) or middle (not first and not last) */ | 
|  | if (!skb || | 
|  | (!(bd_status & (R_F | R_L))) || | 
|  | (bd_status & R_ERRORS_FATAL)) { | 
|  | if (netif_msg_rx_err(ugeth)) | 
|  | ugeth_err("%s, %d: ERROR!!! skb - 0x%08x", | 
|  | __FUNCTION__, __LINE__, (u32) skb); | 
|  | if (skb) | 
|  | dev_kfree_skb_any(skb); | 
|  |  | 
|  | ugeth->rx_skbuff[rxQ][ugeth->skb_currx[rxQ]] = NULL; | 
|  | dev->stats.rx_dropped++; | 
|  | } else { | 
|  | dev->stats.rx_packets++; | 
|  | howmany++; | 
|  |  | 
|  | /* Prep the skb for the packet */ | 
|  | skb_put(skb, length); | 
|  |  | 
|  | /* Tell the skb what kind of packet this is */ | 
|  | skb->protocol = eth_type_trans(skb, ugeth->dev); | 
|  |  | 
|  | dev->stats.rx_bytes += length; | 
|  | /* Send the packet up the stack */ | 
|  | #ifdef CONFIG_UGETH_NAPI | 
|  | netif_receive_skb(skb); | 
|  | #else | 
|  | netif_rx(skb); | 
|  | #endif				/* CONFIG_UGETH_NAPI */ | 
|  | } | 
|  |  | 
|  | ugeth->dev->last_rx = jiffies; | 
|  |  | 
|  | skb = get_new_skb(ugeth, bd); | 
|  | if (!skb) { | 
|  | if (netif_msg_rx_err(ugeth)) | 
|  | ugeth_warn("%s: No Rx Data Buffer", __FUNCTION__); | 
|  | dev->stats.rx_dropped++; | 
|  | break; | 
|  | } | 
|  |  | 
|  | ugeth->rx_skbuff[rxQ][ugeth->skb_currx[rxQ]] = skb; | 
|  |  | 
|  | /* update to point at the next skb */ | 
|  | ugeth->skb_currx[rxQ] = | 
|  | (ugeth->skb_currx[rxQ] + | 
|  | 1) & RX_RING_MOD_MASK(ugeth->ug_info->bdRingLenRx[rxQ]); | 
|  |  | 
|  | if (bd_status & R_W) | 
|  | bd = ugeth->p_rx_bd_ring[rxQ]; | 
|  | else | 
|  | bd += sizeof(struct qe_bd); | 
|  |  | 
|  | bd_status = in_be32((u32 *)bd); | 
|  | } | 
|  |  | 
|  | ugeth->rxBd[rxQ] = bd; | 
|  | return howmany; | 
|  | } | 
|  |  | 
|  | static int ucc_geth_tx(struct net_device *dev, u8 txQ) | 
|  | { | 
|  | /* Start from the next BD that should be filled */ | 
|  | struct ucc_geth_private *ugeth = netdev_priv(dev); | 
|  | u8 *bd;			/* BD pointer */ | 
|  | u32 bd_status; | 
|  |  | 
|  | bd = ugeth->confBd[txQ]; | 
|  | bd_status = in_be32((u32 *)bd); | 
|  |  | 
|  | /* Normal processing. */ | 
|  | while ((bd_status & T_R) == 0) { | 
|  | /* BD contains already transmitted buffer.   */ | 
|  | /* Handle the transmitted buffer and release */ | 
|  | /* the BD to be used with the current frame  */ | 
|  |  | 
|  | if ((bd == ugeth->txBd[txQ]) && (netif_queue_stopped(dev) == 0)) | 
|  | break; | 
|  |  | 
|  | dev->stats.tx_packets++; | 
|  |  | 
|  | /* Free the sk buffer associated with this TxBD */ | 
|  | dev_kfree_skb_irq(ugeth-> | 
|  | tx_skbuff[txQ][ugeth->skb_dirtytx[txQ]]); | 
|  | ugeth->tx_skbuff[txQ][ugeth->skb_dirtytx[txQ]] = NULL; | 
|  | ugeth->skb_dirtytx[txQ] = | 
|  | (ugeth->skb_dirtytx[txQ] + | 
|  | 1) & TX_RING_MOD_MASK(ugeth->ug_info->bdRingLenTx[txQ]); | 
|  |  | 
|  | /* We freed a buffer, so now we can restart transmission */ | 
|  | if (netif_queue_stopped(dev)) | 
|  | netif_wake_queue(dev); | 
|  |  | 
|  | /* Advance the confirmation BD pointer */ | 
|  | if (!(bd_status & T_W)) | 
|  | bd += sizeof(struct qe_bd); | 
|  | else | 
|  | bd = ugeth->p_tx_bd_ring[txQ]; | 
|  | bd_status = in_be32((u32 *)bd); | 
|  | } | 
|  | ugeth->confBd[txQ] = bd; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_UGETH_NAPI | 
|  | static int ucc_geth_poll(struct napi_struct *napi, int budget) | 
|  | { | 
|  | struct ucc_geth_private *ugeth = container_of(napi, struct ucc_geth_private, napi); | 
|  | struct net_device *dev = ugeth->dev; | 
|  | struct ucc_geth_info *ug_info; | 
|  | int howmany, i; | 
|  |  | 
|  | ug_info = ugeth->ug_info; | 
|  |  | 
|  | howmany = 0; | 
|  | for (i = 0; i < ug_info->numQueuesRx; i++) | 
|  | howmany += ucc_geth_rx(ugeth, i, budget - howmany); | 
|  |  | 
|  | if (howmany < budget) { | 
|  | struct ucc_fast_private *uccf; | 
|  | u32 uccm; | 
|  |  | 
|  | netif_rx_complete(dev, napi); | 
|  | uccf = ugeth->uccf; | 
|  | uccm = in_be32(uccf->p_uccm); | 
|  | uccm |= UCCE_RX_EVENTS; | 
|  | out_be32(uccf->p_uccm, uccm); | 
|  | } | 
|  |  | 
|  | return howmany; | 
|  | } | 
|  | #endif				/* CONFIG_UGETH_NAPI */ | 
|  |  | 
|  | static irqreturn_t ucc_geth_irq_handler(int irq, void *info) | 
|  | { | 
|  | struct net_device *dev = info; | 
|  | struct ucc_geth_private *ugeth = netdev_priv(dev); | 
|  | struct ucc_fast_private *uccf; | 
|  | struct ucc_geth_info *ug_info; | 
|  | register u32 ucce; | 
|  | register u32 uccm; | 
|  | #ifndef CONFIG_UGETH_NAPI | 
|  | register u32 rx_mask; | 
|  | #endif | 
|  | register u32 tx_mask; | 
|  | u8 i; | 
|  |  | 
|  | ugeth_vdbg("%s: IN", __FUNCTION__); | 
|  |  | 
|  | uccf = ugeth->uccf; | 
|  | ug_info = ugeth->ug_info; | 
|  |  | 
|  | /* read and clear events */ | 
|  | ucce = (u32) in_be32(uccf->p_ucce); | 
|  | uccm = (u32) in_be32(uccf->p_uccm); | 
|  | ucce &= uccm; | 
|  | out_be32(uccf->p_ucce, ucce); | 
|  |  | 
|  | /* check for receive events that require processing */ | 
|  | if (ucce & UCCE_RX_EVENTS) { | 
|  | #ifdef CONFIG_UGETH_NAPI | 
|  | if (netif_rx_schedule_prep(dev, &ugeth->napi)) { | 
|  | uccm &= ~UCCE_RX_EVENTS; | 
|  | out_be32(uccf->p_uccm, uccm); | 
|  | __netif_rx_schedule(dev, &ugeth->napi); | 
|  | } | 
|  | #else | 
|  | rx_mask = UCCE_RXBF_SINGLE_MASK; | 
|  | for (i = 0; i < ug_info->numQueuesRx; i++) { | 
|  | if (ucce & rx_mask) | 
|  | ucc_geth_rx(ugeth, i, (int)ugeth->ug_info->bdRingLenRx[i]); | 
|  | ucce &= ~rx_mask; | 
|  | rx_mask <<= 1; | 
|  | } | 
|  | #endif /* CONFIG_UGETH_NAPI */ | 
|  | } | 
|  |  | 
|  | /* Tx event processing */ | 
|  | if (ucce & UCCE_TX_EVENTS) { | 
|  | spin_lock(&ugeth->lock); | 
|  | tx_mask = UCCE_TXBF_SINGLE_MASK; | 
|  | for (i = 0; i < ug_info->numQueuesTx; i++) { | 
|  | if (ucce & tx_mask) | 
|  | ucc_geth_tx(dev, i); | 
|  | ucce &= ~tx_mask; | 
|  | tx_mask <<= 1; | 
|  | } | 
|  | spin_unlock(&ugeth->lock); | 
|  | } | 
|  |  | 
|  | /* Errors and other events */ | 
|  | if (ucce & UCCE_OTHER) { | 
|  | if (ucce & UCCE_BSY) { | 
|  | dev->stats.rx_errors++; | 
|  | } | 
|  | if (ucce & UCCE_TXE) { | 
|  | dev->stats.tx_errors++; | 
|  | } | 
|  | } | 
|  |  | 
|  | return IRQ_HANDLED; | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_NET_POLL_CONTROLLER | 
|  | /* | 
|  | * Polling 'interrupt' - used by things like netconsole to send skbs | 
|  | * without having to re-enable interrupts. It's not called while | 
|  | * the interrupt routine is executing. | 
|  | */ | 
|  | static void ucc_netpoll(struct net_device *dev) | 
|  | { | 
|  | struct ucc_geth_private *ugeth = netdev_priv(dev); | 
|  | int irq = ugeth->ug_info->uf_info.irq; | 
|  |  | 
|  | disable_irq(irq); | 
|  | ucc_geth_irq_handler(irq, dev); | 
|  | enable_irq(irq); | 
|  | } | 
|  | #endif /* CONFIG_NET_POLL_CONTROLLER */ | 
|  |  | 
|  | /* Called when something needs to use the ethernet device */ | 
|  | /* Returns 0 for success. */ | 
|  | static int ucc_geth_open(struct net_device *dev) | 
|  | { | 
|  | struct ucc_geth_private *ugeth = netdev_priv(dev); | 
|  | int err; | 
|  |  | 
|  | ugeth_vdbg("%s: IN", __FUNCTION__); | 
|  |  | 
|  | /* Test station address */ | 
|  | if (dev->dev_addr[0] & ENET_GROUP_ADDR) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Multicast address used for station address" | 
|  | " - is this what you wanted?", __FUNCTION__); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | err = ucc_struct_init(ugeth); | 
|  | if (err) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Cannot configure internal struct, aborting.", dev->name); | 
|  | return err; | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_UGETH_NAPI | 
|  | napi_enable(&ugeth->napi); | 
|  | #endif | 
|  | err = ucc_geth_startup(ugeth); | 
|  | if (err) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Cannot configure net device, aborting.", | 
|  | dev->name); | 
|  | goto out_err; | 
|  | } | 
|  |  | 
|  | err = adjust_enet_interface(ugeth); | 
|  | if (err) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Cannot configure net device, aborting.", | 
|  | dev->name); | 
|  | goto out_err; | 
|  | } | 
|  |  | 
|  | /*       Set MACSTNADDR1, MACSTNADDR2                */ | 
|  | /* For more details see the hardware spec.           */ | 
|  | init_mac_station_addr_regs(dev->dev_addr[0], | 
|  | dev->dev_addr[1], | 
|  | dev->dev_addr[2], | 
|  | dev->dev_addr[3], | 
|  | dev->dev_addr[4], | 
|  | dev->dev_addr[5], | 
|  | &ugeth->ug_regs->macstnaddr1, | 
|  | &ugeth->ug_regs->macstnaddr2); | 
|  |  | 
|  | err = init_phy(dev); | 
|  | if (err) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Cannot initialize PHY, aborting.", dev->name); | 
|  | goto out_err; | 
|  | } | 
|  |  | 
|  | phy_start(ugeth->phydev); | 
|  |  | 
|  | err = | 
|  | request_irq(ugeth->ug_info->uf_info.irq, ucc_geth_irq_handler, 0, | 
|  | "UCC Geth", dev); | 
|  | if (err) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Cannot get IRQ for net device, aborting.", | 
|  | dev->name); | 
|  | ucc_geth_stop(ugeth); | 
|  | goto out_err; | 
|  | } | 
|  |  | 
|  | err = ugeth_enable(ugeth, COMM_DIR_RX_AND_TX); | 
|  | if (err) { | 
|  | if (netif_msg_ifup(ugeth)) | 
|  | ugeth_err("%s: Cannot enable net device, aborting.", dev->name); | 
|  | ucc_geth_stop(ugeth); | 
|  | goto out_err; | 
|  | } | 
|  |  | 
|  | netif_start_queue(dev); | 
|  |  | 
|  | return err; | 
|  |  | 
|  | out_err: | 
|  | #ifdef CONFIG_UGETH_NAPI | 
|  | napi_disable(&ugeth->napi); | 
|  | #endif | 
|  | return err; | 
|  | } | 
|  |  | 
|  | /* Stops the kernel queue, and halts the controller */ | 
|  | static int ucc_geth_close(struct net_device *dev) | 
|  | { | 
|  | struct ucc_geth_private *ugeth = netdev_priv(dev); | 
|  |  | 
|  | ugeth_vdbg("%s: IN", __FUNCTION__); | 
|  |  | 
|  | #ifdef CONFIG_UGETH_NAPI | 
|  | napi_disable(&ugeth->napi); | 
|  | #endif | 
|  |  | 
|  | ucc_geth_stop(ugeth); | 
|  |  | 
|  | phy_disconnect(ugeth->phydev); | 
|  | ugeth->phydev = NULL; | 
|  |  | 
|  | netif_stop_queue(dev); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static phy_interface_t to_phy_interface(const char *phy_connection_type) | 
|  | { | 
|  | if (strcasecmp(phy_connection_type, "mii") == 0) | 
|  | return PHY_INTERFACE_MODE_MII; | 
|  | if (strcasecmp(phy_connection_type, "gmii") == 0) | 
|  | return PHY_INTERFACE_MODE_GMII; | 
|  | if (strcasecmp(phy_connection_type, "tbi") == 0) | 
|  | return PHY_INTERFACE_MODE_TBI; | 
|  | if (strcasecmp(phy_connection_type, "rmii") == 0) | 
|  | return PHY_INTERFACE_MODE_RMII; | 
|  | if (strcasecmp(phy_connection_type, "rgmii") == 0) | 
|  | return PHY_INTERFACE_MODE_RGMII; | 
|  | if (strcasecmp(phy_connection_type, "rgmii-id") == 0) | 
|  | return PHY_INTERFACE_MODE_RGMII_ID; | 
|  | if (strcasecmp(phy_connection_type, "rgmii-txid") == 0) | 
|  | return PHY_INTERFACE_MODE_RGMII_TXID; | 
|  | if (strcasecmp(phy_connection_type, "rgmii-rxid") == 0) | 
|  | return PHY_INTERFACE_MODE_RGMII_RXID; | 
|  | if (strcasecmp(phy_connection_type, "rtbi") == 0) | 
|  | return PHY_INTERFACE_MODE_RTBI; | 
|  |  | 
|  | return PHY_INTERFACE_MODE_MII; | 
|  | } | 
|  |  | 
|  | static int ucc_geth_probe(struct of_device* ofdev, const struct of_device_id *match) | 
|  | { | 
|  | struct device *device = &ofdev->dev; | 
|  | struct device_node *np = ofdev->node; | 
|  | struct device_node *mdio; | 
|  | struct net_device *dev = NULL; | 
|  | struct ucc_geth_private *ugeth = NULL; | 
|  | struct ucc_geth_info *ug_info; | 
|  | struct resource res; | 
|  | struct device_node *phy; | 
|  | int err, ucc_num, max_speed = 0; | 
|  | const phandle *ph; | 
|  | const unsigned int *prop; | 
|  | const char *sprop; | 
|  | const void *mac_addr; | 
|  | phy_interface_t phy_interface; | 
|  | static const int enet_to_speed[] = { | 
|  | SPEED_10, SPEED_10, SPEED_10, | 
|  | SPEED_100, SPEED_100, SPEED_100, | 
|  | SPEED_1000, SPEED_1000, SPEED_1000, SPEED_1000, | 
|  | }; | 
|  | static const phy_interface_t enet_to_phy_interface[] = { | 
|  | PHY_INTERFACE_MODE_MII, PHY_INTERFACE_MODE_RMII, | 
|  | PHY_INTERFACE_MODE_RGMII, PHY_INTERFACE_MODE_MII, | 
|  | PHY_INTERFACE_MODE_RMII, PHY_INTERFACE_MODE_RGMII, | 
|  | PHY_INTERFACE_MODE_GMII, PHY_INTERFACE_MODE_RGMII, | 
|  | PHY_INTERFACE_MODE_TBI, PHY_INTERFACE_MODE_RTBI, | 
|  | }; | 
|  |  | 
|  | ugeth_vdbg("%s: IN", __FUNCTION__); | 
|  |  | 
|  | prop = of_get_property(np, "device-id", NULL); | 
|  | ucc_num = *prop - 1; | 
|  | if ((ucc_num < 0) || (ucc_num > 7)) | 
|  | return -ENODEV; | 
|  |  | 
|  | ug_info = &ugeth_info[ucc_num]; | 
|  | if (ug_info == NULL) { | 
|  | if (netif_msg_probe(&debug)) | 
|  | ugeth_err("%s: [%d] Missing additional data!", | 
|  | __FUNCTION__, ucc_num); | 
|  | return -ENODEV; | 
|  | } | 
|  |  | 
|  | ug_info->uf_info.ucc_num = ucc_num; | 
|  |  | 
|  | sprop = of_get_property(np, "rx-clock-name", NULL); | 
|  | if (sprop) { | 
|  | ug_info->uf_info.rx_clock = qe_clock_source(sprop); | 
|  | if ((ug_info->uf_info.rx_clock < QE_CLK_NONE) || | 
|  | (ug_info->uf_info.rx_clock > QE_CLK24)) { | 
|  | printk(KERN_ERR | 
|  | "ucc_geth: invalid rx-clock-name property\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  | } else { | 
|  | prop = of_get_property(np, "rx-clock", NULL); | 
|  | if (!prop) { | 
|  | /* If both rx-clock-name and rx-clock are missing, | 
|  | we want to tell people to use rx-clock-name. */ | 
|  | printk(KERN_ERR | 
|  | "ucc_geth: missing rx-clock-name property\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  | if ((*prop < QE_CLK_NONE) || (*prop > QE_CLK24)) { | 
|  | printk(KERN_ERR | 
|  | "ucc_geth: invalid rx-clock propperty\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  | ug_info->uf_info.rx_clock = *prop; | 
|  | } | 
|  |  | 
|  | sprop = of_get_property(np, "tx-clock-name", NULL); | 
|  | if (sprop) { | 
|  | ug_info->uf_info.tx_clock = qe_clock_source(sprop); | 
|  | if ((ug_info->uf_info.tx_clock < QE_CLK_NONE) || | 
|  | (ug_info->uf_info.tx_clock > QE_CLK24)) { | 
|  | printk(KERN_ERR | 
|  | "ucc_geth: invalid tx-clock-name property\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  | } else { | 
|  | prop = of_get_property(np, "rx-clock", NULL); | 
|  | if (!prop) { | 
|  | printk(KERN_ERR | 
|  | "ucc_geth: mising tx-clock-name property\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  | if ((*prop < QE_CLK_NONE) || (*prop > QE_CLK24)) { | 
|  | printk(KERN_ERR | 
|  | "ucc_geth: invalid tx-clock property\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  | ug_info->uf_info.tx_clock = *prop; | 
|  | } | 
|  |  | 
|  | err = of_address_to_resource(np, 0, &res); | 
|  | if (err) | 
|  | return -EINVAL; | 
|  |  | 
|  | ug_info->uf_info.regs = res.start; | 
|  | ug_info->uf_info.irq = irq_of_parse_and_map(np, 0); | 
|  |  | 
|  | ph = of_get_property(np, "phy-handle", NULL); | 
|  | phy = of_find_node_by_phandle(*ph); | 
|  |  | 
|  | if (phy == NULL) | 
|  | return -ENODEV; | 
|  |  | 
|  | /* set the PHY address */ | 
|  | prop = of_get_property(phy, "reg", NULL); | 
|  | if (prop == NULL) | 
|  | return -1; | 
|  | ug_info->phy_address = *prop; | 
|  |  | 
|  | /* get the phy interface type, or default to MII */ | 
|  | prop = of_get_property(np, "phy-connection-type", NULL); | 
|  | if (!prop) { | 
|  | /* handle interface property present in old trees */ | 
|  | prop = of_get_property(phy, "interface", NULL); | 
|  | if (prop != NULL) { | 
|  | phy_interface = enet_to_phy_interface[*prop]; | 
|  | max_speed = enet_to_speed[*prop]; | 
|  | } else | 
|  | phy_interface = PHY_INTERFACE_MODE_MII; | 
|  | } else { | 
|  | phy_interface = to_phy_interface((const char *)prop); | 
|  | } | 
|  |  | 
|  | /* get speed, or derive from PHY interface */ | 
|  | if (max_speed == 0) | 
|  | switch (phy_interface) { | 
|  | case PHY_INTERFACE_MODE_GMII: | 
|  | case PHY_INTERFACE_MODE_RGMII: | 
|  | case PHY_INTERFACE_MODE_RGMII_ID: | 
|  | case PHY_INTERFACE_MODE_RGMII_RXID: | 
|  | case PHY_INTERFACE_MODE_RGMII_TXID: | 
|  | case PHY_INTERFACE_MODE_TBI: | 
|  | case PHY_INTERFACE_MODE_RTBI: | 
|  | max_speed = SPEED_1000; | 
|  | break; | 
|  | default: | 
|  | max_speed = SPEED_100; | 
|  | break; | 
|  | } | 
|  |  | 
|  | if (max_speed == SPEED_1000) { | 
|  | /* configure muram FIFOs for gigabit operation */ | 
|  | ug_info->uf_info.urfs = UCC_GETH_URFS_GIGA_INIT; | 
|  | ug_info->uf_info.urfet = UCC_GETH_URFET_GIGA_INIT; | 
|  | ug_info->uf_info.urfset = UCC_GETH_URFSET_GIGA_INIT; | 
|  | ug_info->uf_info.utfs = UCC_GETH_UTFS_GIGA_INIT; | 
|  | ug_info->uf_info.utfet = UCC_GETH_UTFET_GIGA_INIT; | 
|  | ug_info->uf_info.utftt = UCC_GETH_UTFTT_GIGA_INIT; | 
|  | } | 
|  |  | 
|  | /* Set the bus id */ | 
|  | mdio = of_get_parent(phy); | 
|  |  | 
|  | if (mdio == NULL) | 
|  | return -1; | 
|  |  | 
|  | err = of_address_to_resource(mdio, 0, &res); | 
|  | of_node_put(mdio); | 
|  |  | 
|  | if (err) | 
|  | return -1; | 
|  |  | 
|  | ug_info->mdio_bus = res.start; | 
|  |  | 
|  | if (netif_msg_probe(&debug)) | 
|  | printk(KERN_INFO "ucc_geth: UCC%1d at 0x%8x (irq = %d) \n", | 
|  | ug_info->uf_info.ucc_num + 1, ug_info->uf_info.regs, | 
|  | ug_info->uf_info.irq); | 
|  |  | 
|  | /* Create an ethernet device instance */ | 
|  | dev = alloc_etherdev(sizeof(*ugeth)); | 
|  |  | 
|  | if (dev == NULL) | 
|  | return -ENOMEM; | 
|  |  | 
|  | ugeth = netdev_priv(dev); | 
|  | spin_lock_init(&ugeth->lock); | 
|  |  | 
|  | /* Create CQs for hash tables */ | 
|  | INIT_LIST_HEAD(&ugeth->group_hash_q); | 
|  | INIT_LIST_HEAD(&ugeth->ind_hash_q); | 
|  |  | 
|  | dev_set_drvdata(device, dev); | 
|  |  | 
|  | /* Set the dev->base_addr to the gfar reg region */ | 
|  | dev->base_addr = (unsigned long)(ug_info->uf_info.regs); | 
|  |  | 
|  | SET_NETDEV_DEV(dev, device); | 
|  |  | 
|  | /* Fill in the dev structure */ | 
|  | uec_set_ethtool_ops(dev); | 
|  | dev->open = ucc_geth_open; | 
|  | dev->hard_start_xmit = ucc_geth_start_xmit; | 
|  | dev->tx_timeout = ucc_geth_timeout; | 
|  | dev->watchdog_timeo = TX_TIMEOUT; | 
|  | #ifdef CONFIG_UGETH_NAPI | 
|  | netif_napi_add(dev, &ugeth->napi, ucc_geth_poll, UCC_GETH_DEV_WEIGHT); | 
|  | #endif				/* CONFIG_UGETH_NAPI */ | 
|  | #ifdef CONFIG_NET_POLL_CONTROLLER | 
|  | dev->poll_controller = ucc_netpoll; | 
|  | #endif | 
|  | dev->stop = ucc_geth_close; | 
|  | //    dev->change_mtu = ucc_geth_change_mtu; | 
|  | dev->mtu = 1500; | 
|  | dev->set_multicast_list = ucc_geth_set_multi; | 
|  |  | 
|  | ugeth->msg_enable = netif_msg_init(debug.msg_enable, UGETH_MSG_DEFAULT); | 
|  | ugeth->phy_interface = phy_interface; | 
|  | ugeth->max_speed = max_speed; | 
|  |  | 
|  | err = register_netdev(dev); | 
|  | if (err) { | 
|  | if (netif_msg_probe(ugeth)) | 
|  | ugeth_err("%s: Cannot register net device, aborting.", | 
|  | dev->name); | 
|  | free_netdev(dev); | 
|  | return err; | 
|  | } | 
|  |  | 
|  | mac_addr = of_get_mac_address(np); | 
|  | if (mac_addr) | 
|  | memcpy(dev->dev_addr, mac_addr, 6); | 
|  |  | 
|  | ugeth->ug_info = ug_info; | 
|  | ugeth->dev = dev; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int ucc_geth_remove(struct of_device* ofdev) | 
|  | { | 
|  | struct device *device = &ofdev->dev; | 
|  | struct net_device *dev = dev_get_drvdata(device); | 
|  | struct ucc_geth_private *ugeth = netdev_priv(dev); | 
|  |  | 
|  | unregister_netdev(dev); | 
|  | free_netdev(dev); | 
|  | ucc_geth_memclean(ugeth); | 
|  | dev_set_drvdata(device, NULL); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static struct of_device_id ucc_geth_match[] = { | 
|  | { | 
|  | .type = "network", | 
|  | .compatible = "ucc_geth", | 
|  | }, | 
|  | {}, | 
|  | }; | 
|  |  | 
|  | MODULE_DEVICE_TABLE(of, ucc_geth_match); | 
|  |  | 
|  | static struct of_platform_driver ucc_geth_driver = { | 
|  | .name		= DRV_NAME, | 
|  | .match_table	= ucc_geth_match, | 
|  | .probe		= ucc_geth_probe, | 
|  | .remove		= ucc_geth_remove, | 
|  | }; | 
|  |  | 
|  | static int __init ucc_geth_init(void) | 
|  | { | 
|  | int i, ret; | 
|  |  | 
|  | ret = uec_mdio_init(); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | if (netif_msg_drv(&debug)) | 
|  | printk(KERN_INFO "ucc_geth: " DRV_DESC "\n"); | 
|  | for (i = 0; i < 8; i++) | 
|  | memcpy(&(ugeth_info[i]), &ugeth_primary_info, | 
|  | sizeof(ugeth_primary_info)); | 
|  |  | 
|  | ret = of_register_platform_driver(&ucc_geth_driver); | 
|  |  | 
|  | if (ret) | 
|  | uec_mdio_exit(); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void __exit ucc_geth_exit(void) | 
|  | { | 
|  | of_unregister_platform_driver(&ucc_geth_driver); | 
|  | uec_mdio_exit(); | 
|  | } | 
|  |  | 
|  | module_init(ucc_geth_init); | 
|  | module_exit(ucc_geth_exit); | 
|  |  | 
|  | MODULE_AUTHOR("Freescale Semiconductor, Inc"); | 
|  | MODULE_DESCRIPTION(DRV_DESC); | 
|  | MODULE_VERSION(DRV_VERSION); | 
|  | MODULE_LICENSE("GPL"); |