From patchwork Fri Jan 6 22:00:14 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Anirudh Venkataramanan X-Patchwork-Id: 13091844 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 225E3C63797 for ; Fri, 6 Jan 2023 21:59:37 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236578AbjAFV7f (ORCPT ); Fri, 6 Jan 2023 16:59:35 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:42356 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235636AbjAFV7W (ORCPT ); Fri, 6 Jan 2023 16:59:22 -0500 Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 307AF7FEEF; Fri, 6 Jan 2023 13:59:18 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1673042358; x=1704578358; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=oDd3aylN3SvVgL+M1nZfwwvrlYyb7VomQOFhKrUb7ZY=; b=SlVYxqIH1XSDS8zIio2hNp1gD+KNxb8f+qK6sQq4NWSZ0fgIr8SZDn69 6Vizi1vCYGiYu8t6pNGbaaVe6bUXHiX8ECAI+Xb2n7NxO9l/343NhdHdB 0nVFuwGgyFCl2r7lAfk4HRFjaVWQM+lzp1pkAaoLolswQeil1lnlU1/3I I5Up1XkPkKVzSvKA9OTIAPoqL+Z1Gfj5Z/+zhIxDlKxmNHO3ZaCzCymZc 6v+uysoDmfjgZ1Bgr6c6++Ivi2hERYVvsVDZlVwLvd+t+p1F2TQc4o1ac Q0/1Iq6gPgWm4mvX/8LDFVr0kO39FnY8sWzA+mvtLVUqMI7t2eO9kwmTU w==; X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="387030700" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="387030700" Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Jan 2023 13:59:16 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="763652877" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="763652877" Received: from avenkata-desk0.sc.intel.com ([172.25.112.60]) by fmsmga002.fm.intel.com with ESMTP; 06 Jan 2023 13:59:15 -0800 From: Anirudh Venkataramanan To: netdev@vger.kernel.org Cc: linux-pci@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-mips@vger.kernel.org, linux-trace-kernel@vger.kernel.org, sparclinux@vger.kernel.org, Anirudh Venkataramanan , Leon Romanovsky Subject: [PATCH net-next 1/7] ethernet: Remove the Sun Cassini driver Date: Fri, 6 Jan 2023 14:00:14 -0800 Message-Id: <20230106220020.1820147-2-anirudh.venkataramanan@intel.com> X-Mailer: git-send-email 2.37.2 In-Reply-To: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> References: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org In a recent patch series that touched this driver [1], it was suggested that this driver should be removed completely. git logs suggest that there hasn't been any significant feature addition, improvement or fixes to user-visible bugs in a while. A web search didn't indicate any recent discussions or any evidence that there are users out there who care about this driver. Thus, remove this driver. Notes: checkpatch complains "WARNING: added, moved or deleted file(s), does MAINTAINERS need updating?". The files being removed don't have their own entries in the MAINTAINERS file, so there's nothing to remove. checkpatch also complains about the long lore link below. [1] https://lore.kernel.org/netdev/99629223-ac1b-0f82-50b8-ea307b3b0197@intel.com/T/#t Suggested-by: Leon Romanovsky Signed-off-by: Anirudh Venkataramanan --- drivers/net/ethernet/sun/Kconfig | 8 - drivers/net/ethernet/sun/Makefile | 1 - drivers/net/ethernet/sun/cassini.c | 5215 ---------------------------- drivers/net/ethernet/sun/cassini.h | 2900 ---------------- 4 files changed, 8124 deletions(-) delete mode 100644 drivers/net/ethernet/sun/cassini.c delete mode 100644 drivers/net/ethernet/sun/cassini.h diff --git a/drivers/net/ethernet/sun/Kconfig b/drivers/net/ethernet/sun/Kconfig index b0d3f9a..57eb168 100644 --- a/drivers/net/ethernet/sun/Kconfig +++ b/drivers/net/ethernet/sun/Kconfig @@ -62,14 +62,6 @@ config SUNGEM Support for the Sun GEM chip, aka Sun GigabitEthernet/P 2.0. See also . -config CASSINI - tristate "Sun Cassini support" - depends on PCI - select CRC32 - help - Support for the Sun Cassini chip, aka Sun GigaSwift Ethernet. See also - . - config SUNVNET_COMMON tristate "Common routines to support Sun Virtual Networking" depends on SUN_LDOMS diff --git a/drivers/net/ethernet/sun/Makefile b/drivers/net/ethernet/sun/Makefile index 9a5249d..fd77810 100644 --- a/drivers/net/ethernet/sun/Makefile +++ b/drivers/net/ethernet/sun/Makefile @@ -7,7 +7,6 @@ obj-$(CONFIG_HAPPYMEAL) += sunhme.o obj-$(CONFIG_SUNQE) += sunqe.o obj-$(CONFIG_SUNBMAC) += sunbmac.o obj-$(CONFIG_SUNGEM) += sungem.o -obj-$(CONFIG_CASSINI) += cassini.o obj-$(CONFIG_SUNVNET_COMMON) += sunvnet_common.o obj-$(CONFIG_SUNVNET) += sunvnet.o obj-$(CONFIG_LDMVSW) += ldmvsw.o diff --git a/drivers/net/ethernet/sun/cassini.c b/drivers/net/ethernet/sun/cassini.c deleted file mode 100644 index 4ef05ba..00000000 --- a/drivers/net/ethernet/sun/cassini.c +++ /dev/null @@ -1,5215 +0,0 @@ -// SPDX-License-Identifier: GPL-2.0+ -/* cassini.c: Sun Microsystems Cassini(+) ethernet driver. - * - * Copyright (C) 2004 Sun Microsystems Inc. - * Copyright (C) 2003 Adrian Sun (asun@darksunrising.com) - * - * This driver uses the sungem driver (c) David Miller - * (davem@redhat.com) as its basis. - * - * The cassini chip has a number of features that distinguish it from - * the gem chip: - * 4 transmit descriptor rings that are used for either QoS (VLAN) or - * load balancing (non-VLAN mode) - * batching of multiple packets - * multiple CPU dispatching - * page-based RX descriptor engine with separate completion rings - * Gigabit support (GMII and PCS interface) - * MIF link up/down detection works - * - * RX is handled by page sized buffers that are attached as fragments to - * the skb. here's what's done: - * -- driver allocates pages at a time and keeps reference counts - * on them. - * -- the upper protocol layers assume that the header is in the skb - * itself. as a result, cassini will copy a small amount (64 bytes) - * to make them happy. - * -- driver appends the rest of the data pages as frags to skbuffs - * and increments the reference count - * -- on page reclamation, the driver swaps the page with a spare page. - * if that page is still in use, it frees its reference to that page, - * and allocates a new page for use. otherwise, it just recycles the - * page. - * - * NOTE: cassini can parse the header. however, it's not worth it - * as long as the network stack requires a header copy. - * - * TX has 4 queues. currently these queues are used in a round-robin - * fashion for load balancing. They can also be used for QoS. for that - * to work, however, QoS information needs to be exposed down to the driver - * level so that subqueues get targeted to particular transmit rings. - * alternatively, the queues can be configured via use of the all-purpose - * ioctl. - * - * RX DATA: the rx completion ring has all the info, but the rx desc - * ring has all of the data. RX can conceivably come in under multiple - * interrupts, but the INT# assignment needs to be set up properly by - * the BIOS and conveyed to the driver. PCI BIOSes don't know how to do - * that. also, the two descriptor rings are designed to distinguish between - * encrypted and non-encrypted packets, but we use them for buffering - * instead. - * - * by default, the selective clear mask is set up to process rx packets. - */ - -#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt - -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include - -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include - -#include - -#include -#include -#include -#include -#include - -#define CAS_NCPUS num_online_cpus() - -#define cas_skb_release(x) netif_rx(x) - -/* select which firmware to use */ -#define USE_HP_WORKAROUND -#define HP_WORKAROUND_DEFAULT /* select which firmware to use as default */ -#define CAS_HP_ALT_FIRMWARE cas_prog_null /* alternate firmware */ - -#include "cassini.h" - -#define USE_TX_COMPWB /* use completion writeback registers */ -#define USE_CSMA_CD_PROTO /* standard CSMA/CD */ -#define USE_RX_BLANK /* hw interrupt mitigation */ -#undef USE_ENTROPY_DEV /* don't test for entropy device */ - -/* NOTE: these aren't useable unless PCI interrupts can be assigned. - * also, we need to make cp->lock finer-grained. - */ -#undef USE_PCI_INTB -#undef USE_PCI_INTC -#undef USE_PCI_INTD -#undef USE_QOS - -#undef USE_VPD_DEBUG /* debug vpd information if defined */ - -/* rx processing options */ -#define USE_PAGE_ORDER /* specify to allocate large rx pages */ -#define RX_DONT_BATCH 0 /* if 1, don't batch flows */ -#define RX_COPY_ALWAYS 0 /* if 0, use frags */ -#define RX_COPY_MIN 64 /* copy a little to make upper layers happy */ -#undef RX_COUNT_BUFFERS /* define to calculate RX buffer stats */ - -#define DRV_MODULE_NAME "cassini" -#define DRV_MODULE_VERSION "1.6" -#define DRV_MODULE_RELDATE "21 May 2008" - -#define CAS_DEF_MSG_ENABLE \ - (NETIF_MSG_DRV | \ - NETIF_MSG_PROBE | \ - NETIF_MSG_LINK | \ - NETIF_MSG_TIMER | \ - NETIF_MSG_IFDOWN | \ - NETIF_MSG_IFUP | \ - NETIF_MSG_RX_ERR | \ - NETIF_MSG_TX_ERR) - -/* length of time before we decide the hardware is borked, - * and dev->tx_timeout() should be called to fix the problem - */ -#define CAS_TX_TIMEOUT (HZ) -#define CAS_LINK_TIMEOUT (22*HZ/10) -#define CAS_LINK_FAST_TIMEOUT (1) - -/* timeout values for state changing. these specify the number - * of 10us delays to be used before giving up. - */ -#define STOP_TRIES_PHY 1000 -#define STOP_TRIES 5000 - -/* specify a minimum frame size to deal with some fifo issues - * max mtu == 2 * page size - ethernet header - 64 - swivel = - * 2 * page_size - 0x50 - */ -#define CAS_MIN_FRAME 97 -#define CAS_1000MB_MIN_FRAME 255 -#define CAS_MIN_MTU 60 -#define CAS_MAX_MTU min(((cp->page_size << 1) - 0x50), 9000) - -#if 1 -/* - * Eliminate these and use separate atomic counters for each, to - * avoid a race condition. - */ -#else -#define CAS_RESET_MTU 1 -#define CAS_RESET_ALL 2 -#define CAS_RESET_SPARE 3 -#endif - -static char version[] = - DRV_MODULE_NAME ".c:v" DRV_MODULE_VERSION " (" DRV_MODULE_RELDATE ")\n"; - -static int cassini_debug = -1; /* -1 == use CAS_DEF_MSG_ENABLE as value */ -static int link_mode; - -MODULE_AUTHOR("Adrian Sun (asun@darksunrising.com)"); -MODULE_DESCRIPTION("Sun Cassini(+) ethernet driver"); -MODULE_LICENSE("GPL"); -MODULE_FIRMWARE("sun/cassini.bin"); -module_param(cassini_debug, int, 0); -MODULE_PARM_DESC(cassini_debug, "Cassini bitmapped debugging message enable value"); -module_param(link_mode, int, 0); -MODULE_PARM_DESC(link_mode, "default link mode"); - -/* - * Work around for a PCS bug in which the link goes down due to the chip - * being confused and never showing a link status of "up." - */ -#define DEFAULT_LINKDOWN_TIMEOUT 5 -/* - * Value in seconds, for user input. - */ -static int linkdown_timeout = DEFAULT_LINKDOWN_TIMEOUT; -module_param(linkdown_timeout, int, 0); -MODULE_PARM_DESC(linkdown_timeout, -"min reset interval in sec. for PCS linkdown issue; disabled if not positive"); - -/* - * value in 'ticks' (units used by jiffies). Set when we init the - * module because 'HZ' in actually a function call on some flavors of - * Linux. This will default to DEFAULT_LINKDOWN_TIMEOUT * HZ. - */ -static int link_transition_timeout; - - - -static u16 link_modes[] = { - BMCR_ANENABLE, /* 0 : autoneg */ - 0, /* 1 : 10bt half duplex */ - BMCR_SPEED100, /* 2 : 100bt half duplex */ - BMCR_FULLDPLX, /* 3 : 10bt full duplex */ - BMCR_SPEED100|BMCR_FULLDPLX, /* 4 : 100bt full duplex */ - CAS_BMCR_SPEED1000|BMCR_FULLDPLX /* 5 : 1000bt full duplex */ -}; - -static const struct pci_device_id cas_pci_tbl[] = { - { PCI_VENDOR_ID_SUN, PCI_DEVICE_ID_SUN_CASSINI, - PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL }, - { PCI_VENDOR_ID_NS, PCI_DEVICE_ID_NS_SATURN, - PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0UL }, - { 0, } -}; - -MODULE_DEVICE_TABLE(pci, cas_pci_tbl); - -static void cas_set_link_modes(struct cas *cp); - -static inline void cas_lock_tx(struct cas *cp) -{ - int i; - - for (i = 0; i < N_TX_RINGS; i++) - spin_lock_nested(&cp->tx_lock[i], i); -} - -/* WTZ: QA was finding deadlock problems with the previous - * versions after long test runs with multiple cards per machine. - * See if replacing cas_lock_all with safer versions helps. The - * symptoms QA is reporting match those we'd expect if interrupts - * aren't being properly restored, and we fixed a previous deadlock - * with similar symptoms by using save/restore versions in other - * places. - */ -#define cas_lock_all_save(cp, flags) \ -do { \ - struct cas *xxxcp = (cp); \ - spin_lock_irqsave(&xxxcp->lock, flags); \ - cas_lock_tx(xxxcp); \ -} while (0) - -static inline void cas_unlock_tx(struct cas *cp) -{ - int i; - - for (i = N_TX_RINGS; i > 0; i--) - spin_unlock(&cp->tx_lock[i - 1]); -} - -#define cas_unlock_all_restore(cp, flags) \ -do { \ - struct cas *xxxcp = (cp); \ - cas_unlock_tx(xxxcp); \ - spin_unlock_irqrestore(&xxxcp->lock, flags); \ -} while (0) - -static void cas_disable_irq(struct cas *cp, const int ring) -{ - /* Make sure we won't get any more interrupts */ - if (ring == 0) { - writel(0xFFFFFFFF, cp->regs + REG_INTR_MASK); - return; - } - - /* disable completion interrupts and selectively mask */ - if (cp->cas_flags & CAS_FLAG_REG_PLUS) { - switch (ring) { -#if defined (USE_PCI_INTB) || defined(USE_PCI_INTC) || defined(USE_PCI_INTD) -#ifdef USE_PCI_INTB - case 1: -#endif -#ifdef USE_PCI_INTC - case 2: -#endif -#ifdef USE_PCI_INTD - case 3: -#endif - writel(INTRN_MASK_CLEAR_ALL | INTRN_MASK_RX_EN, - cp->regs + REG_PLUS_INTRN_MASK(ring)); - break; -#endif - default: - writel(INTRN_MASK_CLEAR_ALL, cp->regs + - REG_PLUS_INTRN_MASK(ring)); - break; - } - } -} - -static inline void cas_mask_intr(struct cas *cp) -{ - int i; - - for (i = 0; i < N_RX_COMP_RINGS; i++) - cas_disable_irq(cp, i); -} - -static void cas_enable_irq(struct cas *cp, const int ring) -{ - if (ring == 0) { /* all but TX_DONE */ - writel(INTR_TX_DONE, cp->regs + REG_INTR_MASK); - return; - } - - if (cp->cas_flags & CAS_FLAG_REG_PLUS) { - switch (ring) { -#if defined (USE_PCI_INTB) || defined(USE_PCI_INTC) || defined(USE_PCI_INTD) -#ifdef USE_PCI_INTB - case 1: -#endif -#ifdef USE_PCI_INTC - case 2: -#endif -#ifdef USE_PCI_INTD - case 3: -#endif - writel(INTRN_MASK_RX_EN, cp->regs + - REG_PLUS_INTRN_MASK(ring)); - break; -#endif - default: - break; - } - } -} - -static inline void cas_unmask_intr(struct cas *cp) -{ - int i; - - for (i = 0; i < N_RX_COMP_RINGS; i++) - cas_enable_irq(cp, i); -} - -static inline void cas_entropy_gather(struct cas *cp) -{ -#ifdef USE_ENTROPY_DEV - if ((cp->cas_flags & CAS_FLAG_ENTROPY_DEV) == 0) - return; - - batch_entropy_store(readl(cp->regs + REG_ENTROPY_IV), - readl(cp->regs + REG_ENTROPY_IV), - sizeof(uint64_t)*8); -#endif -} - -static inline void cas_entropy_reset(struct cas *cp) -{ -#ifdef USE_ENTROPY_DEV - if ((cp->cas_flags & CAS_FLAG_ENTROPY_DEV) == 0) - return; - - writel(BIM_LOCAL_DEV_PAD | BIM_LOCAL_DEV_PROM | BIM_LOCAL_DEV_EXT, - cp->regs + REG_BIM_LOCAL_DEV_EN); - writeb(ENTROPY_RESET_STC_MODE, cp->regs + REG_ENTROPY_RESET); - writeb(0x55, cp->regs + REG_ENTROPY_RAND_REG); - - /* if we read back 0x0, we don't have an entropy device */ - if (readb(cp->regs + REG_ENTROPY_RAND_REG) == 0) - cp->cas_flags &= ~CAS_FLAG_ENTROPY_DEV; -#endif -} - -/* access to the phy. the following assumes that we've initialized the MIF to - * be in frame rather than bit-bang mode - */ -static u16 cas_phy_read(struct cas *cp, int reg) -{ - u32 cmd; - int limit = STOP_TRIES_PHY; - - cmd = MIF_FRAME_ST | MIF_FRAME_OP_READ; - cmd |= CAS_BASE(MIF_FRAME_PHY_ADDR, cp->phy_addr); - cmd |= CAS_BASE(MIF_FRAME_REG_ADDR, reg); - cmd |= MIF_FRAME_TURN_AROUND_MSB; - writel(cmd, cp->regs + REG_MIF_FRAME); - - /* poll for completion */ - while (limit-- > 0) { - udelay(10); - cmd = readl(cp->regs + REG_MIF_FRAME); - if (cmd & MIF_FRAME_TURN_AROUND_LSB) - return cmd & MIF_FRAME_DATA_MASK; - } - return 0xFFFF; /* -1 */ -} - -static int cas_phy_write(struct cas *cp, int reg, u16 val) -{ - int limit = STOP_TRIES_PHY; - u32 cmd; - - cmd = MIF_FRAME_ST | MIF_FRAME_OP_WRITE; - cmd |= CAS_BASE(MIF_FRAME_PHY_ADDR, cp->phy_addr); - cmd |= CAS_BASE(MIF_FRAME_REG_ADDR, reg); - cmd |= MIF_FRAME_TURN_AROUND_MSB; - cmd |= val & MIF_FRAME_DATA_MASK; - writel(cmd, cp->regs + REG_MIF_FRAME); - - /* poll for completion */ - while (limit-- > 0) { - udelay(10); - cmd = readl(cp->regs + REG_MIF_FRAME); - if (cmd & MIF_FRAME_TURN_AROUND_LSB) - return 0; - } - return -1; -} - -static void cas_phy_powerup(struct cas *cp) -{ - u16 ctl = cas_phy_read(cp, MII_BMCR); - - if ((ctl & BMCR_PDOWN) == 0) - return; - ctl &= ~BMCR_PDOWN; - cas_phy_write(cp, MII_BMCR, ctl); -} - -static void cas_phy_powerdown(struct cas *cp) -{ - u16 ctl = cas_phy_read(cp, MII_BMCR); - - if (ctl & BMCR_PDOWN) - return; - ctl |= BMCR_PDOWN; - cas_phy_write(cp, MII_BMCR, ctl); -} - -/* cp->lock held. note: the last put_page will free the buffer */ -static int cas_page_free(struct cas *cp, cas_page_t *page) -{ - dma_unmap_page(&cp->pdev->dev, page->dma_addr, cp->page_size, - DMA_FROM_DEVICE); - __free_pages(page->buffer, cp->page_order); - kfree(page); - return 0; -} - -#ifdef RX_COUNT_BUFFERS -#define RX_USED_ADD(x, y) ((x)->used += (y)) -#define RX_USED_SET(x, y) ((x)->used = (y)) -#else -#define RX_USED_ADD(x, y) do { } while(0) -#define RX_USED_SET(x, y) do { } while(0) -#endif - -/* local page allocation routines for the receive buffers. jumbo pages - * require at least 8K contiguous and 8K aligned buffers. - */ -static cas_page_t *cas_page_alloc(struct cas *cp, const gfp_t flags) -{ - cas_page_t *page; - - page = kmalloc(sizeof(cas_page_t), flags); - if (!page) - return NULL; - - INIT_LIST_HEAD(&page->list); - RX_USED_SET(page, 0); - page->buffer = alloc_pages(flags, cp->page_order); - if (!page->buffer) - goto page_err; - page->dma_addr = dma_map_page(&cp->pdev->dev, page->buffer, 0, - cp->page_size, DMA_FROM_DEVICE); - return page; - -page_err: - kfree(page); - return NULL; -} - -/* initialize spare pool of rx buffers, but allocate during the open */ -static void cas_spare_init(struct cas *cp) -{ - spin_lock(&cp->rx_inuse_lock); - INIT_LIST_HEAD(&cp->rx_inuse_list); - spin_unlock(&cp->rx_inuse_lock); - - spin_lock(&cp->rx_spare_lock); - INIT_LIST_HEAD(&cp->rx_spare_list); - cp->rx_spares_needed = RX_SPARE_COUNT; - spin_unlock(&cp->rx_spare_lock); -} - -/* used on close. free all the spare buffers. */ -static void cas_spare_free(struct cas *cp) -{ - struct list_head list, *elem, *tmp; - - /* free spare buffers */ - INIT_LIST_HEAD(&list); - spin_lock(&cp->rx_spare_lock); - list_splice_init(&cp->rx_spare_list, &list); - spin_unlock(&cp->rx_spare_lock); - list_for_each_safe(elem, tmp, &list) { - cas_page_free(cp, list_entry(elem, cas_page_t, list)); - } - - INIT_LIST_HEAD(&list); -#if 1 - /* - * Looks like Adrian had protected this with a different - * lock than used everywhere else to manipulate this list. - */ - spin_lock(&cp->rx_inuse_lock); - list_splice_init(&cp->rx_inuse_list, &list); - spin_unlock(&cp->rx_inuse_lock); -#else - spin_lock(&cp->rx_spare_lock); - list_splice_init(&cp->rx_inuse_list, &list); - spin_unlock(&cp->rx_spare_lock); -#endif - list_for_each_safe(elem, tmp, &list) { - cas_page_free(cp, list_entry(elem, cas_page_t, list)); - } -} - -/* replenish spares if needed */ -static void cas_spare_recover(struct cas *cp, const gfp_t flags) -{ - struct list_head list, *elem, *tmp; - int needed, i; - - /* check inuse list. if we don't need any more free buffers, - * just free it - */ - - /* make a local copy of the list */ - INIT_LIST_HEAD(&list); - spin_lock(&cp->rx_inuse_lock); - list_splice_init(&cp->rx_inuse_list, &list); - spin_unlock(&cp->rx_inuse_lock); - - list_for_each_safe(elem, tmp, &list) { - cas_page_t *page = list_entry(elem, cas_page_t, list); - - /* - * With the lockless pagecache, cassini buffering scheme gets - * slightly less accurate: we might find that a page has an - * elevated reference count here, due to a speculative ref, - * and skip it as in-use. Ideally we would be able to reclaim - * it. However this would be such a rare case, it doesn't - * matter too much as we should pick it up the next time round. - * - * Importantly, if we find that the page has a refcount of 1 - * here (our refcount), then we know it is definitely not inuse - * so we can reuse it. - */ - if (page_count(page->buffer) > 1) - continue; - - list_del(elem); - spin_lock(&cp->rx_spare_lock); - if (cp->rx_spares_needed > 0) { - list_add(elem, &cp->rx_spare_list); - cp->rx_spares_needed--; - spin_unlock(&cp->rx_spare_lock); - } else { - spin_unlock(&cp->rx_spare_lock); - cas_page_free(cp, page); - } - } - - /* put any inuse buffers back on the list */ - if (!list_empty(&list)) { - spin_lock(&cp->rx_inuse_lock); - list_splice(&list, &cp->rx_inuse_list); - spin_unlock(&cp->rx_inuse_lock); - } - - spin_lock(&cp->rx_spare_lock); - needed = cp->rx_spares_needed; - spin_unlock(&cp->rx_spare_lock); - if (!needed) - return; - - /* we still need spares, so try to allocate some */ - INIT_LIST_HEAD(&list); - i = 0; - while (i < needed) { - cas_page_t *spare = cas_page_alloc(cp, flags); - if (!spare) - break; - list_add(&spare->list, &list); - i++; - } - - spin_lock(&cp->rx_spare_lock); - list_splice(&list, &cp->rx_spare_list); - cp->rx_spares_needed -= i; - spin_unlock(&cp->rx_spare_lock); -} - -/* pull a page from the list. */ -static cas_page_t *cas_page_dequeue(struct cas *cp) -{ - struct list_head *entry; - int recover; - - spin_lock(&cp->rx_spare_lock); - if (list_empty(&cp->rx_spare_list)) { - /* try to do a quick recovery */ - spin_unlock(&cp->rx_spare_lock); - cas_spare_recover(cp, GFP_ATOMIC); - spin_lock(&cp->rx_spare_lock); - if (list_empty(&cp->rx_spare_list)) { - netif_err(cp, rx_err, cp->dev, - "no spare buffers available\n"); - spin_unlock(&cp->rx_spare_lock); - return NULL; - } - } - - entry = cp->rx_spare_list.next; - list_del(entry); - recover = ++cp->rx_spares_needed; - spin_unlock(&cp->rx_spare_lock); - - /* trigger the timer to do the recovery */ - if ((recover & (RX_SPARE_RECOVER_VAL - 1)) == 0) { -#if 1 - atomic_inc(&cp->reset_task_pending); - atomic_inc(&cp->reset_task_pending_spare); - schedule_work(&cp->reset_task); -#else - atomic_set(&cp->reset_task_pending, CAS_RESET_SPARE); - schedule_work(&cp->reset_task); -#endif - } - return list_entry(entry, cas_page_t, list); -} - - -static void cas_mif_poll(struct cas *cp, const int enable) -{ - u32 cfg; - - cfg = readl(cp->regs + REG_MIF_CFG); - cfg &= (MIF_CFG_MDIO_0 | MIF_CFG_MDIO_1); - - if (cp->phy_type & CAS_PHY_MII_MDIO1) - cfg |= MIF_CFG_PHY_SELECT; - - /* poll and interrupt on link status change. */ - if (enable) { - cfg |= MIF_CFG_POLL_EN; - cfg |= CAS_BASE(MIF_CFG_POLL_REG, MII_BMSR); - cfg |= CAS_BASE(MIF_CFG_POLL_PHY, cp->phy_addr); - } - writel((enable) ? ~(BMSR_LSTATUS | BMSR_ANEGCOMPLETE) : 0xFFFF, - cp->regs + REG_MIF_MASK); - writel(cfg, cp->regs + REG_MIF_CFG); -} - -/* Must be invoked under cp->lock */ -static void cas_begin_auto_negotiation(struct cas *cp, - const struct ethtool_link_ksettings *ep) -{ - u16 ctl; -#if 1 - int lcntl; - int changed = 0; - int oldstate = cp->lstate; - int link_was_not_down = !(oldstate == link_down); -#endif - /* Setup link parameters */ - if (!ep) - goto start_aneg; - lcntl = cp->link_cntl; - if (ep->base.autoneg == AUTONEG_ENABLE) { - cp->link_cntl = BMCR_ANENABLE; - } else { - u32 speed = ep->base.speed; - cp->link_cntl = 0; - if (speed == SPEED_100) - cp->link_cntl |= BMCR_SPEED100; - else if (speed == SPEED_1000) - cp->link_cntl |= CAS_BMCR_SPEED1000; - if (ep->base.duplex == DUPLEX_FULL) - cp->link_cntl |= BMCR_FULLDPLX; - } -#if 1 - changed = (lcntl != cp->link_cntl); -#endif -start_aneg: - if (cp->lstate == link_up) { - netdev_info(cp->dev, "PCS link down\n"); - } else { - if (changed) { - netdev_info(cp->dev, "link configuration changed\n"); - } - } - cp->lstate = link_down; - cp->link_transition = LINK_TRANSITION_LINK_DOWN; - if (!cp->hw_running) - return; -#if 1 - /* - * WTZ: If the old state was link_up, we turn off the carrier - * to replicate everything we do elsewhere on a link-down - * event when we were already in a link-up state.. - */ - if (oldstate == link_up) - netif_carrier_off(cp->dev); - if (changed && link_was_not_down) { - /* - * WTZ: This branch will simply schedule a full reset after - * we explicitly changed link modes in an ioctl. See if this - * fixes the link-problems we were having for forced mode. - */ - atomic_inc(&cp->reset_task_pending); - atomic_inc(&cp->reset_task_pending_all); - schedule_work(&cp->reset_task); - cp->timer_ticks = 0; - mod_timer(&cp->link_timer, jiffies + CAS_LINK_TIMEOUT); - return; - } -#endif - if (cp->phy_type & CAS_PHY_SERDES) { - u32 val = readl(cp->regs + REG_PCS_MII_CTRL); - - if (cp->link_cntl & BMCR_ANENABLE) { - val |= (PCS_MII_RESTART_AUTONEG | PCS_MII_AUTONEG_EN); - cp->lstate = link_aneg; - } else { - if (cp->link_cntl & BMCR_FULLDPLX) - val |= PCS_MII_CTRL_DUPLEX; - val &= ~PCS_MII_AUTONEG_EN; - cp->lstate = link_force_ok; - } - cp->link_transition = LINK_TRANSITION_LINK_CONFIG; - writel(val, cp->regs + REG_PCS_MII_CTRL); - - } else { - cas_mif_poll(cp, 0); - ctl = cas_phy_read(cp, MII_BMCR); - ctl &= ~(BMCR_FULLDPLX | BMCR_SPEED100 | - CAS_BMCR_SPEED1000 | BMCR_ANENABLE); - ctl |= cp->link_cntl; - if (ctl & BMCR_ANENABLE) { - ctl |= BMCR_ANRESTART; - cp->lstate = link_aneg; - } else { - cp->lstate = link_force_ok; - } - cp->link_transition = LINK_TRANSITION_LINK_CONFIG; - cas_phy_write(cp, MII_BMCR, ctl); - cas_mif_poll(cp, 1); - } - - cp->timer_ticks = 0; - mod_timer(&cp->link_timer, jiffies + CAS_LINK_TIMEOUT); -} - -/* Must be invoked under cp->lock. */ -static int cas_reset_mii_phy(struct cas *cp) -{ - int limit = STOP_TRIES_PHY; - u16 val; - - cas_phy_write(cp, MII_BMCR, BMCR_RESET); - udelay(100); - while (--limit) { - val = cas_phy_read(cp, MII_BMCR); - if ((val & BMCR_RESET) == 0) - break; - udelay(10); - } - return limit <= 0; -} - -static void cas_saturn_firmware_init(struct cas *cp) -{ - const struct firmware *fw; - const char fw_name[] = "sun/cassini.bin"; - int err; - - if (PHY_NS_DP83065 != cp->phy_id) - return; - - err = request_firmware(&fw, fw_name, &cp->pdev->dev); - if (err) { - pr_err("Failed to load firmware \"%s\"\n", - fw_name); - return; - } - if (fw->size < 2) { - pr_err("bogus length %zu in \"%s\"\n", - fw->size, fw_name); - goto out; - } - cp->fw_load_addr= fw->data[1] << 8 | fw->data[0]; - cp->fw_size = fw->size - 2; - cp->fw_data = vmalloc(cp->fw_size); - if (!cp->fw_data) - goto out; - memcpy(cp->fw_data, &fw->data[2], cp->fw_size); -out: - release_firmware(fw); -} - -static void cas_saturn_firmware_load(struct cas *cp) -{ - int i; - - if (!cp->fw_data) - return; - - cas_phy_powerdown(cp); - - /* expanded memory access mode */ - cas_phy_write(cp, DP83065_MII_MEM, 0x0); - - /* pointer configuration for new firmware */ - cas_phy_write(cp, DP83065_MII_REGE, 0x8ff9); - cas_phy_write(cp, DP83065_MII_REGD, 0xbd); - cas_phy_write(cp, DP83065_MII_REGE, 0x8ffa); - cas_phy_write(cp, DP83065_MII_REGD, 0x82); - cas_phy_write(cp, DP83065_MII_REGE, 0x8ffb); - cas_phy_write(cp, DP83065_MII_REGD, 0x0); - cas_phy_write(cp, DP83065_MII_REGE, 0x8ffc); - cas_phy_write(cp, DP83065_MII_REGD, 0x39); - - /* download new firmware */ - cas_phy_write(cp, DP83065_MII_MEM, 0x1); - cas_phy_write(cp, DP83065_MII_REGE, cp->fw_load_addr); - for (i = 0; i < cp->fw_size; i++) - cas_phy_write(cp, DP83065_MII_REGD, cp->fw_data[i]); - - /* enable firmware */ - cas_phy_write(cp, DP83065_MII_REGE, 0x8ff8); - cas_phy_write(cp, DP83065_MII_REGD, 0x1); -} - - -/* phy initialization */ -static void cas_phy_init(struct cas *cp) -{ - u16 val; - - /* if we're in MII/GMII mode, set up phy */ - if (CAS_PHY_MII(cp->phy_type)) { - writel(PCS_DATAPATH_MODE_MII, - cp->regs + REG_PCS_DATAPATH_MODE); - - cas_mif_poll(cp, 0); - cas_reset_mii_phy(cp); /* take out of isolate mode */ - - if (PHY_LUCENT_B0 == cp->phy_id) { - /* workaround link up/down issue with lucent */ - cas_phy_write(cp, LUCENT_MII_REG, 0x8000); - cas_phy_write(cp, MII_BMCR, 0x00f1); - cas_phy_write(cp, LUCENT_MII_REG, 0x0); - - } else if (PHY_BROADCOM_B0 == (cp->phy_id & 0xFFFFFFFC)) { - /* workarounds for broadcom phy */ - cas_phy_write(cp, BROADCOM_MII_REG8, 0x0C20); - cas_phy_write(cp, BROADCOM_MII_REG7, 0x0012); - cas_phy_write(cp, BROADCOM_MII_REG5, 0x1804); - cas_phy_write(cp, BROADCOM_MII_REG7, 0x0013); - cas_phy_write(cp, BROADCOM_MII_REG5, 0x1204); - cas_phy_write(cp, BROADCOM_MII_REG7, 0x8006); - cas_phy_write(cp, BROADCOM_MII_REG5, 0x0132); - cas_phy_write(cp, BROADCOM_MII_REG7, 0x8006); - cas_phy_write(cp, BROADCOM_MII_REG5, 0x0232); - cas_phy_write(cp, BROADCOM_MII_REG7, 0x201F); - cas_phy_write(cp, BROADCOM_MII_REG5, 0x0A20); - - } else if (PHY_BROADCOM_5411 == cp->phy_id) { - val = cas_phy_read(cp, BROADCOM_MII_REG4); - val = cas_phy_read(cp, BROADCOM_MII_REG4); - if (val & 0x0080) { - /* link workaround */ - cas_phy_write(cp, BROADCOM_MII_REG4, - val & ~0x0080); - } - - } else if (cp->cas_flags & CAS_FLAG_SATURN) { - writel((cp->phy_type & CAS_PHY_MII_MDIO0) ? - SATURN_PCFG_FSI : 0x0, - cp->regs + REG_SATURN_PCFG); - - /* load firmware to address 10Mbps auto-negotiation - * issue. NOTE: this will need to be changed if the - * default firmware gets fixed. - */ - if (PHY_NS_DP83065 == cp->phy_id) { - cas_saturn_firmware_load(cp); - } - cas_phy_powerup(cp); - } - - /* advertise capabilities */ - val = cas_phy_read(cp, MII_BMCR); - val &= ~BMCR_ANENABLE; - cas_phy_write(cp, MII_BMCR, val); - udelay(10); - - cas_phy_write(cp, MII_ADVERTISE, - cas_phy_read(cp, MII_ADVERTISE) | - (ADVERTISE_10HALF | ADVERTISE_10FULL | - ADVERTISE_100HALF | ADVERTISE_100FULL | - CAS_ADVERTISE_PAUSE | - CAS_ADVERTISE_ASYM_PAUSE)); - - if (cp->cas_flags & CAS_FLAG_1000MB_CAP) { - /* make sure that we don't advertise half - * duplex to avoid a chip issue - */ - val = cas_phy_read(cp, CAS_MII_1000_CTRL); - val &= ~CAS_ADVERTISE_1000HALF; - val |= CAS_ADVERTISE_1000FULL; - cas_phy_write(cp, CAS_MII_1000_CTRL, val); - } - - } else { - /* reset pcs for serdes */ - u32 val; - int limit; - - writel(PCS_DATAPATH_MODE_SERDES, - cp->regs + REG_PCS_DATAPATH_MODE); - - /* enable serdes pins on saturn */ - if (cp->cas_flags & CAS_FLAG_SATURN) - writel(0, cp->regs + REG_SATURN_PCFG); - - /* Reset PCS unit. */ - val = readl(cp->regs + REG_PCS_MII_CTRL); - val |= PCS_MII_RESET; - writel(val, cp->regs + REG_PCS_MII_CTRL); - - limit = STOP_TRIES; - while (--limit > 0) { - udelay(10); - if ((readl(cp->regs + REG_PCS_MII_CTRL) & - PCS_MII_RESET) == 0) - break; - } - if (limit <= 0) - netdev_warn(cp->dev, "PCS reset bit would not clear [%08x]\n", - readl(cp->regs + REG_PCS_STATE_MACHINE)); - - /* Make sure PCS is disabled while changing advertisement - * configuration. - */ - writel(0x0, cp->regs + REG_PCS_CFG); - - /* Advertise all capabilities except half-duplex. */ - val = readl(cp->regs + REG_PCS_MII_ADVERT); - val &= ~PCS_MII_ADVERT_HD; - val |= (PCS_MII_ADVERT_FD | PCS_MII_ADVERT_SYM_PAUSE | - PCS_MII_ADVERT_ASYM_PAUSE); - writel(val, cp->regs + REG_PCS_MII_ADVERT); - - /* enable PCS */ - writel(PCS_CFG_EN, cp->regs + REG_PCS_CFG); - - /* pcs workaround: enable sync detect */ - writel(PCS_SERDES_CTRL_SYNCD_EN, - cp->regs + REG_PCS_SERDES_CTRL); - } -} - - -static int cas_pcs_link_check(struct cas *cp) -{ - u32 stat, state_machine; - int retval = 0; - - /* The link status bit latches on zero, so you must - * read it twice in such a case to see a transition - * to the link being up. - */ - stat = readl(cp->regs + REG_PCS_MII_STATUS); - if ((stat & PCS_MII_STATUS_LINK_STATUS) == 0) - stat = readl(cp->regs + REG_PCS_MII_STATUS); - - /* The remote-fault indication is only valid - * when autoneg has completed. - */ - if ((stat & (PCS_MII_STATUS_AUTONEG_COMP | - PCS_MII_STATUS_REMOTE_FAULT)) == - (PCS_MII_STATUS_AUTONEG_COMP | PCS_MII_STATUS_REMOTE_FAULT)) - netif_info(cp, link, cp->dev, "PCS RemoteFault\n"); - - /* work around link detection issue by querying the PCS state - * machine directly. - */ - state_machine = readl(cp->regs + REG_PCS_STATE_MACHINE); - if ((state_machine & PCS_SM_LINK_STATE_MASK) != SM_LINK_STATE_UP) { - stat &= ~PCS_MII_STATUS_LINK_STATUS; - } else if (state_machine & PCS_SM_WORD_SYNC_STATE_MASK) { - stat |= PCS_MII_STATUS_LINK_STATUS; - } - - if (stat & PCS_MII_STATUS_LINK_STATUS) { - if (cp->lstate != link_up) { - if (cp->opened) { - cp->lstate = link_up; - cp->link_transition = LINK_TRANSITION_LINK_UP; - - cas_set_link_modes(cp); - netif_carrier_on(cp->dev); - } - } - } else if (cp->lstate == link_up) { - cp->lstate = link_down; - if (link_transition_timeout != 0 && - cp->link_transition != LINK_TRANSITION_REQUESTED_RESET && - !cp->link_transition_jiffies_valid) { - /* - * force a reset, as a workaround for the - * link-failure problem. May want to move this to a - * point a bit earlier in the sequence. If we had - * generated a reset a short time ago, we'll wait for - * the link timer to check the status until a - * timer expires (link_transistion_jiffies_valid is - * true when the timer is running.) Instead of using - * a system timer, we just do a check whenever the - * link timer is running - this clears the flag after - * a suitable delay. - */ - retval = 1; - cp->link_transition = LINK_TRANSITION_REQUESTED_RESET; - cp->link_transition_jiffies = jiffies; - cp->link_transition_jiffies_valid = 1; - } else { - cp->link_transition = LINK_TRANSITION_ON_FAILURE; - } - netif_carrier_off(cp->dev); - if (cp->opened) - netif_info(cp, link, cp->dev, "PCS link down\n"); - - /* Cassini only: if you force a mode, there can be - * sync problems on link down. to fix that, the following - * things need to be checked: - * 1) read serialink state register - * 2) read pcs status register to verify link down. - * 3) if link down and serial link == 0x03, then you need - * to global reset the chip. - */ - if ((cp->cas_flags & CAS_FLAG_REG_PLUS) == 0) { - /* should check to see if we're in a forced mode */ - stat = readl(cp->regs + REG_PCS_SERDES_STATE); - if (stat == 0x03) - return 1; - } - } else if (cp->lstate == link_down) { - if (link_transition_timeout != 0 && - cp->link_transition != LINK_TRANSITION_REQUESTED_RESET && - !cp->link_transition_jiffies_valid) { - /* force a reset, as a workaround for the - * link-failure problem. May want to move - * this to a point a bit earlier in the - * sequence. - */ - retval = 1; - cp->link_transition = LINK_TRANSITION_REQUESTED_RESET; - cp->link_transition_jiffies = jiffies; - cp->link_transition_jiffies_valid = 1; - } else { - cp->link_transition = LINK_TRANSITION_STILL_FAILED; - } - } - - return retval; -} - -static int cas_pcs_interrupt(struct net_device *dev, - struct cas *cp, u32 status) -{ - u32 stat = readl(cp->regs + REG_PCS_INTR_STATUS); - - if ((stat & PCS_INTR_STATUS_LINK_CHANGE) == 0) - return 0; - return cas_pcs_link_check(cp); -} - -static int cas_txmac_interrupt(struct net_device *dev, - struct cas *cp, u32 status) -{ - u32 txmac_stat = readl(cp->regs + REG_MAC_TX_STATUS); - - if (!txmac_stat) - return 0; - - netif_printk(cp, intr, KERN_DEBUG, cp->dev, - "txmac interrupt, txmac_stat: 0x%x\n", txmac_stat); - - /* Defer timer expiration is quite normal, - * don't even log the event. - */ - if ((txmac_stat & MAC_TX_DEFER_TIMER) && - !(txmac_stat & ~MAC_TX_DEFER_TIMER)) - return 0; - - spin_lock(&cp->stat_lock[0]); - if (txmac_stat & MAC_TX_UNDERRUN) { - netdev_err(dev, "TX MAC xmit underrun\n"); - cp->net_stats[0].tx_fifo_errors++; - } - - if (txmac_stat & MAC_TX_MAX_PACKET_ERR) { - netdev_err(dev, "TX MAC max packet size error\n"); - cp->net_stats[0].tx_errors++; - } - - /* The rest are all cases of one of the 16-bit TX - * counters expiring. - */ - if (txmac_stat & MAC_TX_COLL_NORMAL) - cp->net_stats[0].collisions += 0x10000; - - if (txmac_stat & MAC_TX_COLL_EXCESS) { - cp->net_stats[0].tx_aborted_errors += 0x10000; - cp->net_stats[0].collisions += 0x10000; - } - - if (txmac_stat & MAC_TX_COLL_LATE) { - cp->net_stats[0].tx_aborted_errors += 0x10000; - cp->net_stats[0].collisions += 0x10000; - } - spin_unlock(&cp->stat_lock[0]); - - /* We do not keep track of MAC_TX_COLL_FIRST and - * MAC_TX_PEAK_ATTEMPTS events. - */ - return 0; -} - -static void cas_load_firmware(struct cas *cp, cas_hp_inst_t *firmware) -{ - cas_hp_inst_t *inst; - u32 val; - int i; - - i = 0; - while ((inst = firmware) && inst->note) { - writel(i, cp->regs + REG_HP_INSTR_RAM_ADDR); - - val = CAS_BASE(HP_INSTR_RAM_HI_VAL, inst->val); - val |= CAS_BASE(HP_INSTR_RAM_HI_MASK, inst->mask); - writel(val, cp->regs + REG_HP_INSTR_RAM_DATA_HI); - - val = CAS_BASE(HP_INSTR_RAM_MID_OUTARG, inst->outarg >> 10); - val |= CAS_BASE(HP_INSTR_RAM_MID_OUTOP, inst->outop); - val |= CAS_BASE(HP_INSTR_RAM_MID_FNEXT, inst->fnext); - val |= CAS_BASE(HP_INSTR_RAM_MID_FOFF, inst->foff); - val |= CAS_BASE(HP_INSTR_RAM_MID_SNEXT, inst->snext); - val |= CAS_BASE(HP_INSTR_RAM_MID_SOFF, inst->soff); - val |= CAS_BASE(HP_INSTR_RAM_MID_OP, inst->op); - writel(val, cp->regs + REG_HP_INSTR_RAM_DATA_MID); - - val = CAS_BASE(HP_INSTR_RAM_LOW_OUTMASK, inst->outmask); - val |= CAS_BASE(HP_INSTR_RAM_LOW_OUTSHIFT, inst->outshift); - val |= CAS_BASE(HP_INSTR_RAM_LOW_OUTEN, inst->outenab); - val |= CAS_BASE(HP_INSTR_RAM_LOW_OUTARG, inst->outarg); - writel(val, cp->regs + REG_HP_INSTR_RAM_DATA_LOW); - ++firmware; - ++i; - } -} - -static void cas_init_rx_dma(struct cas *cp) -{ - u64 desc_dma = cp->block_dvma; - u32 val; - int i, size; - - /* rx free descriptors */ - val = CAS_BASE(RX_CFG_SWIVEL, RX_SWIVEL_OFF_VAL); - val |= CAS_BASE(RX_CFG_DESC_RING, RX_DESC_RINGN_INDEX(0)); - val |= CAS_BASE(RX_CFG_COMP_RING, RX_COMP_RINGN_INDEX(0)); - if ((N_RX_DESC_RINGS > 1) && - (cp->cas_flags & CAS_FLAG_REG_PLUS)) /* do desc 2 */ - val |= CAS_BASE(RX_CFG_DESC_RING1, RX_DESC_RINGN_INDEX(1)); - writel(val, cp->regs + REG_RX_CFG); - - val = (unsigned long) cp->init_rxds[0] - - (unsigned long) cp->init_block; - writel((desc_dma + val) >> 32, cp->regs + REG_RX_DB_HI); - writel((desc_dma + val) & 0xffffffff, cp->regs + REG_RX_DB_LOW); - writel(RX_DESC_RINGN_SIZE(0) - 4, cp->regs + REG_RX_KICK); - - if (cp->cas_flags & CAS_FLAG_REG_PLUS) { - /* rx desc 2 is for IPSEC packets. however, - * we don't it that for that purpose. - */ - val = (unsigned long) cp->init_rxds[1] - - (unsigned long) cp->init_block; - writel((desc_dma + val) >> 32, cp->regs + REG_PLUS_RX_DB1_HI); - writel((desc_dma + val) & 0xffffffff, cp->regs + - REG_PLUS_RX_DB1_LOW); - writel(RX_DESC_RINGN_SIZE(1) - 4, cp->regs + - REG_PLUS_RX_KICK1); - } - - /* rx completion registers */ - val = (unsigned long) cp->init_rxcs[0] - - (unsigned long) cp->init_block; - writel((desc_dma + val) >> 32, cp->regs + REG_RX_CB_HI); - writel((desc_dma + val) & 0xffffffff, cp->regs + REG_RX_CB_LOW); - - if (cp->cas_flags & CAS_FLAG_REG_PLUS) { - /* rx comp 2-4 */ - for (i = 1; i < MAX_RX_COMP_RINGS; i++) { - val = (unsigned long) cp->init_rxcs[i] - - (unsigned long) cp->init_block; - writel((desc_dma + val) >> 32, cp->regs + - REG_PLUS_RX_CBN_HI(i)); - writel((desc_dma + val) & 0xffffffff, cp->regs + - REG_PLUS_RX_CBN_LOW(i)); - } - } - - /* read selective clear regs to prevent spurious interrupts - * on reset because complete == kick. - * selective clear set up to prevent interrupts on resets - */ - readl(cp->regs + REG_INTR_STATUS_ALIAS); - writel(INTR_RX_DONE | INTR_RX_BUF_UNAVAIL, cp->regs + REG_ALIAS_CLEAR); - - /* set up pause thresholds */ - val = CAS_BASE(RX_PAUSE_THRESH_OFF, - cp->rx_pause_off / RX_PAUSE_THRESH_QUANTUM); - val |= CAS_BASE(RX_PAUSE_THRESH_ON, - cp->rx_pause_on / RX_PAUSE_THRESH_QUANTUM); - writel(val, cp->regs + REG_RX_PAUSE_THRESH); - - /* zero out dma reassembly buffers */ - for (i = 0; i < 64; i++) { - writel(i, cp->regs + REG_RX_TABLE_ADDR); - writel(0x0, cp->regs + REG_RX_TABLE_DATA_LOW); - writel(0x0, cp->regs + REG_RX_TABLE_DATA_MID); - writel(0x0, cp->regs + REG_RX_TABLE_DATA_HI); - } - - /* make sure address register is 0 for normal operation */ - writel(0x0, cp->regs + REG_RX_CTRL_FIFO_ADDR); - writel(0x0, cp->regs + REG_RX_IPP_FIFO_ADDR); - - /* interrupt mitigation */ -#ifdef USE_RX_BLANK - val = CAS_BASE(RX_BLANK_INTR_TIME, RX_BLANK_INTR_TIME_VAL); - val |= CAS_BASE(RX_BLANK_INTR_PKT, RX_BLANK_INTR_PKT_VAL); - writel(val, cp->regs + REG_RX_BLANK); -#else - writel(0x0, cp->regs + REG_RX_BLANK); -#endif - - /* interrupt generation as a function of low water marks for - * free desc and completion entries. these are used to trigger - * housekeeping for rx descs. we don't use the free interrupt - * as it's not very useful - */ - /* val = CAS_BASE(RX_AE_THRESH_FREE, RX_AE_FREEN_VAL(0)); */ - val = CAS_BASE(RX_AE_THRESH_COMP, RX_AE_COMP_VAL); - writel(val, cp->regs + REG_RX_AE_THRESH); - if (cp->cas_flags & CAS_FLAG_REG_PLUS) { - val = CAS_BASE(RX_AE1_THRESH_FREE, RX_AE_FREEN_VAL(1)); - writel(val, cp->regs + REG_PLUS_RX_AE1_THRESH); - } - - /* Random early detect registers. useful for congestion avoidance. - * this should be tunable. - */ - writel(0x0, cp->regs + REG_RX_RED); - - /* receive page sizes. default == 2K (0x800) */ - val = 0; - if (cp->page_size == 0x1000) - val = 0x1; - else if (cp->page_size == 0x2000) - val = 0x2; - else if (cp->page_size == 0x4000) - val = 0x3; - - /* round mtu + offset. constrain to page size. */ - size = cp->dev->mtu + 64; - if (size > cp->page_size) - size = cp->page_size; - - if (size <= 0x400) - i = 0x0; - else if (size <= 0x800) - i = 0x1; - else if (size <= 0x1000) - i = 0x2; - else - i = 0x3; - - cp->mtu_stride = 1 << (i + 10); - val = CAS_BASE(RX_PAGE_SIZE, val); - val |= CAS_BASE(RX_PAGE_SIZE_MTU_STRIDE, i); - val |= CAS_BASE(RX_PAGE_SIZE_MTU_COUNT, cp->page_size >> (i + 10)); - val |= CAS_BASE(RX_PAGE_SIZE_MTU_OFF, 0x1); - writel(val, cp->regs + REG_RX_PAGE_SIZE); - - /* enable the header parser if desired */ - if (&CAS_HP_FIRMWARE[0] == &cas_prog_null[0]) - return; - - val = CAS_BASE(HP_CFG_NUM_CPU, CAS_NCPUS > 63 ? 0 : CAS_NCPUS); - val |= HP_CFG_PARSE_EN | HP_CFG_SYN_INC_MASK; - val |= CAS_BASE(HP_CFG_TCP_THRESH, HP_TCP_THRESH_VAL); - writel(val, cp->regs + REG_HP_CFG); -} - -static inline void cas_rxc_init(struct cas_rx_comp *rxc) -{ - memset(rxc, 0, sizeof(*rxc)); - rxc->word4 = cpu_to_le64(RX_COMP4_ZERO); -} - -/* NOTE: we use the ENC RX DESC ring for spares. the rx_page[0,1] - * flipping is protected by the fact that the chip will not - * hand back the same page index while it's being processed. - */ -static inline cas_page_t *cas_page_spare(struct cas *cp, const int index) -{ - cas_page_t *page = cp->rx_pages[1][index]; - cas_page_t *new; - - if (page_count(page->buffer) == 1) - return page; - - new = cas_page_dequeue(cp); - if (new) { - spin_lock(&cp->rx_inuse_lock); - list_add(&page->list, &cp->rx_inuse_list); - spin_unlock(&cp->rx_inuse_lock); - } - return new; -} - -/* this needs to be changed if we actually use the ENC RX DESC ring */ -static cas_page_t *cas_page_swap(struct cas *cp, const int ring, - const int index) -{ - cas_page_t **page0 = cp->rx_pages[0]; - cas_page_t **page1 = cp->rx_pages[1]; - - /* swap if buffer is in use */ - if (page_count(page0[index]->buffer) > 1) { - cas_page_t *new = cas_page_spare(cp, index); - if (new) { - page1[index] = page0[index]; - page0[index] = new; - } - } - RX_USED_SET(page0[index], 0); - return page0[index]; -} - -static void cas_clean_rxds(struct cas *cp) -{ - /* only clean ring 0 as ring 1 is used for spare buffers */ - struct cas_rx_desc *rxd = cp->init_rxds[0]; - int i, size; - - /* release all rx flows */ - for (i = 0; i < N_RX_FLOWS; i++) { - struct sk_buff *skb; - while ((skb = __skb_dequeue(&cp->rx_flows[i]))) { - cas_skb_release(skb); - } - } - - /* initialize descriptors */ - size = RX_DESC_RINGN_SIZE(0); - for (i = 0; i < size; i++) { - cas_page_t *page = cas_page_swap(cp, 0, i); - rxd[i].buffer = cpu_to_le64(page->dma_addr); - rxd[i].index = cpu_to_le64(CAS_BASE(RX_INDEX_NUM, i) | - CAS_BASE(RX_INDEX_RING, 0)); - } - - cp->rx_old[0] = RX_DESC_RINGN_SIZE(0) - 4; - cp->rx_last[0] = 0; - cp->cas_flags &= ~CAS_FLAG_RXD_POST(0); -} - -static void cas_clean_rxcs(struct cas *cp) -{ - int i, j; - - /* take ownership of rx comp descriptors */ - memset(cp->rx_cur, 0, sizeof(*cp->rx_cur)*N_RX_COMP_RINGS); - memset(cp->rx_new, 0, sizeof(*cp->rx_new)*N_RX_COMP_RINGS); - for (i = 0; i < N_RX_COMP_RINGS; i++) { - struct cas_rx_comp *rxc = cp->init_rxcs[i]; - for (j = 0; j < RX_COMP_RINGN_SIZE(i); j++) { - cas_rxc_init(rxc + j); - } - } -} - -#if 0 -/* When we get a RX fifo overflow, the RX unit is probably hung - * so we do the following. - * - * If any part of the reset goes wrong, we return 1 and that causes the - * whole chip to be reset. - */ -static int cas_rxmac_reset(struct cas *cp) -{ - struct net_device *dev = cp->dev; - int limit; - u32 val; - - /* First, reset MAC RX. */ - writel(cp->mac_rx_cfg & ~MAC_RX_CFG_EN, cp->regs + REG_MAC_RX_CFG); - for (limit = 0; limit < STOP_TRIES; limit++) { - if (!(readl(cp->regs + REG_MAC_RX_CFG) & MAC_RX_CFG_EN)) - break; - udelay(10); - } - if (limit == STOP_TRIES) { - netdev_err(dev, "RX MAC will not disable, resetting whole chip\n"); - return 1; - } - - /* Second, disable RX DMA. */ - writel(0, cp->regs + REG_RX_CFG); - for (limit = 0; limit < STOP_TRIES; limit++) { - if (!(readl(cp->regs + REG_RX_CFG) & RX_CFG_DMA_EN)) - break; - udelay(10); - } - if (limit == STOP_TRIES) { - netdev_err(dev, "RX DMA will not disable, resetting whole chip\n"); - return 1; - } - - mdelay(5); - - /* Execute RX reset command. */ - writel(SW_RESET_RX, cp->regs + REG_SW_RESET); - for (limit = 0; limit < STOP_TRIES; limit++) { - if (!(readl(cp->regs + REG_SW_RESET) & SW_RESET_RX)) - break; - udelay(10); - } - if (limit == STOP_TRIES) { - netdev_err(dev, "RX reset command will not execute, resetting whole chip\n"); - return 1; - } - - /* reset driver rx state */ - cas_clean_rxds(cp); - cas_clean_rxcs(cp); - - /* Now, reprogram the rest of RX unit. */ - cas_init_rx_dma(cp); - - /* re-enable */ - val = readl(cp->regs + REG_RX_CFG); - writel(val | RX_CFG_DMA_EN, cp->regs + REG_RX_CFG); - writel(MAC_RX_FRAME_RECV, cp->regs + REG_MAC_RX_MASK); - val = readl(cp->regs + REG_MAC_RX_CFG); - writel(val | MAC_RX_CFG_EN, cp->regs + REG_MAC_RX_CFG); - return 0; -} -#endif - -static int cas_rxmac_interrupt(struct net_device *dev, struct cas *cp, - u32 status) -{ - u32 stat = readl(cp->regs + REG_MAC_RX_STATUS); - - if (!stat) - return 0; - - netif_dbg(cp, intr, cp->dev, "rxmac interrupt, stat: 0x%x\n", stat); - - /* these are all rollovers */ - spin_lock(&cp->stat_lock[0]); - if (stat & MAC_RX_ALIGN_ERR) - cp->net_stats[0].rx_frame_errors += 0x10000; - - if (stat & MAC_RX_CRC_ERR) - cp->net_stats[0].rx_crc_errors += 0x10000; - - if (stat & MAC_RX_LEN_ERR) - cp->net_stats[0].rx_length_errors += 0x10000; - - if (stat & MAC_RX_OVERFLOW) { - cp->net_stats[0].rx_over_errors++; - cp->net_stats[0].rx_fifo_errors++; - } - - /* We do not track MAC_RX_FRAME_COUNT and MAC_RX_VIOL_ERR - * events. - */ - spin_unlock(&cp->stat_lock[0]); - return 0; -} - -static int cas_mac_interrupt(struct net_device *dev, struct cas *cp, - u32 status) -{ - u32 stat = readl(cp->regs + REG_MAC_CTRL_STATUS); - - if (!stat) - return 0; - - netif_printk(cp, intr, KERN_DEBUG, cp->dev, - "mac interrupt, stat: 0x%x\n", stat); - - /* This interrupt is just for pause frame and pause - * tracking. It is useful for diagnostics and debug - * but probably by default we will mask these events. - */ - if (stat & MAC_CTRL_PAUSE_STATE) - cp->pause_entered++; - - if (stat & MAC_CTRL_PAUSE_RECEIVED) - cp->pause_last_time_recvd = (stat >> 16); - - return 0; -} - - -/* Must be invoked under cp->lock. */ -static inline int cas_mdio_link_not_up(struct cas *cp) -{ - u16 val; - - switch (cp->lstate) { - case link_force_ret: - netif_info(cp, link, cp->dev, "Autoneg failed again, keeping forced mode\n"); - cas_phy_write(cp, MII_BMCR, cp->link_fcntl); - cp->timer_ticks = 5; - cp->lstate = link_force_ok; - cp->link_transition = LINK_TRANSITION_LINK_CONFIG; - break; - - case link_aneg: - val = cas_phy_read(cp, MII_BMCR); - - /* Try forced modes. we try things in the following order: - * 1000 full -> 100 full/half -> 10 half - */ - val &= ~(BMCR_ANRESTART | BMCR_ANENABLE); - val |= BMCR_FULLDPLX; - val |= (cp->cas_flags & CAS_FLAG_1000MB_CAP) ? - CAS_BMCR_SPEED1000 : BMCR_SPEED100; - cas_phy_write(cp, MII_BMCR, val); - cp->timer_ticks = 5; - cp->lstate = link_force_try; - cp->link_transition = LINK_TRANSITION_LINK_CONFIG; - break; - - case link_force_try: - /* Downgrade from 1000 to 100 to 10 Mbps if necessary. */ - val = cas_phy_read(cp, MII_BMCR); - cp->timer_ticks = 5; - if (val & CAS_BMCR_SPEED1000) { /* gigabit */ - val &= ~CAS_BMCR_SPEED1000; - val |= (BMCR_SPEED100 | BMCR_FULLDPLX); - cas_phy_write(cp, MII_BMCR, val); - break; - } - - if (val & BMCR_SPEED100) { - if (val & BMCR_FULLDPLX) /* fd failed */ - val &= ~BMCR_FULLDPLX; - else { /* 100Mbps failed */ - val &= ~BMCR_SPEED100; - } - cas_phy_write(cp, MII_BMCR, val); - break; - } - break; - default: - break; - } - return 0; -} - - -/* must be invoked with cp->lock held */ -static int cas_mii_link_check(struct cas *cp, const u16 bmsr) -{ - int restart; - - if (bmsr & BMSR_LSTATUS) { - /* Ok, here we got a link. If we had it due to a forced - * fallback, and we were configured for autoneg, we - * retry a short autoneg pass. If you know your hub is - * broken, use ethtool ;) - */ - if ((cp->lstate == link_force_try) && - (cp->link_cntl & BMCR_ANENABLE)) { - cp->lstate = link_force_ret; - cp->link_transition = LINK_TRANSITION_LINK_CONFIG; - cas_mif_poll(cp, 0); - cp->link_fcntl = cas_phy_read(cp, MII_BMCR); - cp->timer_ticks = 5; - if (cp->opened) - netif_info(cp, link, cp->dev, - "Got link after fallback, retrying autoneg once...\n"); - cas_phy_write(cp, MII_BMCR, - cp->link_fcntl | BMCR_ANENABLE | - BMCR_ANRESTART); - cas_mif_poll(cp, 1); - - } else if (cp->lstate != link_up) { - cp->lstate = link_up; - cp->link_transition = LINK_TRANSITION_LINK_UP; - - if (cp->opened) { - cas_set_link_modes(cp); - netif_carrier_on(cp->dev); - } - } - return 0; - } - - /* link not up. if the link was previously up, we restart the - * whole process - */ - restart = 0; - if (cp->lstate == link_up) { - cp->lstate = link_down; - cp->link_transition = LINK_TRANSITION_LINK_DOWN; - - netif_carrier_off(cp->dev); - if (cp->opened) - netif_info(cp, link, cp->dev, "Link down\n"); - restart = 1; - - } else if (++cp->timer_ticks > 10) - cas_mdio_link_not_up(cp); - - return restart; -} - -static int cas_mif_interrupt(struct net_device *dev, struct cas *cp, - u32 status) -{ - u32 stat = readl(cp->regs + REG_MIF_STATUS); - u16 bmsr; - - /* check for a link change */ - if (CAS_VAL(MIF_STATUS_POLL_STATUS, stat) == 0) - return 0; - - bmsr = CAS_VAL(MIF_STATUS_POLL_DATA, stat); - return cas_mii_link_check(cp, bmsr); -} - -static int cas_pci_interrupt(struct net_device *dev, struct cas *cp, - u32 status) -{ - u32 stat = readl(cp->regs + REG_PCI_ERR_STATUS); - - if (!stat) - return 0; - - netdev_err(dev, "PCI error [%04x:%04x]", - stat, readl(cp->regs + REG_BIM_DIAG)); - - /* cassini+ has this reserved */ - if ((stat & PCI_ERR_BADACK) && - ((cp->cas_flags & CAS_FLAG_REG_PLUS) == 0)) - pr_cont(" "); - - if (stat & PCI_ERR_DTRTO) - pr_cont(" "); - if (stat & PCI_ERR_OTHER) - pr_cont(" "); - if (stat & PCI_ERR_BIM_DMA_WRITE) - pr_cont(" "); - if (stat & PCI_ERR_BIM_DMA_READ) - pr_cont(" "); - pr_cont("\n"); - - if (stat & PCI_ERR_OTHER) { - int pci_errs; - - /* Interrogate PCI config space for the - * true cause. - */ - pci_errs = pci_status_get_and_clear_errors(cp->pdev); - - netdev_err(dev, "PCI status errors[%04x]\n", pci_errs); - if (pci_errs & PCI_STATUS_PARITY) - netdev_err(dev, "PCI parity error detected\n"); - if (pci_errs & PCI_STATUS_SIG_TARGET_ABORT) - netdev_err(dev, "PCI target abort\n"); - if (pci_errs & PCI_STATUS_REC_TARGET_ABORT) - netdev_err(dev, "PCI master acks target abort\n"); - if (pci_errs & PCI_STATUS_REC_MASTER_ABORT) - netdev_err(dev, "PCI master abort\n"); - if (pci_errs & PCI_STATUS_SIG_SYSTEM_ERROR) - netdev_err(dev, "PCI system error SERR#\n"); - if (pci_errs & PCI_STATUS_DETECTED_PARITY) - netdev_err(dev, "PCI parity error\n"); - } - - /* For all PCI errors, we should reset the chip. */ - return 1; -} - -/* All non-normal interrupt conditions get serviced here. - * Returns non-zero if we should just exit the interrupt - * handler right now (ie. if we reset the card which invalidates - * all of the other original irq status bits). - */ -static int cas_abnormal_irq(struct net_device *dev, struct cas *cp, - u32 status) -{ - if (status & INTR_RX_TAG_ERROR) { - /* corrupt RX tag framing */ - netif_printk(cp, rx_err, KERN_DEBUG, cp->dev, - "corrupt rx tag framing\n"); - spin_lock(&cp->stat_lock[0]); - cp->net_stats[0].rx_errors++; - spin_unlock(&cp->stat_lock[0]); - goto do_reset; - } - - if (status & INTR_RX_LEN_MISMATCH) { - /* length mismatch. */ - netif_printk(cp, rx_err, KERN_DEBUG, cp->dev, - "length mismatch for rx frame\n"); - spin_lock(&cp->stat_lock[0]); - cp->net_stats[0].rx_errors++; - spin_unlock(&cp->stat_lock[0]); - goto do_reset; - } - - if (status & INTR_PCS_STATUS) { - if (cas_pcs_interrupt(dev, cp, status)) - goto do_reset; - } - - if (status & INTR_TX_MAC_STATUS) { - if (cas_txmac_interrupt(dev, cp, status)) - goto do_reset; - } - - if (status & INTR_RX_MAC_STATUS) { - if (cas_rxmac_interrupt(dev, cp, status)) - goto do_reset; - } - - if (status & INTR_MAC_CTRL_STATUS) { - if (cas_mac_interrupt(dev, cp, status)) - goto do_reset; - } - - if (status & INTR_MIF_STATUS) { - if (cas_mif_interrupt(dev, cp, status)) - goto do_reset; - } - - if (status & INTR_PCI_ERROR_STATUS) { - if (cas_pci_interrupt(dev, cp, status)) - goto do_reset; - } - return 0; - -do_reset: -#if 1 - atomic_inc(&cp->reset_task_pending); - atomic_inc(&cp->reset_task_pending_all); - netdev_err(dev, "reset called in cas_abnormal_irq [0x%x]\n", status); - schedule_work(&cp->reset_task); -#else - atomic_set(&cp->reset_task_pending, CAS_RESET_ALL); - netdev_err(dev, "reset called in cas_abnormal_irq\n"); - schedule_work(&cp->reset_task); -#endif - return 1; -} - -/* NOTE: CAS_TABORT returns 1 or 2 so that it can be used when - * determining whether to do a netif_stop/wakeup - */ -#define CAS_TABORT(x) (((x)->cas_flags & CAS_FLAG_TARGET_ABORT) ? 2 : 1) -#define CAS_ROUND_PAGE(x) (((x) + PAGE_SIZE - 1) & PAGE_MASK) -static inline int cas_calc_tabort(struct cas *cp, const unsigned long addr, - const int len) -{ - unsigned long off = addr + len; - - if (CAS_TABORT(cp) == 1) - return 0; - if ((CAS_ROUND_PAGE(off) - off) > TX_TARGET_ABORT_LEN) - return 0; - return TX_TARGET_ABORT_LEN; -} - -static inline void cas_tx_ringN(struct cas *cp, int ring, int limit) -{ - struct cas_tx_desc *txds; - struct sk_buff **skbs; - struct net_device *dev = cp->dev; - int entry, count; - - spin_lock(&cp->tx_lock[ring]); - txds = cp->init_txds[ring]; - skbs = cp->tx_skbs[ring]; - entry = cp->tx_old[ring]; - - count = TX_BUFF_COUNT(ring, entry, limit); - while (entry != limit) { - struct sk_buff *skb = skbs[entry]; - dma_addr_t daddr; - u32 dlen; - int frag; - - if (!skb) { - /* this should never occur */ - entry = TX_DESC_NEXT(ring, entry); - continue; - } - - /* however, we might get only a partial skb release. */ - count -= skb_shinfo(skb)->nr_frags + - + cp->tx_tiny_use[ring][entry].nbufs + 1; - if (count < 0) - break; - - netif_printk(cp, tx_done, KERN_DEBUG, cp->dev, - "tx[%d] done, slot %d\n", ring, entry); - - skbs[entry] = NULL; - cp->tx_tiny_use[ring][entry].nbufs = 0; - - for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { - struct cas_tx_desc *txd = txds + entry; - - daddr = le64_to_cpu(txd->buffer); - dlen = CAS_VAL(TX_DESC_BUFLEN, - le64_to_cpu(txd->control)); - dma_unmap_page(&cp->pdev->dev, daddr, dlen, - DMA_TO_DEVICE); - entry = TX_DESC_NEXT(ring, entry); - - /* tiny buffer may follow */ - if (cp->tx_tiny_use[ring][entry].used) { - cp->tx_tiny_use[ring][entry].used = 0; - entry = TX_DESC_NEXT(ring, entry); - } - } - - spin_lock(&cp->stat_lock[ring]); - cp->net_stats[ring].tx_packets++; - cp->net_stats[ring].tx_bytes += skb->len; - spin_unlock(&cp->stat_lock[ring]); - dev_consume_skb_irq(skb); - } - cp->tx_old[ring] = entry; - - /* this is wrong for multiple tx rings. the net device needs - * multiple queues for this to do the right thing. we wait - * for 2*packets to be available when using tiny buffers - */ - if (netif_queue_stopped(dev) && - (TX_BUFFS_AVAIL(cp, ring) > CAS_TABORT(cp)*(MAX_SKB_FRAGS + 1))) - netif_wake_queue(dev); - spin_unlock(&cp->tx_lock[ring]); -} - -static void cas_tx(struct net_device *dev, struct cas *cp, - u32 status) -{ - int limit, ring; -#ifdef USE_TX_COMPWB - u64 compwb = le64_to_cpu(cp->init_block->tx_compwb); -#endif - netif_printk(cp, intr, KERN_DEBUG, cp->dev, - "tx interrupt, status: 0x%x, %llx\n", - status, (unsigned long long)compwb); - /* process all the rings */ - for (ring = 0; ring < N_TX_RINGS; ring++) { -#ifdef USE_TX_COMPWB - /* use the completion writeback registers */ - limit = (CAS_VAL(TX_COMPWB_MSB, compwb) << 8) | - CAS_VAL(TX_COMPWB_LSB, compwb); - compwb = TX_COMPWB_NEXT(compwb); -#else - limit = readl(cp->regs + REG_TX_COMPN(ring)); -#endif - if (cp->tx_old[ring] != limit) - cas_tx_ringN(cp, ring, limit); - } -} - - -static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, - int entry, const u64 *words, - struct sk_buff **skbref) -{ - int dlen, hlen, len, i, alloclen; - int off, swivel = RX_SWIVEL_OFF_VAL; - struct cas_page *page; - struct sk_buff *skb; - void *crcaddr; - __sum16 csum; - char *p; - - hlen = CAS_VAL(RX_COMP2_HDR_SIZE, words[1]); - dlen = CAS_VAL(RX_COMP1_DATA_SIZE, words[0]); - len = hlen + dlen; - - if (RX_COPY_ALWAYS || (words[2] & RX_COMP3_SMALL_PKT)) - alloclen = len; - else - alloclen = max(hlen, RX_COPY_MIN); - - skb = netdev_alloc_skb(cp->dev, alloclen + swivel + cp->crc_size); - if (skb == NULL) - return -1; - - *skbref = skb; - skb_reserve(skb, swivel); - - p = skb->data; - crcaddr = NULL; - if (hlen) { /* always copy header pages */ - i = CAS_VAL(RX_COMP2_HDR_INDEX, words[1]); - page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; - off = CAS_VAL(RX_COMP2_HDR_OFF, words[1]) * 0x100 + - swivel; - - i = hlen; - if (!dlen) /* attach FCS */ - i += cp->crc_size; - dma_sync_single_for_cpu(&cp->pdev->dev, page->dma_addr + off, - i, DMA_FROM_DEVICE); - memcpy(p, page_address(page->buffer) + off, i); - dma_sync_single_for_device(&cp->pdev->dev, - page->dma_addr + off, i, - DMA_FROM_DEVICE); - RX_USED_ADD(page, 0x100); - p += hlen; - swivel = 0; - } - - - if (alloclen < (hlen + dlen)) { - skb_frag_t *frag = skb_shinfo(skb)->frags; - - /* normal or jumbo packets. we use frags */ - i = CAS_VAL(RX_COMP1_DATA_INDEX, words[0]); - page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; - off = CAS_VAL(RX_COMP1_DATA_OFF, words[0]) + swivel; - - hlen = min(cp->page_size - off, dlen); - if (hlen < 0) { - netif_printk(cp, rx_err, KERN_DEBUG, cp->dev, - "rx page overflow: %d\n", hlen); - dev_kfree_skb_irq(skb); - return -1; - } - i = hlen; - if (i == dlen) /* attach FCS */ - i += cp->crc_size; - dma_sync_single_for_cpu(&cp->pdev->dev, page->dma_addr + off, - i, DMA_FROM_DEVICE); - - /* make sure we always copy a header */ - swivel = 0; - if (p == (char *) skb->data) { /* not split */ - memcpy(p, page_address(page->buffer) + off, - RX_COPY_MIN); - dma_sync_single_for_device(&cp->pdev->dev, - page->dma_addr + off, i, - DMA_FROM_DEVICE); - off += RX_COPY_MIN; - swivel = RX_COPY_MIN; - RX_USED_ADD(page, cp->mtu_stride); - } else { - RX_USED_ADD(page, hlen); - } - skb_put(skb, alloclen); - - skb_shinfo(skb)->nr_frags++; - skb->data_len += hlen - swivel; - skb->truesize += hlen - swivel; - skb->len += hlen - swivel; - - __skb_frag_set_page(frag, page->buffer); - __skb_frag_ref(frag); - skb_frag_off_set(frag, off); - skb_frag_size_set(frag, hlen - swivel); - - /* any more data? */ - if ((words[0] & RX_COMP1_SPLIT_PKT) && ((dlen -= hlen) > 0)) { - hlen = dlen; - off = 0; - - i = CAS_VAL(RX_COMP2_NEXT_INDEX, words[1]); - page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; - dma_sync_single_for_cpu(&cp->pdev->dev, - page->dma_addr, - hlen + cp->crc_size, - DMA_FROM_DEVICE); - dma_sync_single_for_device(&cp->pdev->dev, - page->dma_addr, - hlen + cp->crc_size, - DMA_FROM_DEVICE); - - skb_shinfo(skb)->nr_frags++; - skb->data_len += hlen; - skb->len += hlen; - frag++; - - __skb_frag_set_page(frag, page->buffer); - __skb_frag_ref(frag); - skb_frag_off_set(frag, 0); - skb_frag_size_set(frag, hlen); - RX_USED_ADD(page, hlen + cp->crc_size); - } - - if (cp->crc_size) - crcaddr = page_address(page->buffer) + off + hlen; - - } else { - /* copying packet */ - if (!dlen) - goto end_copy_pkt; - - i = CAS_VAL(RX_COMP1_DATA_INDEX, words[0]); - page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; - off = CAS_VAL(RX_COMP1_DATA_OFF, words[0]) + swivel; - hlen = min(cp->page_size - off, dlen); - if (hlen < 0) { - netif_printk(cp, rx_err, KERN_DEBUG, cp->dev, - "rx page overflow: %d\n", hlen); - dev_kfree_skb_irq(skb); - return -1; - } - i = hlen; - if (i == dlen) /* attach FCS */ - i += cp->crc_size; - dma_sync_single_for_cpu(&cp->pdev->dev, page->dma_addr + off, - i, DMA_FROM_DEVICE); - memcpy(p, page_address(page->buffer) + off, i); - dma_sync_single_for_device(&cp->pdev->dev, - page->dma_addr + off, i, - DMA_FROM_DEVICE); - if (p == (char *) skb->data) /* not split */ - RX_USED_ADD(page, cp->mtu_stride); - else - RX_USED_ADD(page, i); - - /* any more data? */ - if ((words[0] & RX_COMP1_SPLIT_PKT) && ((dlen -= hlen) > 0)) { - p += hlen; - i = CAS_VAL(RX_COMP2_NEXT_INDEX, words[1]); - page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; - dma_sync_single_for_cpu(&cp->pdev->dev, - page->dma_addr, - dlen + cp->crc_size, - DMA_FROM_DEVICE); - memcpy(p, page_address(page->buffer), dlen + cp->crc_size); - dma_sync_single_for_device(&cp->pdev->dev, - page->dma_addr, - dlen + cp->crc_size, - DMA_FROM_DEVICE); - RX_USED_ADD(page, dlen + cp->crc_size); - } -end_copy_pkt: - if (cp->crc_size) - crcaddr = skb->data + alloclen; - - skb_put(skb, alloclen); - } - - csum = (__force __sum16)htons(CAS_VAL(RX_COMP4_TCP_CSUM, words[3])); - if (cp->crc_size) { - /* checksum includes FCS. strip it out. */ - csum = csum_fold(csum_partial(crcaddr, cp->crc_size, - csum_unfold(csum))); - } - skb->protocol = eth_type_trans(skb, cp->dev); - if (skb->protocol == htons(ETH_P_IP)) { - skb->csum = csum_unfold(~csum); - skb->ip_summed = CHECKSUM_COMPLETE; - } else - skb_checksum_none_assert(skb); - return len; -} - - -/* we can handle up to 64 rx flows at a time. we do the same thing - * as nonreassm except that we batch up the buffers. - * NOTE: we currently just treat each flow as a bunch of packets that - * we pass up. a better way would be to coalesce the packets - * into a jumbo packet. to do that, we need to do the following: - * 1) the first packet will have a clean split between header and - * data. save both. - * 2) each time the next flow packet comes in, extend the - * data length and merge the checksums. - * 3) on flow release, fix up the header. - * 4) make sure the higher layer doesn't care. - * because packets get coalesced, we shouldn't run into fragment count - * issues. - */ -static inline void cas_rx_flow_pkt(struct cas *cp, const u64 *words, - struct sk_buff *skb) -{ - int flowid = CAS_VAL(RX_COMP3_FLOWID, words[2]) & (N_RX_FLOWS - 1); - struct sk_buff_head *flow = &cp->rx_flows[flowid]; - - /* this is protected at a higher layer, so no need to - * do any additional locking here. stick the buffer - * at the end. - */ - __skb_queue_tail(flow, skb); - if (words[0] & RX_COMP1_RELEASE_FLOW) { - while ((skb = __skb_dequeue(flow))) { - cas_skb_release(skb); - } - } -} - -/* put rx descriptor back on ring. if a buffer is in use by a higher - * layer, this will need to put in a replacement. - */ -static void cas_post_page(struct cas *cp, const int ring, const int index) -{ - cas_page_t *new; - int entry; - - entry = cp->rx_old[ring]; - - new = cas_page_swap(cp, ring, index); - cp->init_rxds[ring][entry].buffer = cpu_to_le64(new->dma_addr); - cp->init_rxds[ring][entry].index = - cpu_to_le64(CAS_BASE(RX_INDEX_NUM, index) | - CAS_BASE(RX_INDEX_RING, ring)); - - entry = RX_DESC_ENTRY(ring, entry + 1); - cp->rx_old[ring] = entry; - - if (entry % 4) - return; - - if (ring == 0) - writel(entry, cp->regs + REG_RX_KICK); - else if ((N_RX_DESC_RINGS > 1) && - (cp->cas_flags & CAS_FLAG_REG_PLUS)) - writel(entry, cp->regs + REG_PLUS_RX_KICK1); -} - - -/* only when things are bad */ -static int cas_post_rxds_ringN(struct cas *cp, int ring, int num) -{ - unsigned int entry, last, count, released; - int cluster; - cas_page_t **page = cp->rx_pages[ring]; - - entry = cp->rx_old[ring]; - - netif_printk(cp, intr, KERN_DEBUG, cp->dev, - "rxd[%d] interrupt, done: %d\n", ring, entry); - - cluster = -1; - count = entry & 0x3; - last = RX_DESC_ENTRY(ring, num ? entry + num - 4: entry - 4); - released = 0; - while (entry != last) { - /* make a new buffer if it's still in use */ - if (page_count(page[entry]->buffer) > 1) { - cas_page_t *new = cas_page_dequeue(cp); - if (!new) { - /* let the timer know that we need to - * do this again - */ - cp->cas_flags |= CAS_FLAG_RXD_POST(ring); - if (!timer_pending(&cp->link_timer)) - mod_timer(&cp->link_timer, jiffies + - CAS_LINK_FAST_TIMEOUT); - cp->rx_old[ring] = entry; - cp->rx_last[ring] = num ? num - released : 0; - return -ENOMEM; - } - spin_lock(&cp->rx_inuse_lock); - list_add(&page[entry]->list, &cp->rx_inuse_list); - spin_unlock(&cp->rx_inuse_lock); - cp->init_rxds[ring][entry].buffer = - cpu_to_le64(new->dma_addr); - page[entry] = new; - - } - - if (++count == 4) { - cluster = entry; - count = 0; - } - released++; - entry = RX_DESC_ENTRY(ring, entry + 1); - } - cp->rx_old[ring] = entry; - - if (cluster < 0) - return 0; - - if (ring == 0) - writel(cluster, cp->regs + REG_RX_KICK); - else if ((N_RX_DESC_RINGS > 1) && - (cp->cas_flags & CAS_FLAG_REG_PLUS)) - writel(cluster, cp->regs + REG_PLUS_RX_KICK1); - return 0; -} - - -/* process a completion ring. packets are set up in three basic ways: - * small packets: should be copied header + data in single buffer. - * large packets: header and data in a single buffer. - * split packets: header in a separate buffer from data. - * data may be in multiple pages. data may be > 256 - * bytes but in a single page. - * - * NOTE: RX page posting is done in this routine as well. while there's - * the capability of using multiple RX completion rings, it isn't - * really worthwhile due to the fact that the page posting will - * force serialization on the single descriptor ring. - */ -static int cas_rx_ringN(struct cas *cp, int ring, int budget) -{ - struct cas_rx_comp *rxcs = cp->init_rxcs[ring]; - int entry, drops; - int npackets = 0; - - netif_printk(cp, intr, KERN_DEBUG, cp->dev, - "rx[%d] interrupt, done: %d/%d\n", - ring, - readl(cp->regs + REG_RX_COMP_HEAD), cp->rx_new[ring]); - - entry = cp->rx_new[ring]; - drops = 0; - while (1) { - struct cas_rx_comp *rxc = rxcs + entry; - struct sk_buff *skb; - int type, len; - u64 words[4]; - int i, dring; - - words[0] = le64_to_cpu(rxc->word1); - words[1] = le64_to_cpu(rxc->word2); - words[2] = le64_to_cpu(rxc->word3); - words[3] = le64_to_cpu(rxc->word4); - - /* don't touch if still owned by hw */ - type = CAS_VAL(RX_COMP1_TYPE, words[0]); - if (type == 0) - break; - - /* hw hasn't cleared the zero bit yet */ - if (words[3] & RX_COMP4_ZERO) { - break; - } - - /* get info on the packet */ - if (words[3] & (RX_COMP4_LEN_MISMATCH | RX_COMP4_BAD)) { - spin_lock(&cp->stat_lock[ring]); - cp->net_stats[ring].rx_errors++; - if (words[3] & RX_COMP4_LEN_MISMATCH) - cp->net_stats[ring].rx_length_errors++; - if (words[3] & RX_COMP4_BAD) - cp->net_stats[ring].rx_crc_errors++; - spin_unlock(&cp->stat_lock[ring]); - - /* We'll just return it to Cassini. */ - drop_it: - spin_lock(&cp->stat_lock[ring]); - ++cp->net_stats[ring].rx_dropped; - spin_unlock(&cp->stat_lock[ring]); - goto next; - } - - len = cas_rx_process_pkt(cp, rxc, entry, words, &skb); - if (len < 0) { - ++drops; - goto drop_it; - } - - /* see if it's a flow re-assembly or not. the driver - * itself handles release back up. - */ - if (RX_DONT_BATCH || (type == 0x2)) { - /* non-reassm: these always get released */ - cas_skb_release(skb); - } else { - cas_rx_flow_pkt(cp, words, skb); - } - - spin_lock(&cp->stat_lock[ring]); - cp->net_stats[ring].rx_packets++; - cp->net_stats[ring].rx_bytes += len; - spin_unlock(&cp->stat_lock[ring]); - - next: - npackets++; - - /* should it be released? */ - if (words[0] & RX_COMP1_RELEASE_HDR) { - i = CAS_VAL(RX_COMP2_HDR_INDEX, words[1]); - dring = CAS_VAL(RX_INDEX_RING, i); - i = CAS_VAL(RX_INDEX_NUM, i); - cas_post_page(cp, dring, i); - } - - if (words[0] & RX_COMP1_RELEASE_DATA) { - i = CAS_VAL(RX_COMP1_DATA_INDEX, words[0]); - dring = CAS_VAL(RX_INDEX_RING, i); - i = CAS_VAL(RX_INDEX_NUM, i); - cas_post_page(cp, dring, i); - } - - if (words[0] & RX_COMP1_RELEASE_NEXT) { - i = CAS_VAL(RX_COMP2_NEXT_INDEX, words[1]); - dring = CAS_VAL(RX_INDEX_RING, i); - i = CAS_VAL(RX_INDEX_NUM, i); - cas_post_page(cp, dring, i); - } - - /* skip to the next entry */ - entry = RX_COMP_ENTRY(ring, entry + 1 + - CAS_VAL(RX_COMP1_SKIP, words[0])); -#ifdef USE_NAPI - if (budget && (npackets >= budget)) - break; -#endif - } - cp->rx_new[ring] = entry; - - if (drops) - netdev_info(cp->dev, "Memory squeeze, deferring packet\n"); - return npackets; -} - - -/* put completion entries back on the ring */ -static void cas_post_rxcs_ringN(struct net_device *dev, - struct cas *cp, int ring) -{ - struct cas_rx_comp *rxc = cp->init_rxcs[ring]; - int last, entry; - - last = cp->rx_cur[ring]; - entry = cp->rx_new[ring]; - netif_printk(cp, intr, KERN_DEBUG, dev, - "rxc[%d] interrupt, done: %d/%d\n", - ring, readl(cp->regs + REG_RX_COMP_HEAD), entry); - - /* zero and re-mark descriptors */ - while (last != entry) { - cas_rxc_init(rxc + last); - last = RX_COMP_ENTRY(ring, last + 1); - } - cp->rx_cur[ring] = last; - - if (ring == 0) - writel(last, cp->regs + REG_RX_COMP_TAIL); - else if (cp->cas_flags & CAS_FLAG_REG_PLUS) - writel(last, cp->regs + REG_PLUS_RX_COMPN_TAIL(ring)); -} - - - -/* cassini can use all four PCI interrupts for the completion ring. - * rings 3 and 4 are identical - */ -#if defined(USE_PCI_INTC) || defined(USE_PCI_INTD) -static inline void cas_handle_irqN(struct net_device *dev, - struct cas *cp, const u32 status, - const int ring) -{ - if (status & (INTR_RX_COMP_FULL_ALT | INTR_RX_COMP_AF_ALT)) - cas_post_rxcs_ringN(dev, cp, ring); -} - -static irqreturn_t cas_interruptN(int irq, void *dev_id) -{ - struct net_device *dev = dev_id; - struct cas *cp = netdev_priv(dev); - unsigned long flags; - int ring = (irq == cp->pci_irq_INTC) ? 2 : 3; - u32 status = readl(cp->regs + REG_PLUS_INTRN_STATUS(ring)); - - /* check for shared irq */ - if (status == 0) - return IRQ_NONE; - - spin_lock_irqsave(&cp->lock, flags); - if (status & INTR_RX_DONE_ALT) { /* handle rx separately */ -#ifdef USE_NAPI - cas_mask_intr(cp); - napi_schedule(&cp->napi); -#else - cas_rx_ringN(cp, ring, 0); -#endif - status &= ~INTR_RX_DONE_ALT; - } - - if (status) - cas_handle_irqN(dev, cp, status, ring); - spin_unlock_irqrestore(&cp->lock, flags); - return IRQ_HANDLED; -} -#endif - -#ifdef USE_PCI_INTB -/* everything but rx packets */ -static inline void cas_handle_irq1(struct cas *cp, const u32 status) -{ - if (status & INTR_RX_BUF_UNAVAIL_1) { - /* Frame arrived, no free RX buffers available. - * NOTE: we can get this on a link transition. */ - cas_post_rxds_ringN(cp, 1, 0); - spin_lock(&cp->stat_lock[1]); - cp->net_stats[1].rx_dropped++; - spin_unlock(&cp->stat_lock[1]); - } - - if (status & INTR_RX_BUF_AE_1) - cas_post_rxds_ringN(cp, 1, RX_DESC_RINGN_SIZE(1) - - RX_AE_FREEN_VAL(1)); - - if (status & (INTR_RX_COMP_AF | INTR_RX_COMP_FULL)) - cas_post_rxcs_ringN(cp, 1); -} - -/* ring 2 handles a few more events than 3 and 4 */ -static irqreturn_t cas_interrupt1(int irq, void *dev_id) -{ - struct net_device *dev = dev_id; - struct cas *cp = netdev_priv(dev); - unsigned long flags; - u32 status = readl(cp->regs + REG_PLUS_INTRN_STATUS(1)); - - /* check for shared interrupt */ - if (status == 0) - return IRQ_NONE; - - spin_lock_irqsave(&cp->lock, flags); - if (status & INTR_RX_DONE_ALT) { /* handle rx separately */ -#ifdef USE_NAPI - cas_mask_intr(cp); - napi_schedule(&cp->napi); -#else - cas_rx_ringN(cp, 1, 0); -#endif - status &= ~INTR_RX_DONE_ALT; - } - if (status) - cas_handle_irq1(cp, status); - spin_unlock_irqrestore(&cp->lock, flags); - return IRQ_HANDLED; -} -#endif - -static inline void cas_handle_irq(struct net_device *dev, - struct cas *cp, const u32 status) -{ - /* housekeeping interrupts */ - if (status & INTR_ERROR_MASK) - cas_abnormal_irq(dev, cp, status); - - if (status & INTR_RX_BUF_UNAVAIL) { - /* Frame arrived, no free RX buffers available. - * NOTE: we can get this on a link transition. - */ - cas_post_rxds_ringN(cp, 0, 0); - spin_lock(&cp->stat_lock[0]); - cp->net_stats[0].rx_dropped++; - spin_unlock(&cp->stat_lock[0]); - } else if (status & INTR_RX_BUF_AE) { - cas_post_rxds_ringN(cp, 0, RX_DESC_RINGN_SIZE(0) - - RX_AE_FREEN_VAL(0)); - } - - if (status & (INTR_RX_COMP_AF | INTR_RX_COMP_FULL)) - cas_post_rxcs_ringN(dev, cp, 0); -} - -static irqreturn_t cas_interrupt(int irq, void *dev_id) -{ - struct net_device *dev = dev_id; - struct cas *cp = netdev_priv(dev); - unsigned long flags; - u32 status = readl(cp->regs + REG_INTR_STATUS); - - if (status == 0) - return IRQ_NONE; - - spin_lock_irqsave(&cp->lock, flags); - if (status & (INTR_TX_ALL | INTR_TX_INTME)) { - cas_tx(dev, cp, status); - status &= ~(INTR_TX_ALL | INTR_TX_INTME); - } - - if (status & INTR_RX_DONE) { -#ifdef USE_NAPI - cas_mask_intr(cp); - napi_schedule(&cp->napi); -#else - cas_rx_ringN(cp, 0, 0); -#endif - status &= ~INTR_RX_DONE; - } - - if (status) - cas_handle_irq(dev, cp, status); - spin_unlock_irqrestore(&cp->lock, flags); - return IRQ_HANDLED; -} - - -#ifdef USE_NAPI -static int cas_poll(struct napi_struct *napi, int budget) -{ - struct cas *cp = container_of(napi, struct cas, napi); - struct net_device *dev = cp->dev; - int i, enable_intr, credits; - u32 status = readl(cp->regs + REG_INTR_STATUS); - unsigned long flags; - - spin_lock_irqsave(&cp->lock, flags); - cas_tx(dev, cp, status); - spin_unlock_irqrestore(&cp->lock, flags); - - /* NAPI rx packets. we spread the credits across all of the - * rxc rings - * - * to make sure we're fair with the work we loop through each - * ring N_RX_COMP_RING times with a request of - * budget / N_RX_COMP_RINGS - */ - enable_intr = 1; - credits = 0; - for (i = 0; i < N_RX_COMP_RINGS; i++) { - int j; - for (j = 0; j < N_RX_COMP_RINGS; j++) { - credits += cas_rx_ringN(cp, j, budget / N_RX_COMP_RINGS); - if (credits >= budget) { - enable_intr = 0; - goto rx_comp; - } - } - } - -rx_comp: - /* final rx completion */ - spin_lock_irqsave(&cp->lock, flags); - if (status) - cas_handle_irq(dev, cp, status); - -#ifdef USE_PCI_INTB - if (N_RX_COMP_RINGS > 1) { - status = readl(cp->regs + REG_PLUS_INTRN_STATUS(1)); - if (status) - cas_handle_irq1(dev, cp, status); - } -#endif - -#ifdef USE_PCI_INTC - if (N_RX_COMP_RINGS > 2) { - status = readl(cp->regs + REG_PLUS_INTRN_STATUS(2)); - if (status) - cas_handle_irqN(dev, cp, status, 2); - } -#endif - -#ifdef USE_PCI_INTD - if (N_RX_COMP_RINGS > 3) { - status = readl(cp->regs + REG_PLUS_INTRN_STATUS(3)); - if (status) - cas_handle_irqN(dev, cp, status, 3); - } -#endif - spin_unlock_irqrestore(&cp->lock, flags); - if (enable_intr) { - napi_complete(napi); - cas_unmask_intr(cp); - } - return credits; -} -#endif - -#ifdef CONFIG_NET_POLL_CONTROLLER -static void cas_netpoll(struct net_device *dev) -{ - struct cas *cp = netdev_priv(dev); - - cas_disable_irq(cp, 0); - cas_interrupt(cp->pdev->irq, dev); - cas_enable_irq(cp, 0); - -#ifdef USE_PCI_INTB - if (N_RX_COMP_RINGS > 1) { - /* cas_interrupt1(); */ - } -#endif -#ifdef USE_PCI_INTC - if (N_RX_COMP_RINGS > 2) { - /* cas_interruptN(); */ - } -#endif -#ifdef USE_PCI_INTD - if (N_RX_COMP_RINGS > 3) { - /* cas_interruptN(); */ - } -#endif -} -#endif - -static void cas_tx_timeout(struct net_device *dev, unsigned int txqueue) -{ - struct cas *cp = netdev_priv(dev); - - netdev_err(dev, "transmit timed out, resetting\n"); - if (!cp->hw_running) { - netdev_err(dev, "hrm.. hw not running!\n"); - return; - } - - netdev_err(dev, "MIF_STATE[%08x]\n", - readl(cp->regs + REG_MIF_STATE_MACHINE)); - - netdev_err(dev, "MAC_STATE[%08x]\n", - readl(cp->regs + REG_MAC_STATE_MACHINE)); - - netdev_err(dev, "TX_STATE[%08x:%08x:%08x] FIFO[%08x:%08x:%08x] SM1[%08x] SM2[%08x]\n", - readl(cp->regs + REG_TX_CFG), - readl(cp->regs + REG_MAC_TX_STATUS), - readl(cp->regs + REG_MAC_TX_CFG), - readl(cp->regs + REG_TX_FIFO_PKT_CNT), - readl(cp->regs + REG_TX_FIFO_WRITE_PTR), - readl(cp->regs + REG_TX_FIFO_READ_PTR), - readl(cp->regs + REG_TX_SM_1), - readl(cp->regs + REG_TX_SM_2)); - - netdev_err(dev, "RX_STATE[%08x:%08x:%08x]\n", - readl(cp->regs + REG_RX_CFG), - readl(cp->regs + REG_MAC_RX_STATUS), - readl(cp->regs + REG_MAC_RX_CFG)); - - netdev_err(dev, "HP_STATE[%08x:%08x:%08x:%08x]\n", - readl(cp->regs + REG_HP_STATE_MACHINE), - readl(cp->regs + REG_HP_STATUS0), - readl(cp->regs + REG_HP_STATUS1), - readl(cp->regs + REG_HP_STATUS2)); - -#if 1 - atomic_inc(&cp->reset_task_pending); - atomic_inc(&cp->reset_task_pending_all); - schedule_work(&cp->reset_task); -#else - atomic_set(&cp->reset_task_pending, CAS_RESET_ALL); - schedule_work(&cp->reset_task); -#endif -} - -static inline int cas_intme(int ring, int entry) -{ - /* Algorithm: IRQ every 1/2 of descriptors. */ - if (!(entry & ((TX_DESC_RINGN_SIZE(ring) >> 1) - 1))) - return 1; - return 0; -} - - -static void cas_write_txd(struct cas *cp, int ring, int entry, - dma_addr_t mapping, int len, u64 ctrl, int last) -{ - struct cas_tx_desc *txd = cp->init_txds[ring] + entry; - - ctrl |= CAS_BASE(TX_DESC_BUFLEN, len); - if (cas_intme(ring, entry)) - ctrl |= TX_DESC_INTME; - if (last) - ctrl |= TX_DESC_EOF; - txd->control = cpu_to_le64(ctrl); - txd->buffer = cpu_to_le64(mapping); -} - -static inline void *tx_tiny_buf(struct cas *cp, const int ring, - const int entry) -{ - return cp->tx_tiny_bufs[ring] + TX_TINY_BUF_LEN*entry; -} - -static inline dma_addr_t tx_tiny_map(struct cas *cp, const int ring, - const int entry, const int tentry) -{ - cp->tx_tiny_use[ring][tentry].nbufs++; - cp->tx_tiny_use[ring][entry].used = 1; - return cp->tx_tiny_dvma[ring] + TX_TINY_BUF_LEN*entry; -} - -static inline int cas_xmit_tx_ringN(struct cas *cp, int ring, - struct sk_buff *skb) -{ - struct net_device *dev = cp->dev; - int entry, nr_frags, frag, tabort, tentry; - dma_addr_t mapping; - unsigned long flags; - u64 ctrl; - u32 len; - - spin_lock_irqsave(&cp->tx_lock[ring], flags); - - /* This is a hard error, log it. */ - if (TX_BUFFS_AVAIL(cp, ring) <= - CAS_TABORT(cp)*(skb_shinfo(skb)->nr_frags + 1)) { - netif_stop_queue(dev); - spin_unlock_irqrestore(&cp->tx_lock[ring], flags); - netdev_err(dev, "BUG! Tx Ring full when queue awake!\n"); - return 1; - } - - ctrl = 0; - if (skb->ip_summed == CHECKSUM_PARTIAL) { - const u64 csum_start_off = skb_checksum_start_offset(skb); - const u64 csum_stuff_off = csum_start_off + skb->csum_offset; - - ctrl = TX_DESC_CSUM_EN | - CAS_BASE(TX_DESC_CSUM_START, csum_start_off) | - CAS_BASE(TX_DESC_CSUM_STUFF, csum_stuff_off); - } - - entry = cp->tx_new[ring]; - cp->tx_skbs[ring][entry] = skb; - - nr_frags = skb_shinfo(skb)->nr_frags; - len = skb_headlen(skb); - mapping = dma_map_page(&cp->pdev->dev, virt_to_page(skb->data), - offset_in_page(skb->data), len, DMA_TO_DEVICE); - - tentry = entry; - tabort = cas_calc_tabort(cp, (unsigned long) skb->data, len); - if (unlikely(tabort)) { - /* NOTE: len is always > tabort */ - cas_write_txd(cp, ring, entry, mapping, len - tabort, - ctrl | TX_DESC_SOF, 0); - entry = TX_DESC_NEXT(ring, entry); - - skb_copy_from_linear_data_offset(skb, len - tabort, - tx_tiny_buf(cp, ring, entry), tabort); - mapping = tx_tiny_map(cp, ring, entry, tentry); - cas_write_txd(cp, ring, entry, mapping, tabort, ctrl, - (nr_frags == 0)); - } else { - cas_write_txd(cp, ring, entry, mapping, len, ctrl | - TX_DESC_SOF, (nr_frags == 0)); - } - entry = TX_DESC_NEXT(ring, entry); - - for (frag = 0; frag < nr_frags; frag++) { - const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; - - len = skb_frag_size(fragp); - mapping = skb_frag_dma_map(&cp->pdev->dev, fragp, 0, len, - DMA_TO_DEVICE); - - tabort = cas_calc_tabort(cp, skb_frag_off(fragp), len); - if (unlikely(tabort)) { - /* NOTE: len is always > tabort */ - cas_write_txd(cp, ring, entry, mapping, len - tabort, - ctrl, 0); - entry = TX_DESC_NEXT(ring, entry); - memcpy_from_page(tx_tiny_buf(cp, ring, entry), - skb_frag_page(fragp), - skb_frag_off(fragp) + len - tabort, - tabort); - mapping = tx_tiny_map(cp, ring, entry, tentry); - len = tabort; - } - - cas_write_txd(cp, ring, entry, mapping, len, ctrl, - (frag + 1 == nr_frags)); - entry = TX_DESC_NEXT(ring, entry); - } - - cp->tx_new[ring] = entry; - if (TX_BUFFS_AVAIL(cp, ring) <= CAS_TABORT(cp)*(MAX_SKB_FRAGS + 1)) - netif_stop_queue(dev); - - netif_printk(cp, tx_queued, KERN_DEBUG, dev, - "tx[%d] queued, slot %d, skblen %d, avail %d\n", - ring, entry, skb->len, TX_BUFFS_AVAIL(cp, ring)); - writel(entry, cp->regs + REG_TX_KICKN(ring)); - spin_unlock_irqrestore(&cp->tx_lock[ring], flags); - return 0; -} - -static netdev_tx_t cas_start_xmit(struct sk_buff *skb, struct net_device *dev) -{ - struct cas *cp = netdev_priv(dev); - - /* this is only used as a load-balancing hint, so it doesn't - * need to be SMP safe - */ - static int ring; - - if (skb_padto(skb, cp->min_frame_size)) - return NETDEV_TX_OK; - - /* XXX: we need some higher-level QoS hooks to steer packets to - * individual queues. - */ - if (cas_xmit_tx_ringN(cp, ring++ & N_TX_RINGS_MASK, skb)) - return NETDEV_TX_BUSY; - return NETDEV_TX_OK; -} - -static void cas_init_tx_dma(struct cas *cp) -{ - u64 desc_dma = cp->block_dvma; - unsigned long off; - u32 val; - int i; - - /* set up tx completion writeback registers. must be 8-byte aligned */ -#ifdef USE_TX_COMPWB - off = offsetof(struct cas_init_block, tx_compwb); - writel((desc_dma + off) >> 32, cp->regs + REG_TX_COMPWB_DB_HI); - writel((desc_dma + off) & 0xffffffff, cp->regs + REG_TX_COMPWB_DB_LOW); -#endif - - /* enable completion writebacks, enable paced mode, - * disable read pipe, and disable pre-interrupt compwbs - */ - val = TX_CFG_COMPWB_Q1 | TX_CFG_COMPWB_Q2 | - TX_CFG_COMPWB_Q3 | TX_CFG_COMPWB_Q4 | - TX_CFG_DMA_RDPIPE_DIS | TX_CFG_PACED_MODE | - TX_CFG_INTR_COMPWB_DIS; - - /* write out tx ring info and tx desc bases */ - for (i = 0; i < MAX_TX_RINGS; i++) { - off = (unsigned long) cp->init_txds[i] - - (unsigned long) cp->init_block; - - val |= CAS_TX_RINGN_BASE(i); - writel((desc_dma + off) >> 32, cp->regs + REG_TX_DBN_HI(i)); - writel((desc_dma + off) & 0xffffffff, cp->regs + - REG_TX_DBN_LOW(i)); - /* don't zero out the kick register here as the system - * will wedge - */ - } - writel(val, cp->regs + REG_TX_CFG); - - /* program max burst sizes. these numbers should be different - * if doing QoS. - */ -#ifdef USE_QOS - writel(0x800, cp->regs + REG_TX_MAXBURST_0); - writel(0x1600, cp->regs + REG_TX_MAXBURST_1); - writel(0x2400, cp->regs + REG_TX_MAXBURST_2); - writel(0x4800, cp->regs + REG_TX_MAXBURST_3); -#else - writel(0x800, cp->regs + REG_TX_MAXBURST_0); - writel(0x800, cp->regs + REG_TX_MAXBURST_1); - writel(0x800, cp->regs + REG_TX_MAXBURST_2); - writel(0x800, cp->regs + REG_TX_MAXBURST_3); -#endif -} - -/* Must be invoked under cp->lock. */ -static inline void cas_init_dma(struct cas *cp) -{ - cas_init_tx_dma(cp); - cas_init_rx_dma(cp); -} - -static void cas_process_mc_list(struct cas *cp) -{ - u16 hash_table[16]; - u32 crc; - struct netdev_hw_addr *ha; - int i = 1; - - memset(hash_table, 0, sizeof(hash_table)); - netdev_for_each_mc_addr(ha, cp->dev) { - if (i <= CAS_MC_EXACT_MATCH_SIZE) { - /* use the alternate mac address registers for the - * first 15 multicast addresses - */ - writel((ha->addr[4] << 8) | ha->addr[5], - cp->regs + REG_MAC_ADDRN(i*3 + 0)); - writel((ha->addr[2] << 8) | ha->addr[3], - cp->regs + REG_MAC_ADDRN(i*3 + 1)); - writel((ha->addr[0] << 8) | ha->addr[1], - cp->regs + REG_MAC_ADDRN(i*3 + 2)); - i++; - } - else { - /* use hw hash table for the next series of - * multicast addresses - */ - crc = ether_crc_le(ETH_ALEN, ha->addr); - crc >>= 24; - hash_table[crc >> 4] |= 1 << (15 - (crc & 0xf)); - } - } - for (i = 0; i < 16; i++) - writel(hash_table[i], cp->regs + REG_MAC_HASH_TABLEN(i)); -} - -/* Must be invoked under cp->lock. */ -static u32 cas_setup_multicast(struct cas *cp) -{ - u32 rxcfg = 0; - int i; - - if (cp->dev->flags & IFF_PROMISC) { - rxcfg |= MAC_RX_CFG_PROMISC_EN; - - } else if (cp->dev->flags & IFF_ALLMULTI) { - for (i=0; i < 16; i++) - writel(0xFFFF, cp->regs + REG_MAC_HASH_TABLEN(i)); - rxcfg |= MAC_RX_CFG_HASH_FILTER_EN; - - } else { - cas_process_mc_list(cp); - rxcfg |= MAC_RX_CFG_HASH_FILTER_EN; - } - - return rxcfg; -} - -/* must be invoked under cp->stat_lock[N_TX_RINGS] */ -static void cas_clear_mac_err(struct cas *cp) -{ - writel(0, cp->regs + REG_MAC_COLL_NORMAL); - writel(0, cp->regs + REG_MAC_COLL_FIRST); - writel(0, cp->regs + REG_MAC_COLL_EXCESS); - writel(0, cp->regs + REG_MAC_COLL_LATE); - writel(0, cp->regs + REG_MAC_TIMER_DEFER); - writel(0, cp->regs + REG_MAC_ATTEMPTS_PEAK); - writel(0, cp->regs + REG_MAC_RECV_FRAME); - writel(0, cp->regs + REG_MAC_LEN_ERR); - writel(0, cp->regs + REG_MAC_ALIGN_ERR); - writel(0, cp->regs + REG_MAC_FCS_ERR); - writel(0, cp->regs + REG_MAC_RX_CODE_ERR); -} - - -static void cas_mac_reset(struct cas *cp) -{ - int i; - - /* do both TX and RX reset */ - writel(0x1, cp->regs + REG_MAC_TX_RESET); - writel(0x1, cp->regs + REG_MAC_RX_RESET); - - /* wait for TX */ - i = STOP_TRIES; - while (i-- > 0) { - if (readl(cp->regs + REG_MAC_TX_RESET) == 0) - break; - udelay(10); - } - - /* wait for RX */ - i = STOP_TRIES; - while (i-- > 0) { - if (readl(cp->regs + REG_MAC_RX_RESET) == 0) - break; - udelay(10); - } - - if (readl(cp->regs + REG_MAC_TX_RESET) | - readl(cp->regs + REG_MAC_RX_RESET)) - netdev_err(cp->dev, "mac tx[%d]/rx[%d] reset failed [%08x]\n", - readl(cp->regs + REG_MAC_TX_RESET), - readl(cp->regs + REG_MAC_RX_RESET), - readl(cp->regs + REG_MAC_STATE_MACHINE)); -} - - -/* Must be invoked under cp->lock. */ -static void cas_init_mac(struct cas *cp) -{ - const unsigned char *e = &cp->dev->dev_addr[0]; - int i; - cas_mac_reset(cp); - - /* setup core arbitration weight register */ - writel(CAWR_RR_DIS, cp->regs + REG_CAWR); - -#if !defined(CONFIG_SPARC64) && !defined(CONFIG_ALPHA) - /* set the infinite burst register for chips that don't have - * pci issues. - */ - if ((cp->cas_flags & CAS_FLAG_TARGET_ABORT) == 0) - writel(INF_BURST_EN, cp->regs + REG_INF_BURST); -#endif - - writel(0x1BF0, cp->regs + REG_MAC_SEND_PAUSE); - - writel(0x00, cp->regs + REG_MAC_IPG0); - writel(0x08, cp->regs + REG_MAC_IPG1); - writel(0x04, cp->regs + REG_MAC_IPG2); - - /* change later for 802.3z */ - writel(0x40, cp->regs + REG_MAC_SLOT_TIME); - - /* min frame + FCS */ - writel(ETH_ZLEN + 4, cp->regs + REG_MAC_FRAMESIZE_MIN); - - /* Ethernet payload + header + FCS + optional VLAN tag. NOTE: we - * specify the maximum frame size to prevent RX tag errors on - * oversized frames. - */ - writel(CAS_BASE(MAC_FRAMESIZE_MAX_BURST, 0x2000) | - CAS_BASE(MAC_FRAMESIZE_MAX_FRAME, - (CAS_MAX_MTU + ETH_HLEN + 4 + 4)), - cp->regs + REG_MAC_FRAMESIZE_MAX); - - /* NOTE: crc_size is used as a surrogate for half-duplex. - * workaround saturn half-duplex issue by increasing preamble - * size to 65 bytes. - */ - if ((cp->cas_flags & CAS_FLAG_SATURN) && cp->crc_size) - writel(0x41, cp->regs + REG_MAC_PA_SIZE); - else - writel(0x07, cp->regs + REG_MAC_PA_SIZE); - writel(0x04, cp->regs + REG_MAC_JAM_SIZE); - writel(0x10, cp->regs + REG_MAC_ATTEMPT_LIMIT); - writel(0x8808, cp->regs + REG_MAC_CTRL_TYPE); - - writel((e[5] | (e[4] << 8)) & 0x3ff, cp->regs + REG_MAC_RANDOM_SEED); - - writel(0, cp->regs + REG_MAC_ADDR_FILTER0); - writel(0, cp->regs + REG_MAC_ADDR_FILTER1); - writel(0, cp->regs + REG_MAC_ADDR_FILTER2); - writel(0, cp->regs + REG_MAC_ADDR_FILTER2_1_MASK); - writel(0, cp->regs + REG_MAC_ADDR_FILTER0_MASK); - - /* setup mac address in perfect filter array */ - for (i = 0; i < 45; i++) - writel(0x0, cp->regs + REG_MAC_ADDRN(i)); - - writel((e[4] << 8) | e[5], cp->regs + REG_MAC_ADDRN(0)); - writel((e[2] << 8) | e[3], cp->regs + REG_MAC_ADDRN(1)); - writel((e[0] << 8) | e[1], cp->regs + REG_MAC_ADDRN(2)); - - writel(0x0001, cp->regs + REG_MAC_ADDRN(42)); - writel(0xc200, cp->regs + REG_MAC_ADDRN(43)); - writel(0x0180, cp->regs + REG_MAC_ADDRN(44)); - - cp->mac_rx_cfg = cas_setup_multicast(cp); - - spin_lock(&cp->stat_lock[N_TX_RINGS]); - cas_clear_mac_err(cp); - spin_unlock(&cp->stat_lock[N_TX_RINGS]); - - /* Setup MAC interrupts. We want to get all of the interesting - * counter expiration events, but we do not want to hear about - * normal rx/tx as the DMA engine tells us that. - */ - writel(MAC_TX_FRAME_XMIT, cp->regs + REG_MAC_TX_MASK); - writel(MAC_RX_FRAME_RECV, cp->regs + REG_MAC_RX_MASK); - - /* Don't enable even the PAUSE interrupts for now, we - * make no use of those events other than to record them. - */ - writel(0xffffffff, cp->regs + REG_MAC_CTRL_MASK); -} - -/* Must be invoked under cp->lock. */ -static void cas_init_pause_thresholds(struct cas *cp) -{ - /* Calculate pause thresholds. Setting the OFF threshold to the - * full RX fifo size effectively disables PAUSE generation - */ - if (cp->rx_fifo_size <= (2 * 1024)) { - cp->rx_pause_off = cp->rx_pause_on = cp->rx_fifo_size; - } else { - int max_frame = (cp->dev->mtu + ETH_HLEN + 4 + 4 + 64) & ~63; - if (max_frame * 3 > cp->rx_fifo_size) { - cp->rx_pause_off = 7104; - cp->rx_pause_on = 960; - } else { - int off = (cp->rx_fifo_size - (max_frame * 2)); - int on = off - max_frame; - cp->rx_pause_off = off; - cp->rx_pause_on = on; - } - } -} - -static int cas_vpd_match(const void __iomem *p, const char *str) -{ - int len = strlen(str) + 1; - int i; - - for (i = 0; i < len; i++) { - if (readb(p + i) != str[i]) - return 0; - } - return 1; -} - - -/* get the mac address by reading the vpd information in the rom. - * also get the phy type and determine if there's an entropy generator. - * NOTE: this is a bit convoluted for the following reasons: - * 1) vpd info has order-dependent mac addresses for multinic cards - * 2) the only way to determine the nic order is to use the slot - * number. - * 3) fiber cards don't have bridges, so their slot numbers don't - * mean anything. - * 4) we don't actually know we have a fiber card until after - * the mac addresses are parsed. - */ -static int cas_get_vpd_info(struct cas *cp, unsigned char *dev_addr, - const int offset) -{ - void __iomem *p = cp->regs + REG_EXPANSION_ROM_RUN_START; - void __iomem *base, *kstart; - int i, len; - int found = 0; -#define VPD_FOUND_MAC 0x01 -#define VPD_FOUND_PHY 0x02 - - int phy_type = CAS_PHY_MII_MDIO0; /* default phy type */ - int mac_off = 0; - -#if defined(CONFIG_SPARC) - const unsigned char *addr; -#endif - - /* give us access to the PROM */ - writel(BIM_LOCAL_DEV_PROM | BIM_LOCAL_DEV_PAD, - cp->regs + REG_BIM_LOCAL_DEV_EN); - - /* check for an expansion rom */ - if (readb(p) != 0x55 || readb(p + 1) != 0xaa) - goto use_random_mac_addr; - - /* search for beginning of vpd */ - base = NULL; - for (i = 2; i < EXPANSION_ROM_SIZE; i++) { - /* check for PCIR */ - if ((readb(p + i + 0) == 0x50) && - (readb(p + i + 1) == 0x43) && - (readb(p + i + 2) == 0x49) && - (readb(p + i + 3) == 0x52)) { - base = p + (readb(p + i + 8) | - (readb(p + i + 9) << 8)); - break; - } - } - - if (!base || (readb(base) != 0x82)) - goto use_random_mac_addr; - - i = (readb(base + 1) | (readb(base + 2) << 8)) + 3; - while (i < EXPANSION_ROM_SIZE) { - if (readb(base + i) != 0x90) /* no vpd found */ - goto use_random_mac_addr; - - /* found a vpd field */ - len = readb(base + i + 1) | (readb(base + i + 2) << 8); - - /* extract keywords */ - kstart = base + i + 3; - p = kstart; - while ((p - kstart) < len) { - int klen = readb(p + 2); - int j; - char type; - - p += 3; - - /* look for the following things: - * -- correct length == 29 - * 3 (type) + 2 (size) + - * 18 (strlen("local-mac-address") + 1) + - * 6 (mac addr) - * -- VPD Instance 'I' - * -- VPD Type Bytes 'B' - * -- VPD data length == 6 - * -- property string == local-mac-address - * - * -- correct length == 24 - * 3 (type) + 2 (size) + - * 12 (strlen("entropy-dev") + 1) + - * 7 (strlen("vms110") + 1) - * -- VPD Instance 'I' - * -- VPD Type String 'B' - * -- VPD data length == 7 - * -- property string == entropy-dev - * - * -- correct length == 18 - * 3 (type) + 2 (size) + - * 9 (strlen("phy-type") + 1) + - * 4 (strlen("pcs") + 1) - * -- VPD Instance 'I' - * -- VPD Type String 'S' - * -- VPD data length == 4 - * -- property string == phy-type - * - * -- correct length == 23 - * 3 (type) + 2 (size) + - * 14 (strlen("phy-interface") + 1) + - * 4 (strlen("pcs") + 1) - * -- VPD Instance 'I' - * -- VPD Type String 'S' - * -- VPD data length == 4 - * -- property string == phy-interface - */ - if (readb(p) != 'I') - goto next; - - /* finally, check string and length */ - type = readb(p + 3); - if (type == 'B') { - if ((klen == 29) && readb(p + 4) == 6 && - cas_vpd_match(p + 5, - "local-mac-address")) { - if (mac_off++ > offset) - goto next; - - /* set mac address */ - for (j = 0; j < 6; j++) - dev_addr[j] = - readb(p + 23 + j); - goto found_mac; - } - } - - if (type != 'S') - goto next; - -#ifdef USE_ENTROPY_DEV - if ((klen == 24) && - cas_vpd_match(p + 5, "entropy-dev") && - cas_vpd_match(p + 17, "vms110")) { - cp->cas_flags |= CAS_FLAG_ENTROPY_DEV; - goto next; - } -#endif - - if (found & VPD_FOUND_PHY) - goto next; - - if ((klen == 18) && readb(p + 4) == 4 && - cas_vpd_match(p + 5, "phy-type")) { - if (cas_vpd_match(p + 14, "pcs")) { - phy_type = CAS_PHY_SERDES; - goto found_phy; - } - } - - if ((klen == 23) && readb(p + 4) == 4 && - cas_vpd_match(p + 5, "phy-interface")) { - if (cas_vpd_match(p + 19, "pcs")) { - phy_type = CAS_PHY_SERDES; - goto found_phy; - } - } -found_mac: - found |= VPD_FOUND_MAC; - goto next; - -found_phy: - found |= VPD_FOUND_PHY; - -next: - p += klen; - } - i += len + 3; - } - -use_random_mac_addr: - if (found & VPD_FOUND_MAC) - goto done; - -#if defined(CONFIG_SPARC) - addr = of_get_property(cp->of_node, "local-mac-address", NULL); - if (addr != NULL) { - memcpy(dev_addr, addr, ETH_ALEN); - goto done; - } -#endif - - /* Sun MAC prefix then 3 random bytes. */ - pr_info("MAC address not found in ROM VPD\n"); - dev_addr[0] = 0x08; - dev_addr[1] = 0x00; - dev_addr[2] = 0x20; - get_random_bytes(dev_addr + 3, 3); - -done: - writel(0, cp->regs + REG_BIM_LOCAL_DEV_EN); - return phy_type; -} - -/* check pci invariants */ -static void cas_check_pci_invariants(struct cas *cp) -{ - struct pci_dev *pdev = cp->pdev; - - cp->cas_flags = 0; - if ((pdev->vendor == PCI_VENDOR_ID_SUN) && - (pdev->device == PCI_DEVICE_ID_SUN_CASSINI)) { - if (pdev->revision >= CAS_ID_REVPLUS) - cp->cas_flags |= CAS_FLAG_REG_PLUS; - if (pdev->revision < CAS_ID_REVPLUS02u) - cp->cas_flags |= CAS_FLAG_TARGET_ABORT; - - /* Original Cassini supports HW CSUM, but it's not - * enabled by default as it can trigger TX hangs. - */ - if (pdev->revision < CAS_ID_REV2) - cp->cas_flags |= CAS_FLAG_NO_HW_CSUM; - } else { - /* Only sun has original cassini chips. */ - cp->cas_flags |= CAS_FLAG_REG_PLUS; - - /* We use a flag because the same phy might be externally - * connected. - */ - if ((pdev->vendor == PCI_VENDOR_ID_NS) && - (pdev->device == PCI_DEVICE_ID_NS_SATURN)) - cp->cas_flags |= CAS_FLAG_SATURN; - } -} - - -static int cas_check_invariants(struct cas *cp) -{ - struct pci_dev *pdev = cp->pdev; - u8 addr[ETH_ALEN]; - u32 cfg; - int i; - - /* get page size for rx buffers. */ - cp->page_order = 0; -#ifdef USE_PAGE_ORDER - if (PAGE_SHIFT < CAS_JUMBO_PAGE_SHIFT) { - /* see if we can allocate larger pages */ - struct page *page = alloc_pages(GFP_ATOMIC, - CAS_JUMBO_PAGE_SHIFT - - PAGE_SHIFT); - if (page) { - __free_pages(page, CAS_JUMBO_PAGE_SHIFT - PAGE_SHIFT); - cp->page_order = CAS_JUMBO_PAGE_SHIFT - PAGE_SHIFT; - } else { - printk("MTU limited to %d bytes\n", CAS_MAX_MTU); - } - } -#endif - cp->page_size = (PAGE_SIZE << cp->page_order); - - /* Fetch the FIFO configurations. */ - cp->tx_fifo_size = readl(cp->regs + REG_TX_FIFO_SIZE) * 64; - cp->rx_fifo_size = RX_FIFO_SIZE; - - /* finish phy determination. MDIO1 takes precedence over MDIO0 if - * they're both connected. - */ - cp->phy_type = cas_get_vpd_info(cp, addr, PCI_SLOT(pdev->devfn)); - eth_hw_addr_set(cp->dev, addr); - if (cp->phy_type & CAS_PHY_SERDES) { - cp->cas_flags |= CAS_FLAG_1000MB_CAP; - return 0; /* no more checking needed */ - } - - /* MII */ - cfg = readl(cp->regs + REG_MIF_CFG); - if (cfg & MIF_CFG_MDIO_1) { - cp->phy_type = CAS_PHY_MII_MDIO1; - } else if (cfg & MIF_CFG_MDIO_0) { - cp->phy_type = CAS_PHY_MII_MDIO0; - } - - cas_mif_poll(cp, 0); - writel(PCS_DATAPATH_MODE_MII, cp->regs + REG_PCS_DATAPATH_MODE); - - for (i = 0; i < 32; i++) { - u32 phy_id; - int j; - - for (j = 0; j < 3; j++) { - cp->phy_addr = i; - phy_id = cas_phy_read(cp, MII_PHYSID1) << 16; - phy_id |= cas_phy_read(cp, MII_PHYSID2); - if (phy_id && (phy_id != 0xFFFFFFFF)) { - cp->phy_id = phy_id; - goto done; - } - } - } - pr_err("MII phy did not respond [%08x]\n", - readl(cp->regs + REG_MIF_STATE_MACHINE)); - return -1; - -done: - /* see if we can do gigabit */ - cfg = cas_phy_read(cp, MII_BMSR); - if ((cfg & CAS_BMSR_1000_EXTEND) && - cas_phy_read(cp, CAS_MII_1000_EXTEND)) - cp->cas_flags |= CAS_FLAG_1000MB_CAP; - return 0; -} - -/* Must be invoked under cp->lock. */ -static inline void cas_start_dma(struct cas *cp) -{ - int i; - u32 val; - int txfailed = 0; - - /* enable dma */ - val = readl(cp->regs + REG_TX_CFG) | TX_CFG_DMA_EN; - writel(val, cp->regs + REG_TX_CFG); - val = readl(cp->regs + REG_RX_CFG) | RX_CFG_DMA_EN; - writel(val, cp->regs + REG_RX_CFG); - - /* enable the mac */ - val = readl(cp->regs + REG_MAC_TX_CFG) | MAC_TX_CFG_EN; - writel(val, cp->regs + REG_MAC_TX_CFG); - val = readl(cp->regs + REG_MAC_RX_CFG) | MAC_RX_CFG_EN; - writel(val, cp->regs + REG_MAC_RX_CFG); - - i = STOP_TRIES; - while (i-- > 0) { - val = readl(cp->regs + REG_MAC_TX_CFG); - if ((val & MAC_TX_CFG_EN)) - break; - udelay(10); - } - if (i < 0) txfailed = 1; - i = STOP_TRIES; - while (i-- > 0) { - val = readl(cp->regs + REG_MAC_RX_CFG); - if ((val & MAC_RX_CFG_EN)) { - if (txfailed) { - netdev_err(cp->dev, - "enabling mac failed [tx:%08x:%08x]\n", - readl(cp->regs + REG_MIF_STATE_MACHINE), - readl(cp->regs + REG_MAC_STATE_MACHINE)); - } - goto enable_rx_done; - } - udelay(10); - } - netdev_err(cp->dev, "enabling mac failed [%s:%08x:%08x]\n", - (txfailed ? "tx,rx" : "rx"), - readl(cp->regs + REG_MIF_STATE_MACHINE), - readl(cp->regs + REG_MAC_STATE_MACHINE)); - -enable_rx_done: - cas_unmask_intr(cp); /* enable interrupts */ - writel(RX_DESC_RINGN_SIZE(0) - 4, cp->regs + REG_RX_KICK); - writel(0, cp->regs + REG_RX_COMP_TAIL); - - if (cp->cas_flags & CAS_FLAG_REG_PLUS) { - if (N_RX_DESC_RINGS > 1) - writel(RX_DESC_RINGN_SIZE(1) - 4, - cp->regs + REG_PLUS_RX_KICK1); - } -} - -/* Must be invoked under cp->lock. */ -static void cas_read_pcs_link_mode(struct cas *cp, int *fd, int *spd, - int *pause) -{ - u32 val = readl(cp->regs + REG_PCS_MII_LPA); - *fd = (val & PCS_MII_LPA_FD) ? 1 : 0; - *pause = (val & PCS_MII_LPA_SYM_PAUSE) ? 0x01 : 0x00; - if (val & PCS_MII_LPA_ASYM_PAUSE) - *pause |= 0x10; - *spd = 1000; -} - -/* Must be invoked under cp->lock. */ -static void cas_read_mii_link_mode(struct cas *cp, int *fd, int *spd, - int *pause) -{ - u32 val; - - *fd = 0; - *spd = 10; - *pause = 0; - - /* use GMII registers */ - val = cas_phy_read(cp, MII_LPA); - if (val & CAS_LPA_PAUSE) - *pause = 0x01; - - if (val & CAS_LPA_ASYM_PAUSE) - *pause |= 0x10; - - if (val & LPA_DUPLEX) - *fd = 1; - if (val & LPA_100) - *spd = 100; - - if (cp->cas_flags & CAS_FLAG_1000MB_CAP) { - val = cas_phy_read(cp, CAS_MII_1000_STATUS); - if (val & (CAS_LPA_1000FULL | CAS_LPA_1000HALF)) - *spd = 1000; - if (val & CAS_LPA_1000FULL) - *fd = 1; - } -} - -/* A link-up condition has occurred, initialize and enable the - * rest of the chip. - * - * Must be invoked under cp->lock. - */ -static void cas_set_link_modes(struct cas *cp) -{ - u32 val; - int full_duplex, speed, pause; - - full_duplex = 0; - speed = 10; - pause = 0; - - if (CAS_PHY_MII(cp->phy_type)) { - cas_mif_poll(cp, 0); - val = cas_phy_read(cp, MII_BMCR); - if (val & BMCR_ANENABLE) { - cas_read_mii_link_mode(cp, &full_duplex, &speed, - &pause); - } else { - if (val & BMCR_FULLDPLX) - full_duplex = 1; - - if (val & BMCR_SPEED100) - speed = 100; - else if (val & CAS_BMCR_SPEED1000) - speed = (cp->cas_flags & CAS_FLAG_1000MB_CAP) ? - 1000 : 100; - } - cas_mif_poll(cp, 1); - - } else { - val = readl(cp->regs + REG_PCS_MII_CTRL); - cas_read_pcs_link_mode(cp, &full_duplex, &speed, &pause); - if ((val & PCS_MII_AUTONEG_EN) == 0) { - if (val & PCS_MII_CTRL_DUPLEX) - full_duplex = 1; - } - } - - netif_info(cp, link, cp->dev, "Link up at %d Mbps, %s-duplex\n", - speed, full_duplex ? "full" : "half"); - - val = MAC_XIF_TX_MII_OUTPUT_EN | MAC_XIF_LINK_LED; - if (CAS_PHY_MII(cp->phy_type)) { - val |= MAC_XIF_MII_BUFFER_OUTPUT_EN; - if (!full_duplex) - val |= MAC_XIF_DISABLE_ECHO; - } - if (full_duplex) - val |= MAC_XIF_FDPLX_LED; - if (speed == 1000) - val |= MAC_XIF_GMII_MODE; - writel(val, cp->regs + REG_MAC_XIF_CFG); - - /* deal with carrier and collision detect. */ - val = MAC_TX_CFG_IPG_EN; - if (full_duplex) { - val |= MAC_TX_CFG_IGNORE_CARRIER; - val |= MAC_TX_CFG_IGNORE_COLL; - } else { -#ifndef USE_CSMA_CD_PROTO - val |= MAC_TX_CFG_NEVER_GIVE_UP_EN; - val |= MAC_TX_CFG_NEVER_GIVE_UP_LIM; -#endif - } - /* val now set up for REG_MAC_TX_CFG */ - - /* If gigabit and half-duplex, enable carrier extension - * mode. increase slot time to 512 bytes as well. - * else, disable it and make sure slot time is 64 bytes. - * also activate checksum bug workaround - */ - if ((speed == 1000) && !full_duplex) { - writel(val | MAC_TX_CFG_CARRIER_EXTEND, - cp->regs + REG_MAC_TX_CFG); - - val = readl(cp->regs + REG_MAC_RX_CFG); - val &= ~MAC_RX_CFG_STRIP_FCS; /* checksum workaround */ - writel(val | MAC_RX_CFG_CARRIER_EXTEND, - cp->regs + REG_MAC_RX_CFG); - - writel(0x200, cp->regs + REG_MAC_SLOT_TIME); - - cp->crc_size = 4; - /* minimum size gigabit frame at half duplex */ - cp->min_frame_size = CAS_1000MB_MIN_FRAME; - - } else { - writel(val, cp->regs + REG_MAC_TX_CFG); - - /* checksum bug workaround. don't strip FCS when in - * half-duplex mode - */ - val = readl(cp->regs + REG_MAC_RX_CFG); - if (full_duplex) { - val |= MAC_RX_CFG_STRIP_FCS; - cp->crc_size = 0; - cp->min_frame_size = CAS_MIN_MTU; - } else { - val &= ~MAC_RX_CFG_STRIP_FCS; - cp->crc_size = 4; - cp->min_frame_size = CAS_MIN_FRAME; - } - writel(val & ~MAC_RX_CFG_CARRIER_EXTEND, - cp->regs + REG_MAC_RX_CFG); - writel(0x40, cp->regs + REG_MAC_SLOT_TIME); - } - - if (netif_msg_link(cp)) { - if (pause & 0x01) { - netdev_info(cp->dev, "Pause is enabled (rxfifo: %d off: %d on: %d)\n", - cp->rx_fifo_size, - cp->rx_pause_off, - cp->rx_pause_on); - } else if (pause & 0x10) { - netdev_info(cp->dev, "TX pause enabled\n"); - } else { - netdev_info(cp->dev, "Pause is disabled\n"); - } - } - - val = readl(cp->regs + REG_MAC_CTRL_CFG); - val &= ~(MAC_CTRL_CFG_SEND_PAUSE_EN | MAC_CTRL_CFG_RECV_PAUSE_EN); - if (pause) { /* symmetric or asymmetric pause */ - val |= MAC_CTRL_CFG_SEND_PAUSE_EN; - if (pause & 0x01) { /* symmetric pause */ - val |= MAC_CTRL_CFG_RECV_PAUSE_EN; - } - } - writel(val, cp->regs + REG_MAC_CTRL_CFG); - cas_start_dma(cp); -} - -/* Must be invoked under cp->lock. */ -static void cas_init_hw(struct cas *cp, int restart_link) -{ - if (restart_link) - cas_phy_init(cp); - - cas_init_pause_thresholds(cp); - cas_init_mac(cp); - cas_init_dma(cp); - - if (restart_link) { - /* Default aneg parameters */ - cp->timer_ticks = 0; - cas_begin_auto_negotiation(cp, NULL); - } else if (cp->lstate == link_up) { - cas_set_link_modes(cp); - netif_carrier_on(cp->dev); - } -} - -/* Must be invoked under cp->lock. on earlier cassini boards, - * SOFT_0 is tied to PCI reset. we use this to force a pci reset, - * let it settle out, and then restore pci state. - */ -static void cas_hard_reset(struct cas *cp) -{ - writel(BIM_LOCAL_DEV_SOFT_0, cp->regs + REG_BIM_LOCAL_DEV_EN); - udelay(20); - pci_restore_state(cp->pdev); -} - - -static void cas_global_reset(struct cas *cp, int blkflag) -{ - int limit; - - /* issue a global reset. don't use RSTOUT. */ - if (blkflag && !CAS_PHY_MII(cp->phy_type)) { - /* For PCS, when the blkflag is set, we should set the - * SW_REST_BLOCK_PCS_SLINK bit to prevent the results of - * the last autonegotiation from being cleared. We'll - * need some special handling if the chip is set into a - * loopback mode. - */ - writel((SW_RESET_TX | SW_RESET_RX | SW_RESET_BLOCK_PCS_SLINK), - cp->regs + REG_SW_RESET); - } else { - writel(SW_RESET_TX | SW_RESET_RX, cp->regs + REG_SW_RESET); - } - - /* need to wait at least 3ms before polling register */ - mdelay(3); - - limit = STOP_TRIES; - while (limit-- > 0) { - u32 val = readl(cp->regs + REG_SW_RESET); - if ((val & (SW_RESET_TX | SW_RESET_RX)) == 0) - goto done; - udelay(10); - } - netdev_err(cp->dev, "sw reset failed\n"); - -done: - /* enable various BIM interrupts */ - writel(BIM_CFG_DPAR_INTR_ENABLE | BIM_CFG_RMA_INTR_ENABLE | - BIM_CFG_RTA_INTR_ENABLE, cp->regs + REG_BIM_CFG); - - /* clear out pci error status mask for handled errors. - * we don't deal with DMA counter overflows as they happen - * all the time. - */ - writel(0xFFFFFFFFU & ~(PCI_ERR_BADACK | PCI_ERR_DTRTO | - PCI_ERR_OTHER | PCI_ERR_BIM_DMA_WRITE | - PCI_ERR_BIM_DMA_READ), cp->regs + - REG_PCI_ERR_STATUS_MASK); - - /* set up for MII by default to address mac rx reset timeout - * issue - */ - writel(PCS_DATAPATH_MODE_MII, cp->regs + REG_PCS_DATAPATH_MODE); -} - -static void cas_reset(struct cas *cp, int blkflag) -{ - u32 val; - - cas_mask_intr(cp); - cas_global_reset(cp, blkflag); - cas_mac_reset(cp); - cas_entropy_reset(cp); - - /* disable dma engines. */ - val = readl(cp->regs + REG_TX_CFG); - val &= ~TX_CFG_DMA_EN; - writel(val, cp->regs + REG_TX_CFG); - - val = readl(cp->regs + REG_RX_CFG); - val &= ~RX_CFG_DMA_EN; - writel(val, cp->regs + REG_RX_CFG); - - /* program header parser */ - if ((cp->cas_flags & CAS_FLAG_TARGET_ABORT) || - (&CAS_HP_ALT_FIRMWARE[0] == &cas_prog_null[0])) { - cas_load_firmware(cp, CAS_HP_FIRMWARE); - } else { - cas_load_firmware(cp, CAS_HP_ALT_FIRMWARE); - } - - /* clear out error registers */ - spin_lock(&cp->stat_lock[N_TX_RINGS]); - cas_clear_mac_err(cp); - spin_unlock(&cp->stat_lock[N_TX_RINGS]); -} - -/* Shut down the chip, must be called with pm_mutex held. */ -static void cas_shutdown(struct cas *cp) -{ - unsigned long flags; - - /* Make us not-running to avoid timers respawning */ - cp->hw_running = 0; - - del_timer_sync(&cp->link_timer); - - /* Stop the reset task */ -#if 0 - while (atomic_read(&cp->reset_task_pending_mtu) || - atomic_read(&cp->reset_task_pending_spare) || - atomic_read(&cp->reset_task_pending_all)) - schedule(); - -#else - while (atomic_read(&cp->reset_task_pending)) - schedule(); -#endif - /* Actually stop the chip */ - cas_lock_all_save(cp, flags); - cas_reset(cp, 0); - if (cp->cas_flags & CAS_FLAG_SATURN) - cas_phy_powerdown(cp); - cas_unlock_all_restore(cp, flags); -} - -static int cas_change_mtu(struct net_device *dev, int new_mtu) -{ - struct cas *cp = netdev_priv(dev); - - dev->mtu = new_mtu; - if (!netif_running(dev) || !netif_device_present(dev)) - return 0; - - /* let the reset task handle it */ -#if 1 - atomic_inc(&cp->reset_task_pending); - if ((cp->phy_type & CAS_PHY_SERDES)) { - atomic_inc(&cp->reset_task_pending_all); - } else { - atomic_inc(&cp->reset_task_pending_mtu); - } - schedule_work(&cp->reset_task); -#else - atomic_set(&cp->reset_task_pending, (cp->phy_type & CAS_PHY_SERDES) ? - CAS_RESET_ALL : CAS_RESET_MTU); - pr_err("reset called in cas_change_mtu\n"); - schedule_work(&cp->reset_task); -#endif - - flush_work(&cp->reset_task); - return 0; -} - -static void cas_clean_txd(struct cas *cp, int ring) -{ - struct cas_tx_desc *txd = cp->init_txds[ring]; - struct sk_buff *skb, **skbs = cp->tx_skbs[ring]; - u64 daddr, dlen; - int i, size; - - size = TX_DESC_RINGN_SIZE(ring); - for (i = 0; i < size; i++) { - int frag; - - if (skbs[i] == NULL) - continue; - - skb = skbs[i]; - skbs[i] = NULL; - - for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { - int ent = i & (size - 1); - - /* first buffer is never a tiny buffer and so - * needs to be unmapped. - */ - daddr = le64_to_cpu(txd[ent].buffer); - dlen = CAS_VAL(TX_DESC_BUFLEN, - le64_to_cpu(txd[ent].control)); - dma_unmap_page(&cp->pdev->dev, daddr, dlen, - DMA_TO_DEVICE); - - if (frag != skb_shinfo(skb)->nr_frags) { - i++; - - /* next buffer might by a tiny buffer. - * skip past it. - */ - ent = i & (size - 1); - if (cp->tx_tiny_use[ring][ent].used) - i++; - } - } - dev_kfree_skb_any(skb); - } - - /* zero out tiny buf usage */ - memset(cp->tx_tiny_use[ring], 0, size*sizeof(*cp->tx_tiny_use[ring])); -} - -/* freed on close */ -static inline void cas_free_rx_desc(struct cas *cp, int ring) -{ - cas_page_t **page = cp->rx_pages[ring]; - int i, size; - - size = RX_DESC_RINGN_SIZE(ring); - for (i = 0; i < size; i++) { - if (page[i]) { - cas_page_free(cp, page[i]); - page[i] = NULL; - } - } -} - -static void cas_free_rxds(struct cas *cp) -{ - int i; - - for (i = 0; i < N_RX_DESC_RINGS; i++) - cas_free_rx_desc(cp, i); -} - -/* Must be invoked under cp->lock. */ -static void cas_clean_rings(struct cas *cp) -{ - int i; - - /* need to clean all tx rings */ - memset(cp->tx_old, 0, sizeof(*cp->tx_old)*N_TX_RINGS); - memset(cp->tx_new, 0, sizeof(*cp->tx_new)*N_TX_RINGS); - for (i = 0; i < N_TX_RINGS; i++) - cas_clean_txd(cp, i); - - /* zero out init block */ - memset(cp->init_block, 0, sizeof(struct cas_init_block)); - cas_clean_rxds(cp); - cas_clean_rxcs(cp); -} - -/* allocated on open */ -static inline int cas_alloc_rx_desc(struct cas *cp, int ring) -{ - cas_page_t **page = cp->rx_pages[ring]; - int size, i = 0; - - size = RX_DESC_RINGN_SIZE(ring); - for (i = 0; i < size; i++) { - if ((page[i] = cas_page_alloc(cp, GFP_KERNEL)) == NULL) - return -1; - } - return 0; -} - -static int cas_alloc_rxds(struct cas *cp) -{ - int i; - - for (i = 0; i < N_RX_DESC_RINGS; i++) { - if (cas_alloc_rx_desc(cp, i) < 0) { - cas_free_rxds(cp); - return -1; - } - } - return 0; -} - -static void cas_reset_task(struct work_struct *work) -{ - struct cas *cp = container_of(work, struct cas, reset_task); -#if 0 - int pending = atomic_read(&cp->reset_task_pending); -#else - int pending_all = atomic_read(&cp->reset_task_pending_all); - int pending_spare = atomic_read(&cp->reset_task_pending_spare); - int pending_mtu = atomic_read(&cp->reset_task_pending_mtu); - - if (pending_all == 0 && pending_spare == 0 && pending_mtu == 0) { - /* We can have more tasks scheduled than actually - * needed. - */ - atomic_dec(&cp->reset_task_pending); - return; - } -#endif - /* The link went down, we reset the ring, but keep - * DMA stopped. Use this function for reset - * on error as well. - */ - if (cp->hw_running) { - unsigned long flags; - - /* Make sure we don't get interrupts or tx packets */ - netif_device_detach(cp->dev); - cas_lock_all_save(cp, flags); - - if (cp->opened) { - /* We call cas_spare_recover when we call cas_open. - * but we do not initialize the lists cas_spare_recover - * uses until cas_open is called. - */ - cas_spare_recover(cp, GFP_ATOMIC); - } -#if 1 - /* test => only pending_spare set */ - if (!pending_all && !pending_mtu) - goto done; -#else - if (pending == CAS_RESET_SPARE) - goto done; -#endif - /* when pending == CAS_RESET_ALL, the following - * call to cas_init_hw will restart auto negotiation. - * Setting the second argument of cas_reset to - * !(pending == CAS_RESET_ALL) will set this argument - * to 1 (avoiding reinitializing the PHY for the normal - * PCS case) when auto negotiation is not restarted. - */ -#if 1 - cas_reset(cp, !(pending_all > 0)); - if (cp->opened) - cas_clean_rings(cp); - cas_init_hw(cp, (pending_all > 0)); -#else - cas_reset(cp, !(pending == CAS_RESET_ALL)); - if (cp->opened) - cas_clean_rings(cp); - cas_init_hw(cp, pending == CAS_RESET_ALL); -#endif - -done: - cas_unlock_all_restore(cp, flags); - netif_device_attach(cp->dev); - } -#if 1 - atomic_sub(pending_all, &cp->reset_task_pending_all); - atomic_sub(pending_spare, &cp->reset_task_pending_spare); - atomic_sub(pending_mtu, &cp->reset_task_pending_mtu); - atomic_dec(&cp->reset_task_pending); -#else - atomic_set(&cp->reset_task_pending, 0); -#endif -} - -static void cas_link_timer(struct timer_list *t) -{ - struct cas *cp = from_timer(cp, t, link_timer); - int mask, pending = 0, reset = 0; - unsigned long flags; - - if (link_transition_timeout != 0 && - cp->link_transition_jiffies_valid && - time_is_before_jiffies(cp->link_transition_jiffies + - link_transition_timeout)) { - /* One-second counter so link-down workaround doesn't - * cause resets to occur so fast as to fool the switch - * into thinking the link is down. - */ - cp->link_transition_jiffies_valid = 0; - } - - if (!cp->hw_running) - return; - - spin_lock_irqsave(&cp->lock, flags); - cas_lock_tx(cp); - cas_entropy_gather(cp); - - /* If the link task is still pending, we just - * reschedule the link timer - */ -#if 1 - if (atomic_read(&cp->reset_task_pending_all) || - atomic_read(&cp->reset_task_pending_spare) || - atomic_read(&cp->reset_task_pending_mtu)) - goto done; -#else - if (atomic_read(&cp->reset_task_pending)) - goto done; -#endif - - /* check for rx cleaning */ - if ((mask = (cp->cas_flags & CAS_FLAG_RXD_POST_MASK))) { - int i, rmask; - - for (i = 0; i < MAX_RX_DESC_RINGS; i++) { - rmask = CAS_FLAG_RXD_POST(i); - if ((mask & rmask) == 0) - continue; - - /* post_rxds will do a mod_timer */ - if (cas_post_rxds_ringN(cp, i, cp->rx_last[i]) < 0) { - pending = 1; - continue; - } - cp->cas_flags &= ~rmask; - } - } - - if (CAS_PHY_MII(cp->phy_type)) { - u16 bmsr; - cas_mif_poll(cp, 0); - bmsr = cas_phy_read(cp, MII_BMSR); - /* WTZ: Solaris driver reads this twice, but that - * may be due to the PCS case and the use of a - * common implementation. Read it twice here to be - * safe. - */ - bmsr = cas_phy_read(cp, MII_BMSR); - cas_mif_poll(cp, 1); - readl(cp->regs + REG_MIF_STATUS); /* avoid dups */ - reset = cas_mii_link_check(cp, bmsr); - } else { - reset = cas_pcs_link_check(cp); - } - - if (reset) - goto done; - - /* check for tx state machine confusion */ - if ((readl(cp->regs + REG_MAC_TX_STATUS) & MAC_TX_FRAME_XMIT) == 0) { - u32 val = readl(cp->regs + REG_MAC_STATE_MACHINE); - u32 wptr, rptr; - int tlm = CAS_VAL(MAC_SM_TLM, val); - - if (((tlm == 0x5) || (tlm == 0x3)) && - (CAS_VAL(MAC_SM_ENCAP_SM, val) == 0)) { - netif_printk(cp, tx_err, KERN_DEBUG, cp->dev, - "tx err: MAC_STATE[%08x]\n", val); - reset = 1; - goto done; - } - - val = readl(cp->regs + REG_TX_FIFO_PKT_CNT); - wptr = readl(cp->regs + REG_TX_FIFO_WRITE_PTR); - rptr = readl(cp->regs + REG_TX_FIFO_READ_PTR); - if ((val == 0) && (wptr != rptr)) { - netif_printk(cp, tx_err, KERN_DEBUG, cp->dev, - "tx err: TX_FIFO[%08x:%08x:%08x]\n", - val, wptr, rptr); - reset = 1; - } - - if (reset) - cas_hard_reset(cp); - } - -done: - if (reset) { -#if 1 - atomic_inc(&cp->reset_task_pending); - atomic_inc(&cp->reset_task_pending_all); - schedule_work(&cp->reset_task); -#else - atomic_set(&cp->reset_task_pending, CAS_RESET_ALL); - pr_err("reset called in cas_link_timer\n"); - schedule_work(&cp->reset_task); -#endif - } - - if (!pending) - mod_timer(&cp->link_timer, jiffies + CAS_LINK_TIMEOUT); - cas_unlock_tx(cp); - spin_unlock_irqrestore(&cp->lock, flags); -} - -/* tiny buffers are used to avoid target abort issues with - * older cassini's - */ -static void cas_tx_tiny_free(struct cas *cp) -{ - struct pci_dev *pdev = cp->pdev; - int i; - - for (i = 0; i < N_TX_RINGS; i++) { - if (!cp->tx_tiny_bufs[i]) - continue; - - dma_free_coherent(&pdev->dev, TX_TINY_BUF_BLOCK, - cp->tx_tiny_bufs[i], cp->tx_tiny_dvma[i]); - cp->tx_tiny_bufs[i] = NULL; - } -} - -static int cas_tx_tiny_alloc(struct cas *cp) -{ - struct pci_dev *pdev = cp->pdev; - int i; - - for (i = 0; i < N_TX_RINGS; i++) { - cp->tx_tiny_bufs[i] = - dma_alloc_coherent(&pdev->dev, TX_TINY_BUF_BLOCK, - &cp->tx_tiny_dvma[i], GFP_KERNEL); - if (!cp->tx_tiny_bufs[i]) { - cas_tx_tiny_free(cp); - return -1; - } - } - return 0; -} - - -static int cas_open(struct net_device *dev) -{ - struct cas *cp = netdev_priv(dev); - int hw_was_up, err; - unsigned long flags; - - mutex_lock(&cp->pm_mutex); - - hw_was_up = cp->hw_running; - - /* The power-management mutex protects the hw_running - * etc. state so it is safe to do this bit without cp->lock - */ - if (!cp->hw_running) { - /* Reset the chip */ - cas_lock_all_save(cp, flags); - /* We set the second arg to cas_reset to zero - * because cas_init_hw below will have its second - * argument set to non-zero, which will force - * autonegotiation to start. - */ - cas_reset(cp, 0); - cp->hw_running = 1; - cas_unlock_all_restore(cp, flags); - } - - err = -ENOMEM; - if (cas_tx_tiny_alloc(cp) < 0) - goto err_unlock; - - /* alloc rx descriptors */ - if (cas_alloc_rxds(cp) < 0) - goto err_tx_tiny; - - /* allocate spares */ - cas_spare_init(cp); - cas_spare_recover(cp, GFP_KERNEL); - - /* We can now request the interrupt as we know it's masked - * on the controller. cassini+ has up to 4 interrupts - * that can be used, but you need to do explicit pci interrupt - * mapping to expose them - */ - if (request_irq(cp->pdev->irq, cas_interrupt, - IRQF_SHARED, dev->name, (void *) dev)) { - netdev_err(cp->dev, "failed to request irq !\n"); - err = -EAGAIN; - goto err_spare; - } - -#ifdef USE_NAPI - napi_enable(&cp->napi); -#endif - /* init hw */ - cas_lock_all_save(cp, flags); - cas_clean_rings(cp); - cas_init_hw(cp, !hw_was_up); - cp->opened = 1; - cas_unlock_all_restore(cp, flags); - - netif_start_queue(dev); - mutex_unlock(&cp->pm_mutex); - return 0; - -err_spare: - cas_spare_free(cp); - cas_free_rxds(cp); -err_tx_tiny: - cas_tx_tiny_free(cp); -err_unlock: - mutex_unlock(&cp->pm_mutex); - return err; -} - -static int cas_close(struct net_device *dev) -{ - unsigned long flags; - struct cas *cp = netdev_priv(dev); - -#ifdef USE_NAPI - napi_disable(&cp->napi); -#endif - /* Make sure we don't get distracted by suspend/resume */ - mutex_lock(&cp->pm_mutex); - - netif_stop_queue(dev); - - /* Stop traffic, mark us closed */ - cas_lock_all_save(cp, flags); - cp->opened = 0; - cas_reset(cp, 0); - cas_phy_init(cp); - cas_begin_auto_negotiation(cp, NULL); - cas_clean_rings(cp); - cas_unlock_all_restore(cp, flags); - - free_irq(cp->pdev->irq, (void *) dev); - cas_spare_free(cp); - cas_free_rxds(cp); - cas_tx_tiny_free(cp); - mutex_unlock(&cp->pm_mutex); - return 0; -} - -static struct { - const char name[ETH_GSTRING_LEN]; -} ethtool_cassini_statnames[] = { - {"collisions"}, - {"rx_bytes"}, - {"rx_crc_errors"}, - {"rx_dropped"}, - {"rx_errors"}, - {"rx_fifo_errors"}, - {"rx_frame_errors"}, - {"rx_length_errors"}, - {"rx_over_errors"}, - {"rx_packets"}, - {"tx_aborted_errors"}, - {"tx_bytes"}, - {"tx_dropped"}, - {"tx_errors"}, - {"tx_fifo_errors"}, - {"tx_packets"} -}; -#define CAS_NUM_STAT_KEYS ARRAY_SIZE(ethtool_cassini_statnames) - -static struct { - const int offsets; /* neg. values for 2nd arg to cas_read_phy */ -} ethtool_register_table[] = { - {-MII_BMSR}, - {-MII_BMCR}, - {REG_CAWR}, - {REG_INF_BURST}, - {REG_BIM_CFG}, - {REG_RX_CFG}, - {REG_HP_CFG}, - {REG_MAC_TX_CFG}, - {REG_MAC_RX_CFG}, - {REG_MAC_CTRL_CFG}, - {REG_MAC_XIF_CFG}, - {REG_MIF_CFG}, - {REG_PCS_CFG}, - {REG_SATURN_PCFG}, - {REG_PCS_MII_STATUS}, - {REG_PCS_STATE_MACHINE}, - {REG_MAC_COLL_EXCESS}, - {REG_MAC_COLL_LATE} -}; -#define CAS_REG_LEN ARRAY_SIZE(ethtool_register_table) -#define CAS_MAX_REGS (sizeof (u32)*CAS_REG_LEN) - -static void cas_read_regs(struct cas *cp, u8 *ptr, int len) -{ - u8 *p; - int i; - unsigned long flags; - - spin_lock_irqsave(&cp->lock, flags); - for (i = 0, p = ptr; i < len ; i ++, p += sizeof(u32)) { - u16 hval; - u32 val; - if (ethtool_register_table[i].offsets < 0) { - hval = cas_phy_read(cp, - -ethtool_register_table[i].offsets); - val = hval; - } else { - val= readl(cp->regs+ethtool_register_table[i].offsets); - } - memcpy(p, (u8 *)&val, sizeof(u32)); - } - spin_unlock_irqrestore(&cp->lock, flags); -} - -static struct net_device_stats *cas_get_stats(struct net_device *dev) -{ - struct cas *cp = netdev_priv(dev); - struct net_device_stats *stats = cp->net_stats; - unsigned long flags; - int i; - unsigned long tmp; - - /* we collate all of the stats into net_stats[N_TX_RING] */ - if (!cp->hw_running) - return stats + N_TX_RINGS; - - /* collect outstanding stats */ - /* WTZ: the Cassini spec gives these as 16 bit counters but - * stored in 32-bit words. Added a mask of 0xffff to be safe, - * in case the chip somehow puts any garbage in the other bits. - * Also, counter usage didn't seem to mach what Adrian did - * in the parts of the code that set these quantities. Made - * that consistent. - */ - spin_lock_irqsave(&cp->stat_lock[N_TX_RINGS], flags); - stats[N_TX_RINGS].rx_crc_errors += - readl(cp->regs + REG_MAC_FCS_ERR) & 0xffff; - stats[N_TX_RINGS].rx_frame_errors += - readl(cp->regs + REG_MAC_ALIGN_ERR) &0xffff; - stats[N_TX_RINGS].rx_length_errors += - readl(cp->regs + REG_MAC_LEN_ERR) & 0xffff; -#if 1 - tmp = (readl(cp->regs + REG_MAC_COLL_EXCESS) & 0xffff) + - (readl(cp->regs + REG_MAC_COLL_LATE) & 0xffff); - stats[N_TX_RINGS].tx_aborted_errors += tmp; - stats[N_TX_RINGS].collisions += - tmp + (readl(cp->regs + REG_MAC_COLL_NORMAL) & 0xffff); -#else - stats[N_TX_RINGS].tx_aborted_errors += - readl(cp->regs + REG_MAC_COLL_EXCESS); - stats[N_TX_RINGS].collisions += readl(cp->regs + REG_MAC_COLL_EXCESS) + - readl(cp->regs + REG_MAC_COLL_LATE); -#endif - cas_clear_mac_err(cp); - - /* saved bits that are unique to ring 0 */ - spin_lock(&cp->stat_lock[0]); - stats[N_TX_RINGS].collisions += stats[0].collisions; - stats[N_TX_RINGS].rx_over_errors += stats[0].rx_over_errors; - stats[N_TX_RINGS].rx_frame_errors += stats[0].rx_frame_errors; - stats[N_TX_RINGS].rx_fifo_errors += stats[0].rx_fifo_errors; - stats[N_TX_RINGS].tx_aborted_errors += stats[0].tx_aborted_errors; - stats[N_TX_RINGS].tx_fifo_errors += stats[0].tx_fifo_errors; - spin_unlock(&cp->stat_lock[0]); - - for (i = 0; i < N_TX_RINGS; i++) { - spin_lock(&cp->stat_lock[i]); - stats[N_TX_RINGS].rx_length_errors += - stats[i].rx_length_errors; - stats[N_TX_RINGS].rx_crc_errors += stats[i].rx_crc_errors; - stats[N_TX_RINGS].rx_packets += stats[i].rx_packets; - stats[N_TX_RINGS].tx_packets += stats[i].tx_packets; - stats[N_TX_RINGS].rx_bytes += stats[i].rx_bytes; - stats[N_TX_RINGS].tx_bytes += stats[i].tx_bytes; - stats[N_TX_RINGS].rx_errors += stats[i].rx_errors; - stats[N_TX_RINGS].tx_errors += stats[i].tx_errors; - stats[N_TX_RINGS].rx_dropped += stats[i].rx_dropped; - stats[N_TX_RINGS].tx_dropped += stats[i].tx_dropped; - memset(stats + i, 0, sizeof(struct net_device_stats)); - spin_unlock(&cp->stat_lock[i]); - } - spin_unlock_irqrestore(&cp->stat_lock[N_TX_RINGS], flags); - return stats + N_TX_RINGS; -} - - -static void cas_set_multicast(struct net_device *dev) -{ - struct cas *cp = netdev_priv(dev); - u32 rxcfg, rxcfg_new; - unsigned long flags; - int limit = STOP_TRIES; - - if (!cp->hw_running) - return; - - spin_lock_irqsave(&cp->lock, flags); - rxcfg = readl(cp->regs + REG_MAC_RX_CFG); - - /* disable RX MAC and wait for completion */ - writel(rxcfg & ~MAC_RX_CFG_EN, cp->regs + REG_MAC_RX_CFG); - while (readl(cp->regs + REG_MAC_RX_CFG) & MAC_RX_CFG_EN) { - if (!limit--) - break; - udelay(10); - } - - /* disable hash filter and wait for completion */ - limit = STOP_TRIES; - rxcfg &= ~(MAC_RX_CFG_PROMISC_EN | MAC_RX_CFG_HASH_FILTER_EN); - writel(rxcfg & ~MAC_RX_CFG_EN, cp->regs + REG_MAC_RX_CFG); - while (readl(cp->regs + REG_MAC_RX_CFG) & MAC_RX_CFG_HASH_FILTER_EN) { - if (!limit--) - break; - udelay(10); - } - - /* program hash filters */ - cp->mac_rx_cfg = rxcfg_new = cas_setup_multicast(cp); - rxcfg |= rxcfg_new; - writel(rxcfg, cp->regs + REG_MAC_RX_CFG); - spin_unlock_irqrestore(&cp->lock, flags); -} - -static void cas_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info) -{ - struct cas *cp = netdev_priv(dev); - strscpy(info->driver, DRV_MODULE_NAME, sizeof(info->driver)); - strscpy(info->version, DRV_MODULE_VERSION, sizeof(info->version)); - strscpy(info->bus_info, pci_name(cp->pdev), sizeof(info->bus_info)); -} - -static int cas_get_link_ksettings(struct net_device *dev, - struct ethtool_link_ksettings *cmd) -{ - struct cas *cp = netdev_priv(dev); - u16 bmcr; - int full_duplex, speed, pause; - unsigned long flags; - enum link_state linkstate = link_up; - u32 supported, advertising; - - advertising = 0; - supported = SUPPORTED_Autoneg; - if (cp->cas_flags & CAS_FLAG_1000MB_CAP) { - supported |= SUPPORTED_1000baseT_Full; - advertising |= ADVERTISED_1000baseT_Full; - } - - /* Record PHY settings if HW is on. */ - spin_lock_irqsave(&cp->lock, flags); - bmcr = 0; - linkstate = cp->lstate; - if (CAS_PHY_MII(cp->phy_type)) { - cmd->base.port = PORT_MII; - cmd->base.phy_address = cp->phy_addr; - advertising |= ADVERTISED_TP | ADVERTISED_MII | - ADVERTISED_10baseT_Half | - ADVERTISED_10baseT_Full | - ADVERTISED_100baseT_Half | - ADVERTISED_100baseT_Full; - - supported |= - (SUPPORTED_10baseT_Half | - SUPPORTED_10baseT_Full | - SUPPORTED_100baseT_Half | - SUPPORTED_100baseT_Full | - SUPPORTED_TP | SUPPORTED_MII); - - if (cp->hw_running) { - cas_mif_poll(cp, 0); - bmcr = cas_phy_read(cp, MII_BMCR); - cas_read_mii_link_mode(cp, &full_duplex, - &speed, &pause); - cas_mif_poll(cp, 1); - } - - } else { - cmd->base.port = PORT_FIBRE; - cmd->base.phy_address = 0; - supported |= SUPPORTED_FIBRE; - advertising |= ADVERTISED_FIBRE; - - if (cp->hw_running) { - /* pcs uses the same bits as mii */ - bmcr = readl(cp->regs + REG_PCS_MII_CTRL); - cas_read_pcs_link_mode(cp, &full_duplex, - &speed, &pause); - } - } - spin_unlock_irqrestore(&cp->lock, flags); - - if (bmcr & BMCR_ANENABLE) { - advertising |= ADVERTISED_Autoneg; - cmd->base.autoneg = AUTONEG_ENABLE; - cmd->base.speed = ((speed == 10) ? - SPEED_10 : - ((speed == 1000) ? - SPEED_1000 : SPEED_100)); - cmd->base.duplex = full_duplex ? DUPLEX_FULL : DUPLEX_HALF; - } else { - cmd->base.autoneg = AUTONEG_DISABLE; - cmd->base.speed = ((bmcr & CAS_BMCR_SPEED1000) ? - SPEED_1000 : - ((bmcr & BMCR_SPEED100) ? - SPEED_100 : SPEED_10)); - cmd->base.duplex = (bmcr & BMCR_FULLDPLX) ? - DUPLEX_FULL : DUPLEX_HALF; - } - if (linkstate != link_up) { - /* Force these to "unknown" if the link is not up and - * autonogotiation in enabled. We can set the link - * speed to 0, but not cmd->duplex, - * because its legal values are 0 and 1. Ethtool will - * print the value reported in parentheses after the - * word "Unknown" for unrecognized values. - * - * If in forced mode, we report the speed and duplex - * settings that we configured. - */ - if (cp->link_cntl & BMCR_ANENABLE) { - cmd->base.speed = 0; - cmd->base.duplex = 0xff; - } else { - cmd->base.speed = SPEED_10; - if (cp->link_cntl & BMCR_SPEED100) { - cmd->base.speed = SPEED_100; - } else if (cp->link_cntl & CAS_BMCR_SPEED1000) { - cmd->base.speed = SPEED_1000; - } - cmd->base.duplex = (cp->link_cntl & BMCR_FULLDPLX) ? - DUPLEX_FULL : DUPLEX_HALF; - } - } - - ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported, - supported); - ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising, - advertising); - - return 0; -} - -static int cas_set_link_ksettings(struct net_device *dev, - const struct ethtool_link_ksettings *cmd) -{ - struct cas *cp = netdev_priv(dev); - unsigned long flags; - u32 speed = cmd->base.speed; - - /* Verify the settings we care about. */ - if (cmd->base.autoneg != AUTONEG_ENABLE && - cmd->base.autoneg != AUTONEG_DISABLE) - return -EINVAL; - - if (cmd->base.autoneg == AUTONEG_DISABLE && - ((speed != SPEED_1000 && - speed != SPEED_100 && - speed != SPEED_10) || - (cmd->base.duplex != DUPLEX_HALF && - cmd->base.duplex != DUPLEX_FULL))) - return -EINVAL; - - /* Apply settings and restart link process. */ - spin_lock_irqsave(&cp->lock, flags); - cas_begin_auto_negotiation(cp, cmd); - spin_unlock_irqrestore(&cp->lock, flags); - return 0; -} - -static int cas_nway_reset(struct net_device *dev) -{ - struct cas *cp = netdev_priv(dev); - unsigned long flags; - - if ((cp->link_cntl & BMCR_ANENABLE) == 0) - return -EINVAL; - - /* Restart link process. */ - spin_lock_irqsave(&cp->lock, flags); - cas_begin_auto_negotiation(cp, NULL); - spin_unlock_irqrestore(&cp->lock, flags); - - return 0; -} - -static u32 cas_get_link(struct net_device *dev) -{ - struct cas *cp = netdev_priv(dev); - return cp->lstate == link_up; -} - -static u32 cas_get_msglevel(struct net_device *dev) -{ - struct cas *cp = netdev_priv(dev); - return cp->msg_enable; -} - -static void cas_set_msglevel(struct net_device *dev, u32 value) -{ - struct cas *cp = netdev_priv(dev); - cp->msg_enable = value; -} - -static int cas_get_regs_len(struct net_device *dev) -{ - struct cas *cp = netdev_priv(dev); - return min_t(int, cp->casreg_len, CAS_MAX_REGS); -} - -static void cas_get_regs(struct net_device *dev, struct ethtool_regs *regs, - void *p) -{ - struct cas *cp = netdev_priv(dev); - regs->version = 0; - /* cas_read_regs handles locks (cp->lock). */ - cas_read_regs(cp, p, regs->len / sizeof(u32)); -} - -static int cas_get_sset_count(struct net_device *dev, int sset) -{ - switch (sset) { - case ETH_SS_STATS: - return CAS_NUM_STAT_KEYS; - default: - return -EOPNOTSUPP; - } -} - -static void cas_get_strings(struct net_device *dev, u32 stringset, u8 *data) -{ - memcpy(data, ðtool_cassini_statnames, - CAS_NUM_STAT_KEYS * ETH_GSTRING_LEN); -} - -static void cas_get_ethtool_stats(struct net_device *dev, - struct ethtool_stats *estats, u64 *data) -{ - struct cas *cp = netdev_priv(dev); - struct net_device_stats *stats = cas_get_stats(cp->dev); - int i = 0; - data[i++] = stats->collisions; - data[i++] = stats->rx_bytes; - data[i++] = stats->rx_crc_errors; - data[i++] = stats->rx_dropped; - data[i++] = stats->rx_errors; - data[i++] = stats->rx_fifo_errors; - data[i++] = stats->rx_frame_errors; - data[i++] = stats->rx_length_errors; - data[i++] = stats->rx_over_errors; - data[i++] = stats->rx_packets; - data[i++] = stats->tx_aborted_errors; - data[i++] = stats->tx_bytes; - data[i++] = stats->tx_dropped; - data[i++] = stats->tx_errors; - data[i++] = stats->tx_fifo_errors; - data[i++] = stats->tx_packets; - BUG_ON(i != CAS_NUM_STAT_KEYS); -} - -static const struct ethtool_ops cas_ethtool_ops = { - .get_drvinfo = cas_get_drvinfo, - .nway_reset = cas_nway_reset, - .get_link = cas_get_link, - .get_msglevel = cas_get_msglevel, - .set_msglevel = cas_set_msglevel, - .get_regs_len = cas_get_regs_len, - .get_regs = cas_get_regs, - .get_sset_count = cas_get_sset_count, - .get_strings = cas_get_strings, - .get_ethtool_stats = cas_get_ethtool_stats, - .get_link_ksettings = cas_get_link_ksettings, - .set_link_ksettings = cas_set_link_ksettings, -}; - -static int cas_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd) -{ - struct cas *cp = netdev_priv(dev); - struct mii_ioctl_data *data = if_mii(ifr); - unsigned long flags; - int rc = -EOPNOTSUPP; - - /* Hold the PM mutex while doing ioctl's or we may collide - * with open/close and power management and oops. - */ - mutex_lock(&cp->pm_mutex); - switch (cmd) { - case SIOCGMIIPHY: /* Get address of MII PHY in use. */ - data->phy_id = cp->phy_addr; - fallthrough; - - case SIOCGMIIREG: /* Read MII PHY register. */ - spin_lock_irqsave(&cp->lock, flags); - cas_mif_poll(cp, 0); - data->val_out = cas_phy_read(cp, data->reg_num & 0x1f); - cas_mif_poll(cp, 1); - spin_unlock_irqrestore(&cp->lock, flags); - rc = 0; - break; - - case SIOCSMIIREG: /* Write MII PHY register. */ - spin_lock_irqsave(&cp->lock, flags); - cas_mif_poll(cp, 0); - rc = cas_phy_write(cp, data->reg_num & 0x1f, data->val_in); - cas_mif_poll(cp, 1); - spin_unlock_irqrestore(&cp->lock, flags); - break; - default: - break; - } - - mutex_unlock(&cp->pm_mutex); - return rc; -} - -/* When this chip sits underneath an Intel 31154 bridge, it is the - * only subordinate device and we can tweak the bridge settings to - * reflect that fact. - */ -static void cas_program_bridge(struct pci_dev *cas_pdev) -{ - struct pci_dev *pdev = cas_pdev->bus->self; - u32 val; - - if (!pdev) - return; - - if (pdev->vendor != 0x8086 || pdev->device != 0x537c) - return; - - /* Clear bit 10 (Bus Parking Control) in the Secondary - * Arbiter Control/Status Register which lives at offset - * 0x41. Using a 32-bit word read/modify/write at 0x40 - * is much simpler so that's how we do this. - */ - pci_read_config_dword(pdev, 0x40, &val); - val &= ~0x00040000; - pci_write_config_dword(pdev, 0x40, val); - - /* Max out the Multi-Transaction Timer settings since - * Cassini is the only device present. - * - * The register is 16-bit and lives at 0x50. When the - * settings are enabled, it extends the GRANT# signal - * for a requestor after a transaction is complete. This - * allows the next request to run without first needing - * to negotiate the GRANT# signal back. - * - * Bits 12:10 define the grant duration: - * - * 1 -- 16 clocks - * 2 -- 32 clocks - * 3 -- 64 clocks - * 4 -- 128 clocks - * 5 -- 256 clocks - * - * All other values are illegal. - * - * Bits 09:00 define which REQ/GNT signal pairs get the - * GRANT# signal treatment. We set them all. - */ - pci_write_config_word(pdev, 0x50, (5 << 10) | 0x3ff); - - /* The Read Prefecth Policy register is 16-bit and sits at - * offset 0x52. It enables a "smart" pre-fetch policy. We - * enable it and max out all of the settings since only one - * device is sitting underneath and thus bandwidth sharing is - * not an issue. - * - * The register has several 3 bit fields, which indicates a - * multiplier applied to the base amount of prefetching the - * chip would do. These fields are at: - * - * 15:13 --- ReRead Primary Bus - * 12:10 --- FirstRead Primary Bus - * 09:07 --- ReRead Secondary Bus - * 06:04 --- FirstRead Secondary Bus - * - * Bits 03:00 control which REQ/GNT pairs the prefetch settings - * get enabled on. Bit 3 is a grouped enabler which controls - * all of the REQ/GNT pairs from [8:3]. Bits 2 to 0 control - * the individual REQ/GNT pairs [2:0]. - */ - pci_write_config_word(pdev, 0x52, - (0x7 << 13) | - (0x7 << 10) | - (0x7 << 7) | - (0x7 << 4) | - (0xf << 0)); - - /* Force cacheline size to 0x8 */ - pci_write_config_byte(pdev, PCI_CACHE_LINE_SIZE, 0x08); - - /* Force latency timer to maximum setting so Cassini can - * sit on the bus as long as it likes. - */ - pci_write_config_byte(pdev, PCI_LATENCY_TIMER, 0xff); -} - -static const struct net_device_ops cas_netdev_ops = { - .ndo_open = cas_open, - .ndo_stop = cas_close, - .ndo_start_xmit = cas_start_xmit, - .ndo_get_stats = cas_get_stats, - .ndo_set_rx_mode = cas_set_multicast, - .ndo_eth_ioctl = cas_ioctl, - .ndo_tx_timeout = cas_tx_timeout, - .ndo_change_mtu = cas_change_mtu, - .ndo_set_mac_address = eth_mac_addr, - .ndo_validate_addr = eth_validate_addr, -#ifdef CONFIG_NET_POLL_CONTROLLER - .ndo_poll_controller = cas_netpoll, -#endif -}; - -static int cas_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) -{ - static int cas_version_printed = 0; - unsigned long casreg_len; - struct net_device *dev; - struct cas *cp; - u16 pci_cmd; - int i, err; - u8 orig_cacheline_size = 0, cas_cacheline_size = 0; - - if (cas_version_printed++ == 0) - pr_info("%s", version); - - err = pci_enable_device(pdev); - if (err) { - dev_err(&pdev->dev, "Cannot enable PCI device, aborting\n"); - return err; - } - - if (!(pci_resource_flags(pdev, 0) & IORESOURCE_MEM)) { - dev_err(&pdev->dev, "Cannot find proper PCI device " - "base address, aborting\n"); - err = -ENODEV; - goto err_out_disable_pdev; - } - - dev = alloc_etherdev(sizeof(*cp)); - if (!dev) { - err = -ENOMEM; - goto err_out_disable_pdev; - } - SET_NETDEV_DEV(dev, &pdev->dev); - - err = pci_request_regions(pdev, dev->name); - if (err) { - dev_err(&pdev->dev, "Cannot obtain PCI resources, aborting\n"); - goto err_out_free_netdev; - } - pci_set_master(pdev); - - /* we must always turn on parity response or else parity - * doesn't get generated properly. disable SERR/PERR as well. - * in addition, we want to turn MWI on. - */ - pci_read_config_word(pdev, PCI_COMMAND, &pci_cmd); - pci_cmd &= ~PCI_COMMAND_SERR; - pci_cmd |= PCI_COMMAND_PARITY; - pci_write_config_word(pdev, PCI_COMMAND, pci_cmd); - if (pci_try_set_mwi(pdev)) - pr_warn("Could not enable MWI for %s\n", pci_name(pdev)); - - cas_program_bridge(pdev); - - /* - * On some architectures, the default cache line size set - * by pci_try_set_mwi reduces perforamnce. We have to increase - * it for this case. To start, we'll print some configuration - * data. - */ -#if 1 - pci_read_config_byte(pdev, PCI_CACHE_LINE_SIZE, - &orig_cacheline_size); - if (orig_cacheline_size < CAS_PREF_CACHELINE_SIZE) { - cas_cacheline_size = - (CAS_PREF_CACHELINE_SIZE < SMP_CACHE_BYTES) ? - CAS_PREF_CACHELINE_SIZE : SMP_CACHE_BYTES; - if (pci_write_config_byte(pdev, - PCI_CACHE_LINE_SIZE, - cas_cacheline_size)) { - dev_err(&pdev->dev, "Could not set PCI cache " - "line size\n"); - goto err_out_free_res; - } - } -#endif - - - /* Configure DMA attributes. */ - err = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64)); - if (err) { - dev_err(&pdev->dev, "No usable DMA configuration, aborting\n"); - goto err_out_free_res; - } - - casreg_len = pci_resource_len(pdev, 0); - - cp = netdev_priv(dev); - cp->pdev = pdev; -#if 1 - /* A value of 0 indicates we never explicitly set it */ - cp->orig_cacheline_size = cas_cacheline_size ? orig_cacheline_size: 0; -#endif - cp->dev = dev; - cp->msg_enable = (cassini_debug < 0) ? CAS_DEF_MSG_ENABLE : - cassini_debug; - -#if defined(CONFIG_SPARC) - cp->of_node = pci_device_to_OF_node(pdev); -#endif - - cp->link_transition = LINK_TRANSITION_UNKNOWN; - cp->link_transition_jiffies_valid = 0; - - spin_lock_init(&cp->lock); - spin_lock_init(&cp->rx_inuse_lock); - spin_lock_init(&cp->rx_spare_lock); - for (i = 0; i < N_TX_RINGS; i++) { - spin_lock_init(&cp->stat_lock[i]); - spin_lock_init(&cp->tx_lock[i]); - } - spin_lock_init(&cp->stat_lock[N_TX_RINGS]); - mutex_init(&cp->pm_mutex); - - timer_setup(&cp->link_timer, cas_link_timer, 0); - -#if 1 - /* Just in case the implementation of atomic operations - * change so that an explicit initialization is necessary. - */ - atomic_set(&cp->reset_task_pending, 0); - atomic_set(&cp->reset_task_pending_all, 0); - atomic_set(&cp->reset_task_pending_spare, 0); - atomic_set(&cp->reset_task_pending_mtu, 0); -#endif - INIT_WORK(&cp->reset_task, cas_reset_task); - - /* Default link parameters */ - if (link_mode >= 0 && link_mode < 6) - cp->link_cntl = link_modes[link_mode]; - else - cp->link_cntl = BMCR_ANENABLE; - cp->lstate = link_down; - cp->link_transition = LINK_TRANSITION_LINK_DOWN; - netif_carrier_off(cp->dev); - cp->timer_ticks = 0; - - /* give us access to cassini registers */ - cp->regs = pci_iomap(pdev, 0, casreg_len); - if (!cp->regs) { - dev_err(&pdev->dev, "Cannot map device registers, aborting\n"); - goto err_out_free_res; - } - cp->casreg_len = casreg_len; - - pci_save_state(pdev); - cas_check_pci_invariants(cp); - cas_hard_reset(cp); - cas_reset(cp, 0); - if (cas_check_invariants(cp)) - goto err_out_iounmap; - if (cp->cas_flags & CAS_FLAG_SATURN) - cas_saturn_firmware_init(cp); - - cp->init_block = - dma_alloc_coherent(&pdev->dev, sizeof(struct cas_init_block), - &cp->block_dvma, GFP_KERNEL); - if (!cp->init_block) { - dev_err(&pdev->dev, "Cannot allocate init block, aborting\n"); - goto err_out_iounmap; - } - - for (i = 0; i < N_TX_RINGS; i++) - cp->init_txds[i] = cp->init_block->txds[i]; - - for (i = 0; i < N_RX_DESC_RINGS; i++) - cp->init_rxds[i] = cp->init_block->rxds[i]; - - for (i = 0; i < N_RX_COMP_RINGS; i++) - cp->init_rxcs[i] = cp->init_block->rxcs[i]; - - for (i = 0; i < N_RX_FLOWS; i++) - skb_queue_head_init(&cp->rx_flows[i]); - - dev->netdev_ops = &cas_netdev_ops; - dev->ethtool_ops = &cas_ethtool_ops; - dev->watchdog_timeo = CAS_TX_TIMEOUT; - -#ifdef USE_NAPI - netif_napi_add(dev, &cp->napi, cas_poll); -#endif - dev->irq = pdev->irq; - dev->dma = 0; - - /* Cassini features. */ - if ((cp->cas_flags & CAS_FLAG_NO_HW_CSUM) == 0) - dev->features |= NETIF_F_HW_CSUM | NETIF_F_SG; - - dev->features |= NETIF_F_HIGHDMA; - - /* MTU range: 60 - varies or 9000 */ - dev->min_mtu = CAS_MIN_MTU; - dev->max_mtu = CAS_MAX_MTU; - - if (register_netdev(dev)) { - dev_err(&pdev->dev, "Cannot register net device, aborting\n"); - goto err_out_free_consistent; - } - - i = readl(cp->regs + REG_BIM_CFG); - netdev_info(dev, "Sun Cassini%s (%sbit/%sMHz PCI/%s) Ethernet[%d] %pM\n", - (cp->cas_flags & CAS_FLAG_REG_PLUS) ? "+" : "", - (i & BIM_CFG_32BIT) ? "32" : "64", - (i & BIM_CFG_66MHZ) ? "66" : "33", - (cp->phy_type == CAS_PHY_SERDES) ? "Fi" : "Cu", pdev->irq, - dev->dev_addr); - - pci_set_drvdata(pdev, dev); - cp->hw_running = 1; - cas_entropy_reset(cp); - cas_phy_init(cp); - cas_begin_auto_negotiation(cp, NULL); - return 0; - -err_out_free_consistent: - dma_free_coherent(&pdev->dev, sizeof(struct cas_init_block), - cp->init_block, cp->block_dvma); - -err_out_iounmap: - mutex_lock(&cp->pm_mutex); - if (cp->hw_running) - cas_shutdown(cp); - mutex_unlock(&cp->pm_mutex); - - pci_iounmap(pdev, cp->regs); - - -err_out_free_res: - pci_release_regions(pdev); - - /* Try to restore it in case the error occurred after we - * set it. - */ - pci_write_config_byte(pdev, PCI_CACHE_LINE_SIZE, orig_cacheline_size); - -err_out_free_netdev: - free_netdev(dev); - -err_out_disable_pdev: - pci_disable_device(pdev); - return -ENODEV; -} - -static void cas_remove_one(struct pci_dev *pdev) -{ - struct net_device *dev = pci_get_drvdata(pdev); - struct cas *cp; - if (!dev) - return; - - cp = netdev_priv(dev); - unregister_netdev(dev); - - vfree(cp->fw_data); - - mutex_lock(&cp->pm_mutex); - cancel_work_sync(&cp->reset_task); - if (cp->hw_running) - cas_shutdown(cp); - mutex_unlock(&cp->pm_mutex); - -#if 1 - if (cp->orig_cacheline_size) { - /* Restore the cache line size if we had modified - * it. - */ - pci_write_config_byte(pdev, PCI_CACHE_LINE_SIZE, - cp->orig_cacheline_size); - } -#endif - dma_free_coherent(&pdev->dev, sizeof(struct cas_init_block), - cp->init_block, cp->block_dvma); - pci_iounmap(pdev, cp->regs); - free_netdev(dev); - pci_release_regions(pdev); - pci_disable_device(pdev); -} - -static int __maybe_unused cas_suspend(struct device *dev_d) -{ - struct net_device *dev = dev_get_drvdata(dev_d); - struct cas *cp = netdev_priv(dev); - unsigned long flags; - - mutex_lock(&cp->pm_mutex); - - /* If the driver is opened, we stop the DMA */ - if (cp->opened) { - netif_device_detach(dev); - - cas_lock_all_save(cp, flags); - - /* We can set the second arg of cas_reset to 0 - * because on resume, we'll call cas_init_hw with - * its second arg set so that autonegotiation is - * restarted. - */ - cas_reset(cp, 0); - cas_clean_rings(cp); - cas_unlock_all_restore(cp, flags); - } - - if (cp->hw_running) - cas_shutdown(cp); - mutex_unlock(&cp->pm_mutex); - - return 0; -} - -static int __maybe_unused cas_resume(struct device *dev_d) -{ - struct net_device *dev = dev_get_drvdata(dev_d); - struct cas *cp = netdev_priv(dev); - - netdev_info(dev, "resuming\n"); - - mutex_lock(&cp->pm_mutex); - cas_hard_reset(cp); - if (cp->opened) { - unsigned long flags; - cas_lock_all_save(cp, flags); - cas_reset(cp, 0); - cp->hw_running = 1; - cas_clean_rings(cp); - cas_init_hw(cp, 1); - cas_unlock_all_restore(cp, flags); - - netif_device_attach(dev); - } - mutex_unlock(&cp->pm_mutex); - return 0; -} - -static SIMPLE_DEV_PM_OPS(cas_pm_ops, cas_suspend, cas_resume); - -static struct pci_driver cas_driver = { - .name = DRV_MODULE_NAME, - .id_table = cas_pci_tbl, - .probe = cas_init_one, - .remove = cas_remove_one, - .driver.pm = &cas_pm_ops, -}; - -static int __init cas_init(void) -{ - if (linkdown_timeout > 0) - link_transition_timeout = linkdown_timeout * HZ; - else - link_transition_timeout = 0; - - return pci_register_driver(&cas_driver); -} - -static void __exit cas_cleanup(void) -{ - pci_unregister_driver(&cas_driver); -} - -module_init(cas_init); -module_exit(cas_cleanup); diff --git a/drivers/net/ethernet/sun/cassini.h b/drivers/net/ethernet/sun/cassini.h deleted file mode 100644 index 2d91f49..00000000 --- a/drivers/net/ethernet/sun/cassini.h +++ /dev/null @@ -1,2900 +0,0 @@ -/* SPDX-License-Identifier: GPL-2.0+ */ -/* $Id: cassini.h,v 1.16 2004/08/17 21:15:16 zaumen Exp $ - * cassini.h: Definitions for Sun Microsystems Cassini(+) ethernet driver. - * - * Copyright (C) 2004 Sun Microsystems Inc. - * Copyright (c) 2003 Adrian Sun (asun@darksunrising.com) - * - * vendor id: 0x108E (Sun Microsystems, Inc.) - * device id: 0xabba (Cassini) - * revision ids: 0x01 = Cassini - * 0x02 = Cassini rev 2 - * 0x10 = Cassini+ - * 0x11 = Cassini+ 0.2u - * - * vendor id: 0x100b (National Semiconductor) - * device id: 0x0035 (DP83065/Saturn) - * revision ids: 0x30 = Saturn B2 - * - * rings are all offset from 0. - * - * there are two clock domains: - * PCI: 33/66MHz clock - * chip: 125MHz clock - */ - -#ifndef _CASSINI_H -#define _CASSINI_H - -/* cassini register map: 2M memory mapped in 32-bit memory space accessible as - * 32-bit words. there is no i/o port access. REG_ addresses are - * shared between cassini and cassini+. REG_PLUS_ addresses only - * appear in cassini+. REG_MINUS_ addresses only appear in cassini. - */ -#define CAS_ID_REV2 0x02 -#define CAS_ID_REVPLUS 0x10 -#define CAS_ID_REVPLUS02u 0x11 -#define CAS_ID_REVSATURNB2 0x30 - -/** global resources **/ - -/* this register sets the weights for the weighted round robin arbiter. e.g., - * if rx weight == 1 and tx weight == 0, rx == 2x tx transfer credit - * for its next turn to access the pci bus. - * map: 0x0 = x1, 0x1 = x2, 0x2 = x4, 0x3 = x8 - * DEFAULT: 0x0, SIZE: 5 bits - */ -#define REG_CAWR 0x0004 /* core arbitration weight */ -#define CAWR_RX_DMA_WEIGHT_SHIFT 0 -#define CAWR_RX_DMA_WEIGHT_MASK 0x03 /* [0:1] */ -#define CAWR_TX_DMA_WEIGHT_SHIFT 2 -#define CAWR_TX_DMA_WEIGHT_MASK 0x0C /* [3:2] */ -#define CAWR_RR_DIS 0x10 /* [4] */ - -/* if enabled, BIM can send bursts across PCI bus > cacheline size. burst - * sizes determined by length of packet or descriptor transfer and the - * max length allowed by the target. - * DEFAULT: 0x0, SIZE: 1 bit - */ -#define REG_INF_BURST 0x0008 /* infinite burst enable reg */ -#define INF_BURST_EN 0x1 /* enable */ - -/* top level interrupts [0-9] are auto-cleared to 0 when the status - * register is read. second level interrupts [13 - 18] are cleared at - * the source. tx completion register 3 is replicated in [19 - 31] - * DEFAULT: 0x00000000, SIZE: 29 bits - */ -#define REG_INTR_STATUS 0x000C /* interrupt status register */ -#define INTR_TX_INTME 0x00000001 /* frame w/ INT ME desc bit set - xferred from host queue to - TX FIFO */ -#define INTR_TX_ALL 0x00000002 /* all xmit frames xferred into - TX FIFO. i.e., - TX Kick == TX complete. if - PACED_MODE set, then TX FIFO - also empty */ -#define INTR_TX_DONE 0x00000004 /* any frame xferred into tx - FIFO */ -#define INTR_TX_TAG_ERROR 0x00000008 /* TX FIFO tag framing - corrupted. FATAL ERROR */ -#define INTR_RX_DONE 0x00000010 /* at least 1 frame xferred - from RX FIFO to host mem. - RX completion reg updated. - may be delayed by recv - intr blanking. */ -#define INTR_RX_BUF_UNAVAIL 0x00000020 /* no more receive buffers. - RX Kick == RX complete */ -#define INTR_RX_TAG_ERROR 0x00000040 /* RX FIFO tag framing - corrupted. FATAL ERROR */ -#define INTR_RX_COMP_FULL 0x00000080 /* no more room in completion - ring to post descriptors. - RX complete head incr to - almost reach RX complete - tail */ -#define INTR_RX_BUF_AE 0x00000100 /* less than the - programmable threshold # - of free descr avail for - hw use */ -#define INTR_RX_COMP_AF 0x00000200 /* less than the - programmable threshold # - of descr spaces for hw - use in completion descr - ring */ -#define INTR_RX_LEN_MISMATCH 0x00000400 /* len field from MAC != - len of non-reassembly pkt - from fifo during DMA or - header parser provides TCP - header and payload size > - MAC packet size. - FATAL ERROR */ -#define INTR_SUMMARY 0x00001000 /* summary interrupt bit. this - bit will be set if an interrupt - generated on the pci bus. useful - when driver is polling for - interrupts */ -#define INTR_PCS_STATUS 0x00002000 /* PCS interrupt status register */ -#define INTR_TX_MAC_STATUS 0x00004000 /* TX MAC status register has at - least 1 unmasked interrupt set */ -#define INTR_RX_MAC_STATUS 0x00008000 /* RX MAC status register has at - least 1 unmasked interrupt set */ -#define INTR_MAC_CTRL_STATUS 0x00010000 /* MAC control status register has - at least 1 unmasked interrupt - set */ -#define INTR_MIF_STATUS 0x00020000 /* MIF status register has at least - 1 unmasked interrupt set */ -#define INTR_PCI_ERROR_STATUS 0x00040000 /* PCI error status register in the - BIF has at least 1 unmasked - interrupt set */ -#define INTR_TX_COMP_3_MASK 0xFFF80000 /* mask for TX completion - 3 reg data */ -#define INTR_TX_COMP_3_SHIFT 19 -#define INTR_ERROR_MASK (INTR_MIF_STATUS | INTR_PCI_ERROR_STATUS | \ - INTR_PCS_STATUS | INTR_RX_LEN_MISMATCH | \ - INTR_TX_MAC_STATUS | INTR_RX_MAC_STATUS | \ - INTR_TX_TAG_ERROR | INTR_RX_TAG_ERROR | \ - INTR_MAC_CTRL_STATUS) - -/* determines which status events will cause an interrupt. layout same - * as REG_INTR_STATUS. - * DEFAULT: 0xFFFFFFFF, SIZE: 16 bits - */ -#define REG_INTR_MASK 0x0010 /* Interrupt mask */ - -/* top level interrupt bits that are cleared during read of REG_INTR_STATUS_ALIAS. - * useful when driver is polling for interrupts. layout same as REG_INTR_MASK. - * DEFAULT: 0x00000000, SIZE: 12 bits - */ -#define REG_ALIAS_CLEAR 0x0014 /* alias clear mask - (used w/ status alias) */ -/* same as REG_INTR_STATUS except that only bits cleared are those selected by - * REG_ALIAS_CLEAR - * DEFAULT: 0x00000000, SIZE: 29 bits - */ -#define REG_INTR_STATUS_ALIAS 0x001C /* interrupt status alias - (selective clear) */ - -/* DEFAULT: 0x0, SIZE: 3 bits */ -#define REG_PCI_ERR_STATUS 0x1000 /* PCI error status */ -#define PCI_ERR_BADACK 0x01 /* reserved in Cassini+. - set if no ACK64# during ABS64 cycle - in Cassini. */ -#define PCI_ERR_DTRTO 0x02 /* delayed xaction timeout. set if - no read retry after 2^15 clocks */ -#define PCI_ERR_OTHER 0x04 /* other PCI errors */ -#define PCI_ERR_BIM_DMA_WRITE 0x08 /* BIM received 0 count DMA write req. - unused in Cassini. */ -#define PCI_ERR_BIM_DMA_READ 0x10 /* BIM received 0 count DMA read req. - unused in Cassini. */ -#define PCI_ERR_BIM_DMA_TIMEOUT 0x20 /* BIM received 255 retries during - DMA. unused in cassini. */ - -/* mask for PCI status events that will set PCI_ERR_STATUS. if cleared, event - * causes an interrupt to be generated. - * DEFAULT: 0x7, SIZE: 3 bits - */ -#define REG_PCI_ERR_STATUS_MASK 0x1004 /* PCI Error status mask */ - -/* used to configure PCI related parameters that are not in PCI config space. - * DEFAULT: 0bxx000, SIZE: 5 bits - */ -#define REG_BIM_CFG 0x1008 /* BIM Configuration */ -#define BIM_CFG_RESERVED0 0x001 /* reserved */ -#define BIM_CFG_RESERVED1 0x002 /* reserved */ -#define BIM_CFG_64BIT_DISABLE 0x004 /* disable 64-bit mode */ -#define BIM_CFG_66MHZ 0x008 /* (ro) 1 = 66MHz, 0 = < 66MHz */ -#define BIM_CFG_32BIT 0x010 /* (ro) 1 = 32-bit slot, 0 = 64-bit */ -#define BIM_CFG_DPAR_INTR_ENABLE 0x020 /* detected parity err enable */ -#define BIM_CFG_RMA_INTR_ENABLE 0x040 /* master abort intr enable */ -#define BIM_CFG_RTA_INTR_ENABLE 0x080 /* target abort intr enable */ -#define BIM_CFG_RESERVED2 0x100 /* reserved */ -#define BIM_CFG_BIM_DISABLE 0x200 /* stop BIM DMA. use before global - reset. reserved in Cassini. */ -#define BIM_CFG_BIM_STATUS 0x400 /* (ro) 1 = BIM DMA suspended. - reserved in Cassini. */ -#define BIM_CFG_PERROR_BLOCK 0x800 /* block PERR# to pci bus. def: 0. - reserved in Cassini. */ - -/* DEFAULT: 0x00000000, SIZE: 32 bits */ -#define REG_BIM_DIAG 0x100C /* BIM Diagnostic */ -#define BIM_DIAG_MSTR_SM_MASK 0x3FFFFF00 /* PCI master controller state - machine bits [21:0] */ -#define BIM_DIAG_BRST_SM_MASK 0x7F /* PCI burst controller state - machine bits [6:0] */ - -/* writing to SW_RESET_TX and SW_RESET_RX will issue a global - * reset. poll until TX and RX read back as 0's for completion. - */ -#define REG_SW_RESET 0x1010 /* Software reset */ -#define SW_RESET_TX 0x00000001 /* reset TX DMA engine. poll until - cleared to 0. */ -#define SW_RESET_RX 0x00000002 /* reset RX DMA engine. poll until - cleared to 0. */ -#define SW_RESET_RSTOUT 0x00000004 /* force RSTOUT# pin active (low). - resets PHY and anything else - connected to RSTOUT#. RSTOUT# - is also activated by local PCI - reset when hot-swap is being - done. */ -#define SW_RESET_BLOCK_PCS_SLINK 0x00000008 /* if a global reset is done with - this bit set, PCS and SLINK - modules won't be reset. - i.e., link won't drop. */ -#define SW_RESET_BREQ_SM_MASK 0x00007F00 /* breq state machine [6:0] */ -#define SW_RESET_PCIARB_SM_MASK 0x00070000 /* pci arbitration state bits: - 0b000: ARB_IDLE1 - 0b001: ARB_IDLE2 - 0b010: ARB_WB_ACK - 0b011: ARB_WB_WAT - 0b100: ARB_RB_ACK - 0b101: ARB_RB_WAT - 0b110: ARB_RB_END - 0b111: ARB_WB_END */ -#define SW_RESET_RDPCI_SM_MASK 0x00300000 /* read pci state bits: - 0b00: RD_PCI_WAT - 0b01: RD_PCI_RDY - 0b11: RD_PCI_ACK */ -#define SW_RESET_RDARB_SM_MASK 0x00C00000 /* read arbitration state bits: - 0b00: AD_IDL_RX - 0b01: AD_ACK_RX - 0b10: AD_ACK_TX - 0b11: AD_IDL_TX */ -#define SW_RESET_WRPCI_SM_MASK 0x06000000 /* write pci state bits - 0b00: WR_PCI_WAT - 0b01: WR_PCI_RDY - 0b11: WR_PCI_ACK */ -#define SW_RESET_WRARB_SM_MASK 0x38000000 /* write arbitration state bits: - 0b000: ARB_IDLE1 - 0b001: ARB_IDLE2 - 0b010: ARB_TX_ACK - 0b011: ARB_TX_WAT - 0b100: ARB_RX_ACK - 0b110: ARB_RX_WAT */ - -/* Cassini only. 64-bit register used to check PCI datapath. when read, - * value written has both lower and upper 32-bit halves rotated to the right - * one bit position. e.g., FFFFFFFF FFFFFFFF -> 7FFFFFFF 7FFFFFFF - */ -#define REG_MINUS_BIM_DATAPATH_TEST 0x1018 /* Cassini: BIM datapath test - Cassini+: reserved */ - -/* output enables are provided for each device's chip select and for the rest - * of the outputs from cassini to its local bus devices. two sw programmable - * bits are connected to general purpus control/status bits. - * DEFAULT: 0x7 - */ -#define REG_BIM_LOCAL_DEV_EN 0x1020 /* BIM local device - output EN. default: 0x7 */ -#define BIM_LOCAL_DEV_PAD 0x01 /* address bus, RW signal, and - OE signal output enable on the - local bus interface. these - are shared between both local - bus devices. tristate when 0. */ -#define BIM_LOCAL_DEV_PROM 0x02 /* PROM chip select */ -#define BIM_LOCAL_DEV_EXT 0x04 /* secondary local bus device chip - select output enable */ -#define BIM_LOCAL_DEV_SOFT_0 0x08 /* sw programmable ctrl bit 0 */ -#define BIM_LOCAL_DEV_SOFT_1 0x10 /* sw programmable ctrl bit 1 */ -#define BIM_LOCAL_DEV_HW_RESET 0x20 /* internal hw reset. Cassini+ only. */ - -/* access 24 entry BIM read and write buffers. put address in REG_BIM_BUFFER_ADDR - * and read/write from/to it REG_BIM_BUFFER_DATA_LOW and _DATA_HI. - * _DATA_HI should be the last access of the sequence. - * DEFAULT: undefined - */ -#define REG_BIM_BUFFER_ADDR 0x1024 /* BIM buffer address. for - purposes. */ -#define BIM_BUFFER_ADDR_MASK 0x3F /* index (0 - 23) of buffer */ -#define BIM_BUFFER_WR_SELECT 0x40 /* write buffer access = 1 - read buffer access = 0 */ -/* DEFAULT: undefined */ -#define REG_BIM_BUFFER_DATA_LOW 0x1028 /* BIM buffer data low */ -#define REG_BIM_BUFFER_DATA_HI 0x102C /* BIM buffer data high */ - -/* set BIM_RAM_BIST_START to start built-in self test for BIM read buffer. - * bit auto-clears when done with status read from _SUMMARY and _PASS bits. - */ -#define REG_BIM_RAM_BIST 0x102C /* BIM RAM (read buffer) BIST - control/status */ -#define BIM_RAM_BIST_RD_START 0x01 /* start BIST for BIM read buffer */ -#define BIM_RAM_BIST_WR_START 0x02 /* start BIST for BIM write buffer. - Cassini only. reserved in - Cassini+. */ -#define BIM_RAM_BIST_RD_PASS 0x04 /* summary BIST pass status for read - buffer. */ -#define BIM_RAM_BIST_WR_PASS 0x08 /* summary BIST pass status for write - buffer. Cassini only. reserved - in Cassini+. */ -#define BIM_RAM_BIST_RD_LOW_PASS 0x10 /* read low bank passes BIST */ -#define BIM_RAM_BIST_RD_HI_PASS 0x20 /* read high bank passes BIST */ -#define BIM_RAM_BIST_WR_LOW_PASS 0x40 /* write low bank passes BIST. - Cassini only. reserved in - Cassini+. */ -#define BIM_RAM_BIST_WR_HI_PASS 0x80 /* write high bank passes BIST. - Cassini only. reserved in - Cassini+. */ - -/* ASUN: i'm not sure what this does as it's not in the spec. - * DEFAULT: 0xFC - */ -#define REG_BIM_DIAG_MUX 0x1030 /* BIM diagnostic probe mux - select register */ - -/* enable probe monitoring mode and select data appearing on the P_A* bus. bit - * values for _SEL_HI_MASK and _SEL_LOW_MASK: - * 0x0: internal probe[7:0] (pci arb state, wtc empty w, wtc full w, wtc empty w, - * wtc empty r, post pci) - * 0x1: internal probe[15:8] (pci wbuf comp, pci wpkt comp, pci rbuf comp, - * pci rpkt comp, txdma wr req, txdma wr ack, - * txdma wr rdy, txdma wr xfr done) - * 0x2: internal probe[23:16] (txdma rd req, txdma rd ack, txdma rd rdy, rxdma rd, - * rd arb state, rd pci state) - * 0x3: internal probe[31:24] (rxdma req, rxdma ack, rxdma rdy, wrarb state, - * wrpci state) - * 0x4: pci io probe[7:0] 0x5: pci io probe[15:8] - * 0x6: pci io probe[23:16] 0x7: pci io probe[31:24] - * 0x8: pci io probe[39:32] 0x9: pci io probe[47:40] - * 0xa: pci io probe[55:48] 0xb: pci io probe[63:56] - * the following are not available in Cassini: - * 0xc: rx probe[7:0] 0xd: tx probe[7:0] - * 0xe: hp probe[7:0] 0xf: mac probe[7:0] - */ -#define REG_PLUS_PROBE_MUX_SELECT 0x1034 /* Cassini+: PROBE MUX SELECT */ -#define PROBE_MUX_EN 0x80000000 /* allow probe signals to be - driven on local bus P_A[15:0] - for debugging */ -#define PROBE_MUX_SUB_MUX_MASK 0x0000FF00 /* select sub module probe signals: - 0x03 = mac[1:0] - 0x0C = rx[1:0] - 0x30 = tx[1:0] - 0xC0 = hp[1:0] */ -#define PROBE_MUX_SEL_HI_MASK 0x000000F0 /* select which module to appear - on P_A[15:8]. see above for - values. */ -#define PROBE_MUX_SEL_LOW_MASK 0x0000000F /* select which module to appear - on P_A[7:0]. see above for - values. */ - -/* values mean the same thing as REG_INTR_MASK excep that it's for INTB. - DEFAULT: 0x1F */ -#define REG_PLUS_INTR_MASK_1 0x1038 /* Cassini+: interrupt mask - register 2 for INTB */ -#define REG_PLUS_INTRN_MASK(x) (REG_PLUS_INTR_MASK_1 + ((x) - 1)*16) -/* bits correspond to both _MASK and _STATUS registers. _ALT corresponds to - * all of the alternate (2-4) INTR registers while _1 corresponds to only - * _MASK_1 and _STATUS_1 registers. - * DEFAULT: 0x7 for MASK registers, 0x0 for ALIAS_CLEAR registers - */ -#define INTR_RX_DONE_ALT 0x01 -#define INTR_RX_COMP_FULL_ALT 0x02 -#define INTR_RX_COMP_AF_ALT 0x04 -#define INTR_RX_BUF_UNAVAIL_1 0x08 -#define INTR_RX_BUF_AE_1 0x10 /* almost empty */ -#define INTRN_MASK_RX_EN 0x80 -#define INTRN_MASK_CLEAR_ALL (INTR_RX_DONE_ALT | \ - INTR_RX_COMP_FULL_ALT | \ - INTR_RX_COMP_AF_ALT | \ - INTR_RX_BUF_UNAVAIL_1 | \ - INTR_RX_BUF_AE_1) -#define REG_PLUS_INTR_STATUS_1 0x103C /* Cassini+: interrupt status - register 2 for INTB. default: 0x1F */ -#define REG_PLUS_INTRN_STATUS(x) (REG_PLUS_INTR_STATUS_1 + ((x) - 1)*16) -#define INTR_STATUS_ALT_INTX_EN 0x80 /* generate INTX when one of the - flags are set. enables desc ring. */ - -#define REG_PLUS_ALIAS_CLEAR_1 0x1040 /* Cassini+: alias clear mask - register 2 for INTB */ -#define REG_PLUS_ALIASN_CLEAR(x) (REG_PLUS_ALIAS_CLEAR_1 + ((x) - 1)*16) - -#define REG_PLUS_INTR_STATUS_ALIAS_1 0x1044 /* Cassini+: interrupt status - register alias 2 for INTB */ -#define REG_PLUS_INTRN_STATUS_ALIAS(x) (REG_PLUS_INTR_STATUS_ALIAS_1 + ((x) - 1)*16) - -#define REG_SATURN_PCFG 0x106c /* pin configuration register for - integrated macphy */ - -#define SATURN_PCFG_TLA 0x00000001 /* 1 = phy actled */ -#define SATURN_PCFG_FLA 0x00000002 /* 1 = phy link10led */ -#define SATURN_PCFG_CLA 0x00000004 /* 1 = phy link100led */ -#define SATURN_PCFG_LLA 0x00000008 /* 1 = phy link1000led */ -#define SATURN_PCFG_RLA 0x00000010 /* 1 = phy duplexled */ -#define SATURN_PCFG_PDS 0x00000020 /* phy debug mode. - 0 = normal */ -#define SATURN_PCFG_MTP 0x00000080 /* test point select */ -#define SATURN_PCFG_GMO 0x00000100 /* GMII observe. 1 = - GMII on SERDES pins for - monitoring. */ -#define SATURN_PCFG_FSI 0x00000200 /* 1 = freeze serdes/gmii. all - pins configed as outputs. - for power saving when using - internal phy. */ -#define SATURN_PCFG_LAD 0x00000800 /* 0 = mac core led ctrl - polarity from strapping - value. - 1 = mac core led ctrl - polarity active low. */ - - -/** transmit dma registers **/ -#define MAX_TX_RINGS_SHIFT 2 -#define MAX_TX_RINGS (1 << MAX_TX_RINGS_SHIFT) -#define MAX_TX_RINGS_MASK (MAX_TX_RINGS - 1) - -/* TX configuration. - * descr ring sizes size = 32 * (1 << n), n < 9. e.g., 0x8 = 8k. default: 0x8 - * DEFAULT: 0x3F000001 - */ -#define REG_TX_CFG 0x2004 /* TX config */ -#define TX_CFG_DMA_EN 0x00000001 /* enable TX DMA. if cleared, DMA - will stop after xfer of current - buffer has been completed. */ -#define TX_CFG_FIFO_PIO_SEL 0x00000002 /* TX DMA FIFO can be - accessed w/ FIFO addr - and data registers. - TX DMA should be - disabled. */ -#define TX_CFG_DESC_RING0_MASK 0x0000003C /* # desc entries in - ring 1. */ -#define TX_CFG_DESC_RING0_SHIFT 2 -#define TX_CFG_DESC_RINGN_MASK(a) (TX_CFG_DESC_RING0_MASK << (a)*4) -#define TX_CFG_DESC_RINGN_SHIFT(a) (TX_CFG_DESC_RING0_SHIFT + (a)*4) -#define TX_CFG_PACED_MODE 0x00100000 /* TX_ALL only set after - TX FIFO becomes empty. - if 0, TX_ALL set - if descr queue empty. */ -#define TX_CFG_DMA_RDPIPE_DIS 0x01000000 /* always set to 1 */ -#define TX_CFG_COMPWB_Q1 0x02000000 /* completion writeback happens at - the end of every packet kicked - through Q1. */ -#define TX_CFG_COMPWB_Q2 0x04000000 /* completion writeback happens at - the end of every packet kicked - through Q2. */ -#define TX_CFG_COMPWB_Q3 0x08000000 /* completion writeback happens at - the end of every packet kicked - through Q3 */ -#define TX_CFG_COMPWB_Q4 0x10000000 /* completion writeback happens at - the end of every packet kicked - through Q4 */ -#define TX_CFG_INTR_COMPWB_DIS 0x20000000 /* disable pre-interrupt completion - writeback */ -#define TX_CFG_CTX_SEL_MASK 0xC0000000 /* selects tx test port - connection - 0b00: tx mac req, - tx mac retry req, - tx ack and tx tag. - 0b01: txdma rd req, - txdma rd ack, - txdma rd rdy, - txdma rd type0 - 0b11: txdma wr req, - txdma wr ack, - txdma wr rdy, - txdma wr xfr done. */ -#define TX_CFG_CTX_SEL_SHIFT 30 - -/* 11-bit counters that point to next location in FIFO to be loaded/retrieved. - * used for diagnostics only. - */ -#define REG_TX_FIFO_WRITE_PTR 0x2014 /* TX FIFO write pointer */ -#define REG_TX_FIFO_SHADOW_WRITE_PTR 0x2018 /* TX FIFO shadow write - pointer. temp hold reg. - diagnostics only. */ -#define REG_TX_FIFO_READ_PTR 0x201C /* TX FIFO read pointer */ -#define REG_TX_FIFO_SHADOW_READ_PTR 0x2020 /* TX FIFO shadow read - pointer */ - -/* (ro) 11-bit up/down counter w/ # of frames currently in TX FIFO */ -#define REG_TX_FIFO_PKT_CNT 0x2024 /* TX FIFO packet counter */ - -/* current state of all state machines in TX */ -#define REG_TX_SM_1 0x2028 /* TX state machine reg #1 */ -#define TX_SM_1_CHAIN_MASK 0x000003FF /* chaining state machine */ -#define TX_SM_1_CSUM_MASK 0x00000C00 /* checksum state machine */ -#define TX_SM_1_FIFO_LOAD_MASK 0x0003F000 /* FIFO load state machine. - = 0x01 when TX disabled. */ -#define TX_SM_1_FIFO_UNLOAD_MASK 0x003C0000 /* FIFO unload state machine */ -#define TX_SM_1_CACHE_MASK 0x03C00000 /* desc. prefetch cache controller - state machine */ -#define TX_SM_1_CBQ_ARB_MASK 0xF8000000 /* CBQ arbiter state machine */ - -#define REG_TX_SM_2 0x202C /* TX state machine reg #2 */ -#define TX_SM_2_COMP_WB_MASK 0x07 /* completion writeback sm */ -#define TX_SM_2_SUB_LOAD_MASK 0x38 /* sub load state machine */ -#define TX_SM_2_KICK_MASK 0xC0 /* kick state machine */ - -/* 64-bit pointer to the transmit data buffer. only the 50 LSB are incremented - * while the upper 23 bits are taken from the TX descriptor - */ -#define REG_TX_DATA_PTR_LOW 0x2030 /* TX data pointer low */ -#define REG_TX_DATA_PTR_HI 0x2034 /* TX data pointer high */ - -/* 13 bit registers written by driver w/ descriptor value that follows - * last valid xmit descriptor. kick # and complete # values are used by - * the xmit dma engine to control tx descr fetching. if > 1 valid - * tx descr is available within the cache line being read, cassini will - * internally cache up to 4 of them. 0 on reset. _KICK = rw, _COMP = ro. - */ -#define REG_TX_KICK0 0x2038 /* TX kick reg #1 */ -#define REG_TX_KICKN(x) (REG_TX_KICK0 + (x)*4) -#define REG_TX_COMP0 0x2048 /* TX completion reg #1 */ -#define REG_TX_COMPN(x) (REG_TX_COMP0 + (x)*4) - -/* values of TX_COMPLETE_1-4 are written. each completion register - * is 2bytes in size and contiguous. 8B allocation w/ 8B alignment. - * NOTE: completion reg values are only written back prior to TX_INTME and - * TX_ALL interrupts. at all other times, the most up-to-date index values - * should be obtained from the REG_TX_COMPLETE_# registers. - * here's the layout: - * offset from base addr completion # byte - * 0 TX_COMPLETE_1_MSB - * 1 TX_COMPLETE_1_LSB - * 2 TX_COMPLETE_2_MSB - * 3 TX_COMPLETE_2_LSB - * 4 TX_COMPLETE_3_MSB - * 5 TX_COMPLETE_3_LSB - * 6 TX_COMPLETE_4_MSB - * 7 TX_COMPLETE_4_LSB - */ -#define TX_COMPWB_SIZE 8 -#define REG_TX_COMPWB_DB_LOW 0x2058 /* TX completion write back - base low */ -#define REG_TX_COMPWB_DB_HI 0x205C /* TX completion write back - base high */ -#define TX_COMPWB_MSB_MASK 0x00000000000000FFULL -#define TX_COMPWB_MSB_SHIFT 0 -#define TX_COMPWB_LSB_MASK 0x000000000000FF00ULL -#define TX_COMPWB_LSB_SHIFT 8 -#define TX_COMPWB_NEXT(x) ((x) >> 16) - -/* 53 MSB used as base address. 11 LSB assumed to be 0. TX desc pointer must - * be 2KB-aligned. */ -#define REG_TX_DB0_LOW 0x2060 /* TX descriptor base low #1 */ -#define REG_TX_DB0_HI 0x2064 /* TX descriptor base hi #1 */ -#define REG_TX_DBN_LOW(x) (REG_TX_DB0_LOW + (x)*8) -#define REG_TX_DBN_HI(x) (REG_TX_DB0_HI + (x)*8) - -/* 16-bit registers hold weights for the weighted round-robin of the - * four CBQ TX descr rings. weights correspond to # bytes xferred from - * host to TXFIFO in a round of WRR arbitration. can be set - * dynamically with new weights set upon completion of the current - * packet transfer from host memory to TXFIFO. a dummy write to any of - * these registers causes a queue1 pre-emption with all historical bw - * deficit data reset to 0 (useful when congestion requires a - * pre-emption/re-allocation of network bandwidth - */ -#define REG_TX_MAXBURST_0 0x2080 /* TX MaxBurst #1 */ -#define REG_TX_MAXBURST_1 0x2084 /* TX MaxBurst #2 */ -#define REG_TX_MAXBURST_2 0x2088 /* TX MaxBurst #3 */ -#define REG_TX_MAXBURST_3 0x208C /* TX MaxBurst #4 */ - -/* diagnostics access to any TX FIFO location. every access is 65 - * bits. _DATA_LOW = 32 LSB, _DATA_HI_T1/T0 = 32 MSB. _TAG = tag bit. - * writing _DATA_HI_T0 sets tag bit low, writing _DATA_HI_T1 sets tag - * bit high. TX_FIFO_PIO_SEL must be set for TX FIFO PIO access. if - * TX FIFO data integrity is desired, TX DMA should be - * disabled. _DATA_HI_Tx should be the last access of the sequence. - */ -#define REG_TX_FIFO_ADDR 0x2104 /* TX FIFO address */ -#define REG_TX_FIFO_TAG 0x2108 /* TX FIFO tag */ -#define REG_TX_FIFO_DATA_LOW 0x210C /* TX FIFO data low */ -#define REG_TX_FIFO_DATA_HI_T1 0x2110 /* TX FIFO data high t1 */ -#define REG_TX_FIFO_DATA_HI_T0 0x2114 /* TX FIFO data high t0 */ -#define REG_TX_FIFO_SIZE 0x2118 /* (ro) TX FIFO size = 0x090 = 9KB */ - -/* 9-bit register controls BIST of TX FIFO. bit set indicates that the BIST - * passed for the specified memory - */ -#define REG_TX_RAMBIST 0x211C /* TX RAMBIST control/status */ -#define TX_RAMBIST_STATE 0x01C0 /* progress state of RAMBIST - controller state machine */ -#define TX_RAMBIST_RAM33A_PASS 0x0020 /* RAM33A passed */ -#define TX_RAMBIST_RAM32A_PASS 0x0010 /* RAM32A passed */ -#define TX_RAMBIST_RAM33B_PASS 0x0008 /* RAM33B passed */ -#define TX_RAMBIST_RAM32B_PASS 0x0004 /* RAM32B passed */ -#define TX_RAMBIST_SUMMARY 0x0002 /* all RAM passed */ -#define TX_RAMBIST_START 0x0001 /* write 1 to start BIST. self - clears on completion. */ - -/** receive dma registers **/ -#define MAX_RX_DESC_RINGS 2 -#define MAX_RX_COMP_RINGS 4 - -/* receive DMA channel configuration. default: 0x80910 - * free ring size = (1 << n)*32 -> [32 - 8k] - * completion ring size = (1 << n)*128 -> [128 - 32k], n < 9 - * DEFAULT: 0x80910 - */ -#define REG_RX_CFG 0x4000 /* RX config */ -#define RX_CFG_DMA_EN 0x00000001 /* enable RX DMA. 0 stops - channel as soon as current - frame xfer has completed. - driver should disable MAC - for 200ms before disabling - RX */ -#define RX_CFG_DESC_RING_MASK 0x0000001E /* # desc entries in RX - free desc ring. - def: 0x8 = 8k */ -#define RX_CFG_DESC_RING_SHIFT 1 -#define RX_CFG_COMP_RING_MASK 0x000001E0 /* # desc entries in RX complete - ring. def: 0x8 = 32k */ -#define RX_CFG_COMP_RING_SHIFT 5 -#define RX_CFG_BATCH_DIS 0x00000200 /* disable receive desc - batching. def: 0x0 = - enabled */ -#define RX_CFG_SWIVEL_MASK 0x00001C00 /* byte offset of the 1st - data byte of the packet - w/in 8 byte boundares. - this swivels the data - DMA'ed to header - buffers, jumbo buffers - when header split is not - requested and MTU sized - buffers. def: 0x2 */ -#define RX_CFG_SWIVEL_SHIFT 10 - -/* cassini+ only */ -#define RX_CFG_DESC_RING1_MASK 0x000F0000 /* # of desc entries in - RX free desc ring 2. - def: 0x8 = 8k */ -#define RX_CFG_DESC_RING1_SHIFT 16 - - -/* the page size register allows cassini chips to do the following with - * received data: - * [--------------------------------------------------------------] page - * [off][buf1][pad][off][buf2][pad][off][buf3][pad][off][buf4][pad] - * |--------------| = PAGE_SIZE_BUFFER_STRIDE - * page = PAGE_SIZE - * offset = PAGE_SIZE_MTU_OFF - * for the above example, MTU_BUFFER_COUNT = 4. - * NOTE: as is apparent, you need to ensure that the following holds: - * MTU_BUFFER_COUNT <= PAGE_SIZE/PAGE_SIZE_BUFFER_STRIDE - * DEFAULT: 0x48002002 (8k pages) - */ -#define REG_RX_PAGE_SIZE 0x4004 /* RX page size */ -#define RX_PAGE_SIZE_MASK 0x00000003 /* size of pages pointed to - by receive descriptors. - if jumbo buffers are - supported the page size - should not be < 8k. - 0b00 = 2k, 0b01 = 4k - 0b10 = 8k, 0b11 = 16k - DEFAULT: 8k */ -#define RX_PAGE_SIZE_SHIFT 0 -#define RX_PAGE_SIZE_MTU_COUNT_MASK 0x00007800 /* # of MTU buffers the hw - packs into a page. - DEFAULT: 4 */ -#define RX_PAGE_SIZE_MTU_COUNT_SHIFT 11 -#define RX_PAGE_SIZE_MTU_STRIDE_MASK 0x18000000 /* # of bytes that separate - each MTU buffer + - offset from each - other. - 0b00 = 1k, 0b01 = 2k - 0b10 = 4k, 0b11 = 8k - DEFAULT: 0x1 */ -#define RX_PAGE_SIZE_MTU_STRIDE_SHIFT 27 -#define RX_PAGE_SIZE_MTU_OFF_MASK 0xC0000000 /* offset in each page that - hw writes the MTU buffer - into. - 0b00 = 0, - 0b01 = 64 bytes - 0b10 = 96, 0b11 = 128 - DEFAULT: 0x1 */ -#define RX_PAGE_SIZE_MTU_OFF_SHIFT 30 - -/* 11-bit counter points to next location in RX FIFO to be loaded/read. - * shadow write pointers enable retries in case of early receive aborts. - * DEFAULT: 0x0. generated on 64-bit boundaries. - */ -#define REG_RX_FIFO_WRITE_PTR 0x4008 /* RX FIFO write pointer */ -#define REG_RX_FIFO_READ_PTR 0x400C /* RX FIFO read pointer */ -#define REG_RX_IPP_FIFO_SHADOW_WRITE_PTR 0x4010 /* RX IPP FIFO shadow write - pointer */ -#define REG_RX_IPP_FIFO_SHADOW_READ_PTR 0x4014 /* RX IPP FIFO shadow read - pointer */ -#define REG_RX_IPP_FIFO_READ_PTR 0x400C /* RX IPP FIFO read - pointer. (8-bit counter) */ - -/* current state of RX DMA state engines + other info - * DEFAULT: 0x0 - */ -#define REG_RX_DEBUG 0x401C /* RX debug */ -#define RX_DEBUG_LOAD_STATE_MASK 0x0000000F /* load state machine w/ MAC: - 0x0 = idle, 0x1 = load_bop - 0x2 = load 1, 0x3 = load 2 - 0x4 = load 3, 0x5 = load 4 - 0x6 = last detect - 0x7 = wait req - 0x8 = wait req statuss 1st - 0x9 = load st - 0xa = bubble mac - 0xb = error */ -#define RX_DEBUG_LM_STATE_MASK 0x00000070 /* load state machine w/ HP and - RX FIFO: - 0x0 = idle, 0x1 = hp xfr - 0x2 = wait hp ready - 0x3 = wait flow code - 0x4 = fifo xfer - 0x5 = make status - 0x6 = csum ready - 0x7 = error */ -#define RX_DEBUG_FC_STATE_MASK 0x000000180 /* flow control state machine - w/ MAC: - 0x0 = idle - 0x1 = wait xoff ack - 0x2 = wait xon - 0x3 = wait xon ack */ -#define RX_DEBUG_DATA_STATE_MASK 0x000001E00 /* unload data state machine - states: - 0x0 = idle data - 0x1 = header begin - 0x2 = xfer header - 0x3 = xfer header ld - 0x4 = mtu begin - 0x5 = xfer mtu - 0x6 = xfer mtu ld - 0x7 = jumbo begin - 0x8 = xfer jumbo - 0x9 = xfer jumbo ld - 0xa = reas begin - 0xb = xfer reas - 0xc = flush tag - 0xd = xfer reas ld - 0xe = error - 0xf = bubble idle */ -#define RX_DEBUG_DESC_STATE_MASK 0x0001E000 /* unload desc state machine - states: - 0x0 = idle desc - 0x1 = wait ack - 0x9 = wait ack 2 - 0x2 = fetch desc 1 - 0xa = fetch desc 2 - 0x3 = load ptrs - 0x4 = wait dma - 0x5 = wait ack batch - 0x6 = post batch - 0x7 = xfr done */ -#define RX_DEBUG_INTR_READ_PTR_MASK 0x30000000 /* interrupt read ptr of the - interrupt queue */ -#define RX_DEBUG_INTR_WRITE_PTR_MASK 0xC0000000 /* interrupt write pointer - of the interrupt queue */ - -/* flow control frames are emitted using two PAUSE thresholds: - * XOFF PAUSE uses pause time value pre-programmed in the Send PAUSE MAC reg - * XON PAUSE uses a pause time of 0. granularity of threshold is 64bytes. - * PAUSE thresholds defined in terms of FIFO occupancy and may be translated - * into FIFO vacancy using RX_FIFO_SIZE. setting ON will trigger XON frames - * when FIFO reaches 0. OFF threshold should not be > size of RX FIFO. max - * value is 0x6F. - * DEFAULT: 0x00078 - */ -#define REG_RX_PAUSE_THRESH 0x4020 /* RX pause thresholds */ -#define RX_PAUSE_THRESH_QUANTUM 64 -#define RX_PAUSE_THRESH_OFF_MASK 0x000001FF /* XOFF PAUSE emitted when - RX FIFO occupancy > - value*64B */ -#define RX_PAUSE_THRESH_OFF_SHIFT 0 -#define RX_PAUSE_THRESH_ON_MASK 0x001FF000 /* XON PAUSE emitted after - emitting XOFF PAUSE when RX - FIFO occupancy falls below - this value*64B. must be - < XOFF threshold. if = - RX_FIFO_SIZE< XON frames are - never emitted. */ -#define RX_PAUSE_THRESH_ON_SHIFT 12 - -/* 13-bit register used to control RX desc fetching and intr generation. if 4+ - * valid RX descriptors are available, Cassini will read 4 at a time. - * writing N means that all desc up to *but* excluding N are available. N must - * be a multiple of 4 (N % 4 = 0). first desc should be cache-line aligned. - * DEFAULT: 0 on reset - */ -#define REG_RX_KICK 0x4024 /* RX kick reg */ - -/* 8KB aligned 64-bit pointer to the base of the RX free/completion rings. - * lower 13 bits of the low register are hard-wired to 0. - */ -#define REG_RX_DB_LOW 0x4028 /* RX descriptor ring - base low */ -#define REG_RX_DB_HI 0x402C /* RX descriptor ring - base hi */ -#define REG_RX_CB_LOW 0x4030 /* RX completion ring - base low */ -#define REG_RX_CB_HI 0x4034 /* RX completion ring - base hi */ -/* 13-bit register indicate desc used by cassini for receive frames. used - * for diagnostic purposes. - * DEFAULT: 0 on reset - */ -#define REG_RX_COMP 0x4038 /* (ro) RX completion */ - -/* HEAD and TAIL are used to control RX desc posting and interrupt - * generation. hw moves the head register to pass ownership to sw. sw - * moves the tail register to pass ownership back to hw. to give all - * entries to hw, set TAIL = HEAD. if HEAD and TAIL indicate that no - * more entries are available, DMA will pause and an interrupt will be - * generated to indicate no more entries are available. sw can use - * this interrupt to reduce the # of times it must update the - * completion tail register. - * DEFAULT: 0 on reset - */ -#define REG_RX_COMP_HEAD 0x403C /* RX completion head */ -#define REG_RX_COMP_TAIL 0x4040 /* RX completion tail */ - -/* values used for receive interrupt blanking. loaded each time the ISR is read - * DEFAULT: 0x00000000 - */ -#define REG_RX_BLANK 0x4044 /* RX blanking register - for ISR read */ -#define RX_BLANK_INTR_PKT_MASK 0x000001FF /* RX_DONE intr asserted if - this many sets of completion - writebacks (up to 2 packets) - occur since the last time - the ISR was read. 0 = no - packet blanking */ -#define RX_BLANK_INTR_PKT_SHIFT 0 -#define RX_BLANK_INTR_TIME_MASK 0x3FFFF000 /* RX_DONE interrupt asserted - if that many clocks were - counted since last time the - ISR was read. - each count is 512 core - clocks (125MHz). 0 = no - time blanking */ -#define RX_BLANK_INTR_TIME_SHIFT 12 - -/* values used for interrupt generation based on threshold values of how - * many free desc and completion entries are available for hw use. - * DEFAULT: 0x00000000 - */ -#define REG_RX_AE_THRESH 0x4048 /* RX almost empty - thresholds */ -#define RX_AE_THRESH_FREE_MASK 0x00001FFF /* RX_BUF_AE will be - generated if # desc - avail for hw use <= - # */ -#define RX_AE_THRESH_FREE_SHIFT 0 -#define RX_AE_THRESH_COMP_MASK 0x0FFFE000 /* RX_COMP_AE will be - generated if # of - completion entries - avail for hw use <= - # */ -#define RX_AE_THRESH_COMP_SHIFT 13 - -/* probabilities for random early drop (RED) thresholds on a FIFO threshold - * basis. probability should increase when the FIFO level increases. control - * packets are never dropped and not counted in stats. probability programmed - * on a 12.5% granularity. e.g., 0x1 = 1/8 packets dropped. - * DEFAULT: 0x00000000 - */ -#define REG_RX_RED 0x404C /* RX random early detect enable */ -#define RX_RED_4K_6K_FIFO_MASK 0x000000FF /* 4KB < FIFO thresh < 6KB */ -#define RX_RED_6K_8K_FIFO_MASK 0x0000FF00 /* 6KB < FIFO thresh < 8KB */ -#define RX_RED_8K_10K_FIFO_MASK 0x00FF0000 /* 8KB < FIFO thresh < 10KB */ -#define RX_RED_10K_12K_FIFO_MASK 0xFF000000 /* 10KB < FIFO thresh < 12KB */ - -/* FIFO fullness levels for RX FIFO, RX control FIFO, and RX IPP FIFO. - * RX control FIFO = # of packets in RX FIFO. - * DEFAULT: 0x0 - */ -#define REG_RX_FIFO_FULLNESS 0x4050 /* (ro) RX FIFO fullness */ -#define RX_FIFO_FULLNESS_RX_FIFO_MASK 0x3FF80000 /* level w/ 8B granularity */ -#define RX_FIFO_FULLNESS_IPP_FIFO_MASK 0x0007FF00 /* level w/ 8B granularity */ -#define RX_FIFO_FULLNESS_RX_PKT_MASK 0x000000FF /* # packets in RX FIFO */ -#define REG_RX_IPP_PACKET_COUNT 0x4054 /* RX IPP packet counter */ -#define REG_RX_WORK_DMA_PTR_LOW 0x4058 /* RX working DMA ptr low */ -#define REG_RX_WORK_DMA_PTR_HI 0x405C /* RX working DMA ptr - high */ - -/* BIST testing ro RX FIFO, RX control FIFO, and RX IPP FIFO. only RX BIST - * START/COMPLETE is writeable. START will clear when the BIST has completed - * checking all 17 RAMS. - * DEFAULT: 0bxxxx xxxxx xxxx xxxx xxxx x000 0000 0000 00x0 - */ -#define REG_RX_BIST 0x4060 /* (ro) RX BIST */ -#define RX_BIST_32A_PASS 0x80000000 /* RX FIFO 32A passed */ -#define RX_BIST_33A_PASS 0x40000000 /* RX FIFO 33A passed */ -#define RX_BIST_32B_PASS 0x20000000 /* RX FIFO 32B passed */ -#define RX_BIST_33B_PASS 0x10000000 /* RX FIFO 33B passed */ -#define RX_BIST_32C_PASS 0x08000000 /* RX FIFO 32C passed */ -#define RX_BIST_33C_PASS 0x04000000 /* RX FIFO 33C passed */ -#define RX_BIST_IPP_32A_PASS 0x02000000 /* RX IPP FIFO 33B passed */ -#define RX_BIST_IPP_33A_PASS 0x01000000 /* RX IPP FIFO 33A passed */ -#define RX_BIST_IPP_32B_PASS 0x00800000 /* RX IPP FIFO 32B passed */ -#define RX_BIST_IPP_33B_PASS 0x00400000 /* RX IPP FIFO 33B passed */ -#define RX_BIST_IPP_32C_PASS 0x00200000 /* RX IPP FIFO 32C passed */ -#define RX_BIST_IPP_33C_PASS 0x00100000 /* RX IPP FIFO 33C passed */ -#define RX_BIST_CTRL_32_PASS 0x00800000 /* RX CTRL FIFO 32 passed */ -#define RX_BIST_CTRL_33_PASS 0x00400000 /* RX CTRL FIFO 33 passed */ -#define RX_BIST_REAS_26A_PASS 0x00200000 /* RX Reas 26A passed */ -#define RX_BIST_REAS_26B_PASS 0x00100000 /* RX Reas 26B passed */ -#define RX_BIST_REAS_27_PASS 0x00080000 /* RX Reas 27 passed */ -#define RX_BIST_STATE_MASK 0x00078000 /* BIST state machine */ -#define RX_BIST_SUMMARY 0x00000002 /* when BIST complete, - summary pass bit - contains AND of BIST - results of all 16 - RAMS */ -#define RX_BIST_START 0x00000001 /* write 1 to start - BIST. self clears - on completion. */ - -/* next location in RX CTRL FIFO that will be loaded w/ data from RX IPP/read - * from to retrieve packet control info. - * DEFAULT: 0 - */ -#define REG_RX_CTRL_FIFO_WRITE_PTR 0x4064 /* (ro) RX control FIFO - write ptr */ -#define REG_RX_CTRL_FIFO_READ_PTR 0x4068 /* (ro) RX control FIFO read - ptr */ - -/* receive interrupt blanking. loaded each time interrupt alias register is - * read. - * DEFAULT: 0x0 - */ -#define REG_RX_BLANK_ALIAS_READ 0x406C /* RX blanking register for - alias read */ -#define RX_BAR_INTR_PACKET_MASK 0x000001FF /* assert RX_DONE if # - completion writebacks - > # since last ISR - read. 0 = no - blanking. up to 2 - packets per - completion wb. */ -#define RX_BAR_INTR_TIME_MASK 0x3FFFF000 /* assert RX_DONE if # - clocks > # since last - ISR read. each count - is 512 core clocks - (125MHz). 0 = no - blanking. */ - -/* diagnostic access to RX FIFO. 32 LSB accessed via DATA_LOW. 32 MSB accessed - * via DATA_HI_T0 or DATA_HI_T1. TAG reads the tag bit. writing HI_T0 - * will unset the tag bit while writing HI_T1 will set the tag bit. to reset - * to normal operation after diagnostics, write to address location 0x0. - * RX_DMA_EN bit must be set to 0x0 for RX FIFO PIO access. DATA_HI should - * be the last write access of a write sequence. - * DEFAULT: undefined - */ -#define REG_RX_FIFO_ADDR 0x4080 /* RX FIFO address */ -#define REG_RX_FIFO_TAG 0x4084 /* RX FIFO tag */ -#define REG_RX_FIFO_DATA_LOW 0x4088 /* RX FIFO data low */ -#define REG_RX_FIFO_DATA_HI_T0 0x408C /* RX FIFO data high T0 */ -#define REG_RX_FIFO_DATA_HI_T1 0x4090 /* RX FIFO data high T1 */ - -/* diagnostic assess to RX CTRL FIFO. 8-bit FIFO_ADDR holds address of - * 81 bit control entry and 6 bit flow id. LOW and MID are both 32-bit - * accesses. HI is 7-bits with 6-bit flow id and 1 bit control - * word. RX_DMA_EN must be 0 for RX CTRL FIFO PIO access. DATA_HI - * should be last write access of the write sequence. - * DEFAULT: undefined - */ -#define REG_RX_CTRL_FIFO_ADDR 0x4094 /* RX Control FIFO and - Batching FIFO addr */ -#define REG_RX_CTRL_FIFO_DATA_LOW 0x4098 /* RX Control FIFO data - low */ -#define REG_RX_CTRL_FIFO_DATA_MID 0x409C /* RX Control FIFO data - mid */ -#define REG_RX_CTRL_FIFO_DATA_HI 0x4100 /* RX Control FIFO data - hi and flow id */ -#define RX_CTRL_FIFO_DATA_HI_CTRL 0x0001 /* upper bit of ctrl word */ -#define RX_CTRL_FIFO_DATA_HI_FLOW_MASK 0x007E /* flow id */ - -/* diagnostic access to RX IPP FIFO. same semantics as RX_FIFO. - * DEFAULT: undefined - */ -#define REG_RX_IPP_FIFO_ADDR 0x4104 /* RX IPP FIFO address */ -#define REG_RX_IPP_FIFO_TAG 0x4108 /* RX IPP FIFO tag */ -#define REG_RX_IPP_FIFO_DATA_LOW 0x410C /* RX IPP FIFO data low */ -#define REG_RX_IPP_FIFO_DATA_HI_T0 0x4110 /* RX IPP FIFO data high - T0 */ -#define REG_RX_IPP_FIFO_DATA_HI_T1 0x4114 /* RX IPP FIFO data high - T1 */ - -/* 64-bit pointer to receive data buffer in host memory used for headers and - * small packets. MSB in high register. loaded by DMA state machine and - * increments as DMA writes receive data. only 50 LSB are incremented. top - * 13 bits taken from RX descriptor. - * DEFAULT: undefined - */ -#define REG_RX_HEADER_PAGE_PTR_LOW 0x4118 /* (ro) RX header page ptr - low */ -#define REG_RX_HEADER_PAGE_PTR_HI 0x411C /* (ro) RX header page ptr - high */ -#define REG_RX_MTU_PAGE_PTR_LOW 0x4120 /* (ro) RX MTU page pointer - low */ -#define REG_RX_MTU_PAGE_PTR_HI 0x4124 /* (ro) RX MTU page pointer - high */ - -/* PIO diagnostic access to RX reassembly DMA Table RAM. 6-bit register holds - * one of 64 79-bit locations in the RX Reassembly DMA table and the addr of - * one of the 64 byte locations in the Batching table. LOW holds 32 LSB. - * MID holds the next 32 LSB. HIGH holds the 15 MSB. RX_DMA_EN must be set - * to 0 for PIO access. DATA_HIGH should be last write of write sequence. - * layout: - * reassmbl ptr [78:15] | reassmbl index [14:1] | reassmbl entry valid [0] - * DEFAULT: undefined - */ -#define REG_RX_TABLE_ADDR 0x4128 /* RX reassembly DMA table - address */ -#define RX_TABLE_ADDR_MASK 0x0000003F /* address mask */ - -#define REG_RX_TABLE_DATA_LOW 0x412C /* RX reassembly DMA table - data low */ -#define REG_RX_TABLE_DATA_MID 0x4130 /* RX reassembly DMA table - data mid */ -#define REG_RX_TABLE_DATA_HI 0x4134 /* RX reassembly DMA table - data high */ - -/* cassini+ only */ -/* 8KB aligned 64-bit pointer to base of RX rings. lower 13 bits hardwired to - * 0. same semantics as primary desc/complete rings. - */ -#define REG_PLUS_RX_DB1_LOW 0x4200 /* RX descriptor ring - 2 base low */ -#define REG_PLUS_RX_DB1_HI 0x4204 /* RX descriptor ring - 2 base high */ -#define REG_PLUS_RX_CB1_LOW 0x4208 /* RX completion ring - 2 base low. 4 total */ -#define REG_PLUS_RX_CB1_HI 0x420C /* RX completion ring - 2 base high. 4 total */ -#define REG_PLUS_RX_CBN_LOW(x) (REG_PLUS_RX_CB1_LOW + 8*((x) - 1)) -#define REG_PLUS_RX_CBN_HI(x) (REG_PLUS_RX_CB1_HI + 8*((x) - 1)) -#define REG_PLUS_RX_KICK1 0x4220 /* RX Kick 2 register */ -#define REG_PLUS_RX_COMP1 0x4224 /* (ro) RX completion 2 - reg */ -#define REG_PLUS_RX_COMP1_HEAD 0x4228 /* (ro) RX completion 2 - head reg. 4 total. */ -#define REG_PLUS_RX_COMP1_TAIL 0x422C /* RX completion 2 - tail reg. 4 total. */ -#define REG_PLUS_RX_COMPN_HEAD(x) (REG_PLUS_RX_COMP1_HEAD + 8*((x) - 1)) -#define REG_PLUS_RX_COMPN_TAIL(x) (REG_PLUS_RX_COMP1_TAIL + 8*((x) - 1)) -#define REG_PLUS_RX_AE1_THRESH 0x4240 /* RX almost empty 2 - thresholds */ -#define RX_AE1_THRESH_FREE_MASK RX_AE_THRESH_FREE_MASK -#define RX_AE1_THRESH_FREE_SHIFT RX_AE_THRESH_FREE_SHIFT - -/** header parser registers **/ - -/* RX parser configuration register. - * DEFAULT: 0x1651004 - */ -#define REG_HP_CFG 0x4140 /* header parser - configuration reg */ -#define HP_CFG_PARSE_EN 0x00000001 /* enab header parsing */ -#define HP_CFG_NUM_CPU_MASK 0x000000FC /* # processors - 0 = 64. 0x3f = 63 */ -#define HP_CFG_NUM_CPU_SHIFT 2 -#define HP_CFG_SYN_INC_MASK 0x00000100 /* SYN bit won't increment - TCP seq # by one when - stored in FDBM */ -#define HP_CFG_TCP_THRESH_MASK 0x000FFE00 /* # bytes of TCP data - needed to be considered - for reassembly */ -#define HP_CFG_TCP_THRESH_SHIFT 9 - -/* access to RX Instruction RAM. 5-bit register/counter holds addr - * of 39 bit entry to be read/written. 32 LSB in _DATA_LOW. 7 MSB in _DATA_HI. - * RX_DMA_EN must be 0 for RX instr PIO access. DATA_HI should be last access - * of sequence. - * DEFAULT: undefined - */ -#define REG_HP_INSTR_RAM_ADDR 0x4144 /* HP instruction RAM - address */ -#define HP_INSTR_RAM_ADDR_MASK 0x01F /* 5-bit mask */ -#define REG_HP_INSTR_RAM_DATA_LOW 0x4148 /* HP instruction RAM - data low */ -#define HP_INSTR_RAM_LOW_OUTMASK_MASK 0x0000FFFF -#define HP_INSTR_RAM_LOW_OUTMASK_SHIFT 0 -#define HP_INSTR_RAM_LOW_OUTSHIFT_MASK 0x000F0000 -#define HP_INSTR_RAM_LOW_OUTSHIFT_SHIFT 16 -#define HP_INSTR_RAM_LOW_OUTEN_MASK 0x00300000 -#define HP_INSTR_RAM_LOW_OUTEN_SHIFT 20 -#define HP_INSTR_RAM_LOW_OUTARG_MASK 0xFFC00000 -#define HP_INSTR_RAM_LOW_OUTARG_SHIFT 22 -#define REG_HP_INSTR_RAM_DATA_MID 0x414C /* HP instruction RAM - data mid */ -#define HP_INSTR_RAM_MID_OUTARG_MASK 0x00000003 -#define HP_INSTR_RAM_MID_OUTARG_SHIFT 0 -#define HP_INSTR_RAM_MID_OUTOP_MASK 0x0000003C -#define HP_INSTR_RAM_MID_OUTOP_SHIFT 2 -#define HP_INSTR_RAM_MID_FNEXT_MASK 0x000007C0 -#define HP_INSTR_RAM_MID_FNEXT_SHIFT 6 -#define HP_INSTR_RAM_MID_FOFF_MASK 0x0003F800 -#define HP_INSTR_RAM_MID_FOFF_SHIFT 11 -#define HP_INSTR_RAM_MID_SNEXT_MASK 0x007C0000 -#define HP_INSTR_RAM_MID_SNEXT_SHIFT 18 -#define HP_INSTR_RAM_MID_SOFF_MASK 0x3F800000 -#define HP_INSTR_RAM_MID_SOFF_SHIFT 23 -#define HP_INSTR_RAM_MID_OP_MASK 0xC0000000 -#define HP_INSTR_RAM_MID_OP_SHIFT 30 -#define REG_HP_INSTR_RAM_DATA_HI 0x4150 /* HP instruction RAM - data high */ -#define HP_INSTR_RAM_HI_VAL_MASK 0x0000FFFF -#define HP_INSTR_RAM_HI_VAL_SHIFT 0 -#define HP_INSTR_RAM_HI_MASK_MASK 0xFFFF0000 -#define HP_INSTR_RAM_HI_MASK_SHIFT 16 - -/* PIO access into RX Header parser data RAM and flow database. - * 11-bit register. Data fills the LSB portion of bus if less than 32 bits. - * DATA_RAM: write RAM_FDB_DATA with index to access DATA_RAM. - * RAM bytes = 4*(x - 1) + [3:0]. e.g., 0 -> [3:0], 31 -> [123:120] - * FLOWDB: write DATA_RAM_FDB register and then read/write FDB1-12 to access - * flow database. - * RX_DMA_EN must be 0 for RX parser RAM PIO access. RX Parser RAM data reg - * should be the last write access of the write sequence. - * DEFAULT: undefined - */ -#define REG_HP_DATA_RAM_FDB_ADDR 0x4154 /* HP data and FDB - RAM address */ -#define HP_DATA_RAM_FDB_DATA_MASK 0x001F /* select 1 of 86 byte - locations in header - parser data ram to - read/write */ -#define HP_DATA_RAM_FDB_FDB_MASK 0x3F00 /* 1 of 64 353-bit locations - in the flow database */ -#define REG_HP_DATA_RAM_DATA 0x4158 /* HP data RAM data */ - -/* HP flow database registers: 1 - 12, 0x415C - 0x4188, 4 8-bit bytes - * FLOW_DB(1) = IP_SA[127:96], FLOW_DB(2) = IP_SA[95:64] - * FLOW_DB(3) = IP_SA[63:32], FLOW_DB(4) = IP_SA[31:0] - * FLOW_DB(5) = IP_DA[127:96], FLOW_DB(6) = IP_DA[95:64] - * FLOW_DB(7) = IP_DA[63:32], FLOW_DB(8) = IP_DA[31:0] - * FLOW_DB(9) = {TCP_SP[15:0],TCP_DP[15:0]} - * FLOW_DB(10) = bit 0 has value for flow valid - * FLOW_DB(11) = TCP_SEQ[63:32], FLOW_DB(12) = TCP_SEQ[31:0] - */ -#define REG_HP_FLOW_DB0 0x415C /* HP flow database 1 reg */ -#define REG_HP_FLOW_DBN(x) (REG_HP_FLOW_DB0 + (x)*4) - -/* diagnostics for RX Header Parser block. - * ASUN: the header parser state machine register is used for diagnostics - * purposes. however, the spec doesn't have any details on it. - */ -#define REG_HP_STATE_MACHINE 0x418C /* (ro) HP state machine */ -#define REG_HP_STATUS0 0x4190 /* (ro) HP status 1 */ -#define HP_STATUS0_SAP_MASK 0xFFFF0000 /* SAP */ -#define HP_STATUS0_L3_OFF_MASK 0x0000FE00 /* L3 offset */ -#define HP_STATUS0_LB_CPUNUM_MASK 0x000001F8 /* load balancing CPU - number */ -#define HP_STATUS0_HRP_OPCODE_MASK 0x00000007 /* HRP opcode */ - -#define REG_HP_STATUS1 0x4194 /* (ro) HP status 2 */ -#define HP_STATUS1_ACCUR2_MASK 0xE0000000 /* accu R2[6:4] */ -#define HP_STATUS1_FLOWID_MASK 0x1F800000 /* flow id */ -#define HP_STATUS1_TCP_OFF_MASK 0x007F0000 /* tcp payload offset */ -#define HP_STATUS1_TCP_SIZE_MASK 0x0000FFFF /* tcp payload size */ - -#define REG_HP_STATUS2 0x4198 /* (ro) HP status 3 */ -#define HP_STATUS2_ACCUR2_MASK 0xF0000000 /* accu R2[3:0] */ -#define HP_STATUS2_CSUM_OFF_MASK 0x07F00000 /* checksum start - start offset */ -#define HP_STATUS2_ACCUR1_MASK 0x000FE000 /* accu R1 */ -#define HP_STATUS2_FORCE_DROP 0x00001000 /* force drop */ -#define HP_STATUS2_BWO_REASSM 0x00000800 /* batching w/o - reassembly */ -#define HP_STATUS2_JH_SPLIT_EN 0x00000400 /* jumbo header split - enable */ -#define HP_STATUS2_FORCE_TCP_NOCHECK 0x00000200 /* force tcp no payload - check */ -#define HP_STATUS2_DATA_MASK_ZERO 0x00000100 /* mask of data length - equal to zero */ -#define HP_STATUS2_FORCE_TCP_CHECK 0x00000080 /* force tcp payload - chk */ -#define HP_STATUS2_MASK_TCP_THRESH 0x00000040 /* mask of payload - threshold */ -#define HP_STATUS2_NO_ASSIST 0x00000020 /* no assist */ -#define HP_STATUS2_CTRL_PACKET_FLAG 0x00000010 /* control packet flag */ -#define HP_STATUS2_TCP_FLAG_CHECK 0x00000008 /* tcp flag check */ -#define HP_STATUS2_SYN_FLAG 0x00000004 /* syn flag */ -#define HP_STATUS2_TCP_CHECK 0x00000002 /* tcp payload chk */ -#define HP_STATUS2_TCP_NOCHECK 0x00000001 /* tcp no payload chk */ - -/* BIST for header parser(HP) and flow database memories (FDBM). set _START - * to start BIST. controller clears _START on completion. _START can also - * be cleared to force termination of BIST. a bit set indicates that that - * memory passed its BIST. - */ -#define REG_HP_RAM_BIST 0x419C /* HP RAM BIST reg */ -#define HP_RAM_BIST_HP_DATA_PASS 0x80000000 /* HP data ram */ -#define HP_RAM_BIST_HP_INSTR0_PASS 0x40000000 /* HP instr ram 0 */ -#define HP_RAM_BIST_HP_INSTR1_PASS 0x20000000 /* HP instr ram 1 */ -#define HP_RAM_BIST_HP_INSTR2_PASS 0x10000000 /* HP instr ram 2 */ -#define HP_RAM_BIST_FDBM_AGE0_PASS 0x08000000 /* FDBM aging RAM0 */ -#define HP_RAM_BIST_FDBM_AGE1_PASS 0x04000000 /* FDBM aging RAM1 */ -#define HP_RAM_BIST_FDBM_FLOWID00_PASS 0x02000000 /* FDBM flowid RAM0 - bank 0 */ -#define HP_RAM_BIST_FDBM_FLOWID10_PASS 0x01000000 /* FDBM flowid RAM1 - bank 0 */ -#define HP_RAM_BIST_FDBM_FLOWID20_PASS 0x00800000 /* FDBM flowid RAM2 - bank 0 */ -#define HP_RAM_BIST_FDBM_FLOWID30_PASS 0x00400000 /* FDBM flowid RAM3 - bank 0 */ -#define HP_RAM_BIST_FDBM_FLOWID01_PASS 0x00200000 /* FDBM flowid RAM0 - bank 1 */ -#define HP_RAM_BIST_FDBM_FLOWID11_PASS 0x00100000 /* FDBM flowid RAM1 - bank 2 */ -#define HP_RAM_BIST_FDBM_FLOWID21_PASS 0x00080000 /* FDBM flowid RAM2 - bank 1 */ -#define HP_RAM_BIST_FDBM_FLOWID31_PASS 0x00040000 /* FDBM flowid RAM3 - bank 1 */ -#define HP_RAM_BIST_FDBM_TCPSEQ_PASS 0x00020000 /* FDBM tcp sequence - RAM */ -#define HP_RAM_BIST_SUMMARY 0x00000002 /* all BIST tests */ -#define HP_RAM_BIST_START 0x00000001 /* start/stop BIST */ - - -/** MAC registers. **/ -/* reset bits are set using a PIO write and self-cleared after the command - * execution has completed. - */ -#define REG_MAC_TX_RESET 0x6000 /* TX MAC software reset - command (default: 0x0) */ -#define REG_MAC_RX_RESET 0x6004 /* RX MAC software reset - command (default: 0x0) */ -/* execute a pause flow control frame transmission - DEFAULT: 0x0XXXX */ -#define REG_MAC_SEND_PAUSE 0x6008 /* send pause command reg */ -#define MAC_SEND_PAUSE_TIME_MASK 0x0000FFFF /* value of pause time - to be sent on network - in units of slot - times */ -#define MAC_SEND_PAUSE_SEND 0x00010000 /* send pause flow ctrl - frame on network */ - -/* bit set indicates that event occurred. auto-cleared when status register - * is read and have corresponding mask bits in mask register. events will - * trigger an interrupt if the corresponding mask bit is 0. - * status register default: 0x00000000 - * mask register default = 0xFFFFFFFF on reset - */ -#define REG_MAC_TX_STATUS 0x6010 /* TX MAC status reg */ -#define MAC_TX_FRAME_XMIT 0x0001 /* successful frame - transmision */ -#define MAC_TX_UNDERRUN 0x0002 /* terminated frame - transmission due to - data starvation in the - xmit data path */ -#define MAC_TX_MAX_PACKET_ERR 0x0004 /* frame exceeds max allowed - length passed to TX MAC - by the DMA engine */ -#define MAC_TX_COLL_NORMAL 0x0008 /* rollover of the normal - collision counter */ -#define MAC_TX_COLL_EXCESS 0x0010 /* rollover of the excessive - collision counter */ -#define MAC_TX_COLL_LATE 0x0020 /* rollover of the late - collision counter */ -#define MAC_TX_COLL_FIRST 0x0040 /* rollover of the first - collision counter */ -#define MAC_TX_DEFER_TIMER 0x0080 /* rollover of the defer - timer */ -#define MAC_TX_PEAK_ATTEMPTS 0x0100 /* rollover of the peak - attempts counter */ - -#define REG_MAC_RX_STATUS 0x6014 /* RX MAC status reg */ -#define MAC_RX_FRAME_RECV 0x0001 /* successful receipt of - a frame */ -#define MAC_RX_OVERFLOW 0x0002 /* dropped frame due to - RX FIFO overflow */ -#define MAC_RX_FRAME_COUNT 0x0004 /* rollover of receive frame - counter */ -#define MAC_RX_ALIGN_ERR 0x0008 /* rollover of alignment - error counter */ -#define MAC_RX_CRC_ERR 0x0010 /* rollover of crc error - counter */ -#define MAC_RX_LEN_ERR 0x0020 /* rollover of length - error counter */ -#define MAC_RX_VIOL_ERR 0x0040 /* rollover of code - violation error */ - -/* DEFAULT: 0xXXXX0000 on reset */ -#define REG_MAC_CTRL_STATUS 0x6018 /* MAC control status reg */ -#define MAC_CTRL_PAUSE_RECEIVED 0x00000001 /* successful - reception of a - pause control - frame */ -#define MAC_CTRL_PAUSE_STATE 0x00000002 /* MAC has made a - transition from - "not paused" to - "paused" */ -#define MAC_CTRL_NOPAUSE_STATE 0x00000004 /* MAC has made a - transition from - "paused" to "not - paused" */ -#define MAC_CTRL_PAUSE_TIME_MASK 0xFFFF0000 /* value of pause time - operand that was - received in the last - pause flow control - frame */ - -/* layout identical to TX MAC[8:0] */ -#define REG_MAC_TX_MASK 0x6020 /* TX MAC mask reg */ -/* layout identical to RX MAC[6:0] */ -#define REG_MAC_RX_MASK 0x6024 /* RX MAC mask reg */ -/* layout identical to CTRL MAC[2:0] */ -#define REG_MAC_CTRL_MASK 0x6028 /* MAC control mask reg */ - -/* to ensure proper operation, CFG_EN must be cleared to 0 and a delay - * imposed before writes to other bits in the TX_MAC_CFG register or any of - * the MAC parameters is performed. delay dependent upon time required to - * transmit a maximum size frame (= MAC_FRAMESIZE_MAX*8/Mbps). e.g., - * the delay for a 1518-byte frame on a 100Mbps network is 125us. - * alternatively, just poll TX_CFG_EN until it reads back as 0. - * NOTE: on half-duplex 1Gbps, TX_CFG_CARRIER_EXTEND and - * RX_CFG_CARRIER_EXTEND should be set and the SLOT_TIME register should - * be 0x200 (slot time of 512 bytes) - */ -#define REG_MAC_TX_CFG 0x6030 /* TX MAC config reg */ -#define MAC_TX_CFG_EN 0x0001 /* enable TX MAC. 0 will - force TXMAC state - machine to remain in - idle state or to - transition to idle state - on completion of an - ongoing packet. */ -#define MAC_TX_CFG_IGNORE_CARRIER 0x0002 /* disable CSMA/CD deferral - process. set to 1 when - full duplex and 0 when - half duplex */ -#define MAC_TX_CFG_IGNORE_COLL 0x0004 /* disable CSMA/CD backoff - algorithm. set to 1 when - full duplex and 0 when - half duplex */ -#define MAC_TX_CFG_IPG_EN 0x0008 /* enable extension of the - Rx-to-TX IPG. after - receiving a frame, TX - MAC will reset its - deferral process to - carrier sense for the - amount of time = IPG0 + - IPG1 and commit to - transmission for time - specified in IPG2. when - 0 or when xmitting frames - back-to-pack (Tx-to-Tx - IPG), TX MAC ignores - IPG0 and will only use - IPG1 for deferral time. - IPG2 still used. */ -#define MAC_TX_CFG_NEVER_GIVE_UP_EN 0x0010 /* TX MAC will not easily - give up on frame - xmission. if backoff - algorithm reaches the - ATTEMPT_LIMIT, it will - clear attempts counter - and continue trying to - send the frame as - specified by - GIVE_UP_LIM. when 0, - TX MAC will execute - standard CSMA/CD prot. */ -#define MAC_TX_CFG_NEVER_GIVE_UP_LIM 0x0020 /* when set, TX MAC will - continue to try to xmit - until successful. when - 0, TX MAC will continue - to try xmitting until - successful or backoff - algorithm reaches - ATTEMPT_LIMIT*16 */ -#define MAC_TX_CFG_NO_BACKOFF 0x0040 /* modify CSMA/CD to disable - backoff algorithm. TX - MAC will not back off - after a xmission attempt - that resulted in a - collision. */ -#define MAC_TX_CFG_SLOW_DOWN 0x0080 /* modify CSMA/CD so that - deferral process is reset - in response to carrier - sense during the entire - duration of IPG. TX MAC - will only commit to frame - xmission after frame - xmission has actually - begun. */ -#define MAC_TX_CFG_NO_FCS 0x0100 /* TX MAC will not generate - CRC for all xmitted - packets. when clear, CRC - generation is dependent - upon NO_CRC bit in the - xmit control word from - TX DMA */ -#define MAC_TX_CFG_CARRIER_EXTEND 0x0200 /* enables xmit part of the - carrier extension - feature. this allows for - longer collision domains - by extending the carrier - and collision window - from the end of FCS until - the end of the slot time - if necessary. Required - for half-duplex at 1Gbps, - clear otherwise. */ - -/* when CRC is not stripped, reassembly packets will not contain the CRC. - * these will be stripped by HRP because it reassembles layer 4 data, and the - * CRC is layer 2. however, non-reassembly packets will still contain the CRC - * when passed to the host. to ensure proper operation, need to wait 3.2ms - * after clearing RX_CFG_EN before writing to any other RX MAC registers - * or other MAC parameters. alternatively, poll RX_CFG_EN until it clears - * to 0. similary, HASH_FILTER_EN and ADDR_FILTER_EN have the same - * restrictions as CFG_EN. - */ -#define REG_MAC_RX_CFG 0x6034 /* RX MAC config reg */ -#define MAC_RX_CFG_EN 0x0001 /* enable RX MAC */ -#define MAC_RX_CFG_STRIP_PAD 0x0002 /* always program to 0. - feature not supported */ -#define MAC_RX_CFG_STRIP_FCS 0x0004 /* RX MAC will strip the - last 4 bytes of a - received frame. */ -#define MAC_RX_CFG_PROMISC_EN 0x0008 /* promiscuous mode */ -#define MAC_RX_CFG_PROMISC_GROUP_EN 0x0010 /* accept all valid - multicast frames (group - bit in DA field set) */ -#define MAC_RX_CFG_HASH_FILTER_EN 0x0020 /* use hash table to filter - multicast addresses */ -#define MAC_RX_CFG_ADDR_FILTER_EN 0x0040 /* cause RX MAC to use - address filtering regs - to filter both unicast - and multicast - addresses */ -#define MAC_RX_CFG_DISABLE_DISCARD 0x0080 /* pass errored frames to - RX DMA by setting BAD - bit but not Abort bit - in the status. CRC, - framing, and length errs - will not increment - error counters. frames - which don't match dest - addr will be passed up - w/ BAD bit set. */ -#define MAC_RX_CFG_CARRIER_EXTEND 0x0100 /* enable reception of - packet bursts generated - by carrier extension - with packet bursting - senders. only applies - to half-duplex 1Gbps */ - -/* DEFAULT: 0x0 */ -#define REG_MAC_CTRL_CFG 0x6038 /* MAC control config reg */ -#define MAC_CTRL_CFG_SEND_PAUSE_EN 0x0001 /* respond to requests for - sending pause flow ctrl - frames */ -#define MAC_CTRL_CFG_RECV_PAUSE_EN 0x0002 /* respond to received - pause flow ctrl frames */ -#define MAC_CTRL_CFG_PASS_CTRL 0x0004 /* pass valid MAC ctrl - packets to RX DMA */ - -/* to ensure proper operation, a global initialization sequence should be - * performed when a loopback config is entered or exited. if programmed after - * a hw or global sw reset, RX/TX MAC software reset and initialization - * should be done to ensure stable clocking. - * DEFAULT: 0x0 - */ -#define REG_MAC_XIF_CFG 0x603C /* XIF config reg */ -#define MAC_XIF_TX_MII_OUTPUT_EN 0x0001 /* enable output drivers - on MII xmit bus */ -#define MAC_XIF_MII_INT_LOOPBACK 0x0002 /* loopback GMII xmit data - path to GMII recv data - path. phy mode register - clock selection must be - set to GMII mode and - GMII_MODE should be set - to 1. in loopback mode, - REFCLK will drive the - entire mac core. 0 for - normal operation. */ -#define MAC_XIF_DISABLE_ECHO 0x0004 /* disables receive data - path during packet - xmission. clear to 0 - in any full duplex mode, - in any loopback mode, - or in half-duplex SERDES - or SLINK modes. set when - in half-duplex when - using external phy. */ -#define MAC_XIF_GMII_MODE 0x0008 /* MAC operates with GMII - clocks and datapath */ -#define MAC_XIF_MII_BUFFER_OUTPUT_EN 0x0010 /* MII_BUF_EN pin. enable - external tristate buffer - on the MII receive - bus. */ -#define MAC_XIF_LINK_LED 0x0020 /* LINKLED# active (low) */ -#define MAC_XIF_FDPLX_LED 0x0040 /* FDPLXLED# active (low) */ - -#define REG_MAC_IPG0 0x6040 /* inter-packet gap0 reg. - recommended: 0x00 */ -#define REG_MAC_IPG1 0x6044 /* inter-packet gap1 reg - recommended: 0x08 */ -#define REG_MAC_IPG2 0x6048 /* inter-packet gap2 reg - recommended: 0x04 */ -#define REG_MAC_SLOT_TIME 0x604C /* slot time reg - recommended: 0x40 */ -#define REG_MAC_FRAMESIZE_MIN 0x6050 /* min frame size reg - recommended: 0x40 */ - -/* FRAMESIZE_MAX holds both the max frame size as well as the max burst size. - * recommended value: 0x2000.05EE - */ -#define REG_MAC_FRAMESIZE_MAX 0x6054 /* max frame size reg */ -#define MAC_FRAMESIZE_MAX_BURST_MASK 0x3FFF0000 /* max burst size */ -#define MAC_FRAMESIZE_MAX_BURST_SHIFT 16 -#define MAC_FRAMESIZE_MAX_FRAME_MASK 0x00007FFF /* max frame size */ -#define MAC_FRAMESIZE_MAX_FRAME_SHIFT 0 -#define REG_MAC_PA_SIZE 0x6058 /* PA size reg. number of - preamble bytes that the - TX MAC will xmit at the - beginning of each frame - value should be 2 or - greater. recommended - value: 0x07 */ -#define REG_MAC_JAM_SIZE 0x605C /* jam size reg. duration - of jam in units of media - byte time. recommended - value: 0x04 */ -#define REG_MAC_ATTEMPT_LIMIT 0x6060 /* attempt limit reg. # - of attempts TX MAC will - make to xmit a frame - before it resets its - attempts counter. after - the limit has been - reached, TX MAC may or - may not drop the frame - dependent upon value - in TX_MAC_CFG. - recommended - value: 0x10 */ -#define REG_MAC_CTRL_TYPE 0x6064 /* MAC control type reg. - type field of a MAC - ctrl frame. recommended - value: 0x8808 */ - -/* mac address registers: 0 - 44, 0x6080 - 0x6130, 4 8-bit bytes. - * register contains comparison - * 0 16 MSB of primary MAC addr [47:32] of DA field - * 1 16 middle bits "" [31:16] of DA field - * 2 16 LSB "" [15:0] of DA field - * 3*x 16MSB of alt MAC addr 1-15 [47:32] of DA field - * 4*x 16 middle bits "" [31:16] - * 5*x 16 LSB "" [15:0] - * 42 16 MSB of MAC CTRL addr [47:32] of DA. - * 43 16 middle bits "" [31:16] - * 44 16 LSB "" [15:0] - * MAC CTRL addr must be the reserved multicast addr for MAC CTRL frames. - * if there is a match, MAC will set the bit for alternative address - * filter pass [15] - - * here is the map of registers given MAC address notation: a:b:c:d:e:f - * ab cd ef - * primary addr reg 2 reg 1 reg 0 - * alt addr 1 reg 5 reg 4 reg 3 - * alt addr x reg 5*x reg 4*x reg 3*x - * ctrl addr reg 44 reg 43 reg 42 - */ -#define REG_MAC_ADDR0 0x6080 /* MAC address 0 reg */ -#define REG_MAC_ADDRN(x) (REG_MAC_ADDR0 + (x)*4) -#define REG_MAC_ADDR_FILTER0 0x614C /* address filter 0 reg - [47:32] */ -#define REG_MAC_ADDR_FILTER1 0x6150 /* address filter 1 reg - [31:16] */ -#define REG_MAC_ADDR_FILTER2 0x6154 /* address filter 2 reg - [15:0] */ -#define REG_MAC_ADDR_FILTER2_1_MASK 0x6158 /* address filter 2 and 1 - mask reg. 8-bit reg - contains nibble mask for - reg 2 and 1. */ -#define REG_MAC_ADDR_FILTER0_MASK 0x615C /* address filter 0 mask - reg */ - -/* hash table registers: 0 - 15, 0x6160 - 0x619C, 4 8-bit bytes - * 16-bit registers contain bits of the hash table. - * reg x -> [16*(15 - x) + 15 : 16*(15 - x)]. - * e.g., 15 -> [15:0], 0 -> [255:240] - */ -#define REG_MAC_HASH_TABLE0 0x6160 /* hash table 0 reg */ -#define REG_MAC_HASH_TABLEN(x) (REG_MAC_HASH_TABLE0 + (x)*4) - -/* statistics registers. these registers generate an interrupt on - * overflow. recommended initialization: 0x0000. most are 16-bits except - * for PEAK_ATTEMPTS register which is 8 bits. - */ -#define REG_MAC_COLL_NORMAL 0x61A0 /* normal collision - counter. */ -#define REG_MAC_COLL_FIRST 0x61A4 /* first attempt - successful collision - counter */ -#define REG_MAC_COLL_EXCESS 0x61A8 /* excessive collision - counter */ -#define REG_MAC_COLL_LATE 0x61AC /* late collision counter */ -#define REG_MAC_TIMER_DEFER 0x61B0 /* defer timer. time base - is the media byte - clock/256 */ -#define REG_MAC_ATTEMPTS_PEAK 0x61B4 /* peak attempts reg */ -#define REG_MAC_RECV_FRAME 0x61B8 /* receive frame counter */ -#define REG_MAC_LEN_ERR 0x61BC /* length error counter */ -#define REG_MAC_ALIGN_ERR 0x61C0 /* alignment error counter */ -#define REG_MAC_FCS_ERR 0x61C4 /* FCS error counter */ -#define REG_MAC_RX_CODE_ERR 0x61C8 /* RX code violation - error counter */ - -/* misc registers */ -#define REG_MAC_RANDOM_SEED 0x61CC /* random number seed reg. - 10-bit register used as a - seed for the random number - generator for the CSMA/CD - backoff algorithm. only - programmed after power-on - reset and should be a - random value which has a - high likelihood of being - unique for each MAC - attached to a network - segment (e.g., 10 LSB of - MAC address) */ - -/* ASUN: there's a PAUSE_TIMER (ro) described, but it's not in the address - * map - */ - -/* 27-bit register has the current state for key state machines in the MAC */ -#define REG_MAC_STATE_MACHINE 0x61D0 /* (ro) state machine reg */ -#define MAC_SM_RLM_MASK 0x07800000 -#define MAC_SM_RLM_SHIFT 23 -#define MAC_SM_RX_FC_MASK 0x00700000 -#define MAC_SM_RX_FC_SHIFT 20 -#define MAC_SM_TLM_MASK 0x000F0000 -#define MAC_SM_TLM_SHIFT 16 -#define MAC_SM_ENCAP_SM_MASK 0x0000F000 -#define MAC_SM_ENCAP_SM_SHIFT 12 -#define MAC_SM_TX_REQ_MASK 0x00000C00 -#define MAC_SM_TX_REQ_SHIFT 10 -#define MAC_SM_TX_FC_MASK 0x000003C0 -#define MAC_SM_TX_FC_SHIFT 6 -#define MAC_SM_FIFO_WRITE_SEL_MASK 0x00000038 -#define MAC_SM_FIFO_WRITE_SEL_SHIFT 3 -#define MAC_SM_TX_FIFO_EMPTY_MASK 0x00000007 -#define MAC_SM_TX_FIFO_EMPTY_SHIFT 0 - -/** MIF registers. the MIF can be programmed in either bit-bang or - * frame mode. - **/ -#define REG_MIF_BIT_BANG_CLOCK 0x6200 /* MIF bit-bang clock. - 1 -> 0 will generate a - rising edge. 0 -> 1 will - generate a falling edge. */ -#define REG_MIF_BIT_BANG_DATA 0x6204 /* MIF bit-bang data. 1-bit - register generates data */ -#define REG_MIF_BIT_BANG_OUTPUT_EN 0x6208 /* MIF bit-bang output - enable. enable when - xmitting data from MIF to - transceiver. */ - -/* 32-bit register serves as an instruction register when the MIF is - * programmed in frame mode. load this register w/ a valid instruction - * (as per IEEE 802.3u MII spec). poll this register to check for instruction - * execution completion. during a read operation, this register will also - * contain the 16-bit data returned by the tranceiver. unless specified - * otherwise, fields are considered "don't care" when polling for - * completion. - */ -#define REG_MIF_FRAME 0x620C /* MIF frame/output reg */ -#define MIF_FRAME_START_MASK 0xC0000000 /* start of frame. - load w/ 01 when - issuing an instr */ -#define MIF_FRAME_ST 0x40000000 /* STart of frame */ -#define MIF_FRAME_OPCODE_MASK 0x30000000 /* opcode. 01 for a - write. 10 for a - read */ -#define MIF_FRAME_OP_READ 0x20000000 /* read OPcode */ -#define MIF_FRAME_OP_WRITE 0x10000000 /* write OPcode */ -#define MIF_FRAME_PHY_ADDR_MASK 0x0F800000 /* phy address. when - issuing an instr, - this field should be - loaded w/ the XCVR - addr */ -#define MIF_FRAME_PHY_ADDR_SHIFT 23 -#define MIF_FRAME_REG_ADDR_MASK 0x007C0000 /* register address. - when issuing an instr, - addr of register - to be read/written */ -#define MIF_FRAME_REG_ADDR_SHIFT 18 -#define MIF_FRAME_TURN_AROUND_MSB 0x00020000 /* turn around, MSB. - when issuing an instr, - set this bit to 1 */ -#define MIF_FRAME_TURN_AROUND_LSB 0x00010000 /* turn around, LSB. - when issuing an instr, - set this bit to 0. - when polling for - completion, 1 means - that instr execution - has been completed */ -#define MIF_FRAME_DATA_MASK 0x0000FFFF /* instruction payload - load with 16-bit data - to be written in - transceiver reg for a - write. doesn't matter - in a read. when - polling for - completion, field is - "don't care" for write - and 16-bit data - returned by the - transceiver for a - read (if valid bit - is set) */ -#define REG_MIF_CFG 0x6210 /* MIF config reg */ -#define MIF_CFG_PHY_SELECT 0x0001 /* 1 -> select MDIO_1 - 0 -> select MDIO_0 */ -#define MIF_CFG_POLL_EN 0x0002 /* enable polling - mechanism. if set, - BB_MODE should be 0 */ -#define MIF_CFG_BB_MODE 0x0004 /* 1 -> bit-bang mode - 0 -> frame mode */ -#define MIF_CFG_POLL_REG_MASK 0x00F8 /* register address to be - used by polling mode. - only meaningful if POLL_EN - is set to 1 */ -#define MIF_CFG_POLL_REG_SHIFT 3 -#define MIF_CFG_MDIO_0 0x0100 /* (ro) dual purpose. - when MDIO_0 is idle, - 1 -> tranceiver is - connected to MDIO_0. - when MIF is communicating - w/ MDIO_0 in bit-bang - mode, this bit indicates - the incoming bit stream - during a read op */ -#define MIF_CFG_MDIO_1 0x0200 /* (ro) dual purpose. - when MDIO_1 is idle, - 1 -> transceiver is - connected to MDIO_1. - when MIF is communicating - w/ MDIO_1 in bit-bang - mode, this bit indicates - the incoming bit stream - during a read op */ -#define MIF_CFG_POLL_PHY_MASK 0x7C00 /* tranceiver address to - be polled */ -#define MIF_CFG_POLL_PHY_SHIFT 10 - -/* 16-bit register used to determine which bits in the POLL_STATUS portion of - * the MIF_STATUS register will cause an interrupt. if a mask bit is 0, - * corresponding bit of the POLL_STATUS will generate a MIF interrupt when - * set. DEFAULT: 0xFFFF - */ -#define REG_MIF_MASK 0x6214 /* MIF mask reg */ - -/* 32-bit register used when in poll mode. auto-cleared after being read */ -#define REG_MIF_STATUS 0x6218 /* MIF status reg */ -#define MIF_STATUS_POLL_DATA_MASK 0xFFFF0000 /* poll data contains - the "latest image" - update of the XCVR - reg being read */ -#define MIF_STATUS_POLL_DATA_SHIFT 16 -#define MIF_STATUS_POLL_STATUS_MASK 0x0000FFFF /* poll status indicates - which bits in the - POLL_DATA field have - changed since the - MIF_STATUS reg was - last read */ -#define MIF_STATUS_POLL_STATUS_SHIFT 0 - -/* 7-bit register has current state for all state machines in the MIF */ -#define REG_MIF_STATE_MACHINE 0x621C /* MIF state machine reg */ -#define MIF_SM_CONTROL_MASK 0x07 /* control state machine - state */ -#define MIF_SM_EXECUTION_MASK 0x60 /* execution state machine - state */ - -/** PCS/Serialink. the following registers are equivalent to the standard - * MII management registers except that they're directly mapped in - * Cassini's register space. - **/ - -/* the auto-negotiation enable bit should be programmed the same at - * the link partner as in the local device to enable auto-negotiation to - * complete. when that bit is reprogrammed, auto-neg/manual config is - * restarted automatically. - * DEFAULT: 0x1040 - */ -#define REG_PCS_MII_CTRL 0x9000 /* PCS MII control reg */ -#define PCS_MII_CTRL_1000_SEL 0x0040 /* reads 1. ignored on - writes */ -#define PCS_MII_CTRL_COLLISION_TEST 0x0080 /* COL signal at the PCS - to MAC interface is - activated regardless - of activity */ -#define PCS_MII_CTRL_DUPLEX 0x0100 /* forced 0x0. PCS - behaviour same for - half and full dplx */ -#define PCS_MII_RESTART_AUTONEG 0x0200 /* self clearing. - restart auto- - negotiation */ -#define PCS_MII_ISOLATE 0x0400 /* read as 0. ignored - on writes */ -#define PCS_MII_POWER_DOWN 0x0800 /* read as 0. ignored - on writes */ -#define PCS_MII_AUTONEG_EN 0x1000 /* default 1. PCS goes - through automatic - link config before it - can be used. when 0, - link can be used - w/out any link config - phase */ -#define PCS_MII_10_100_SEL 0x2000 /* read as 0. ignored on - writes */ -#define PCS_MII_RESET 0x8000 /* reset PCS. self-clears - when done */ - -/* DEFAULT: 0x0108 */ -#define REG_PCS_MII_STATUS 0x9004 /* PCS MII status reg */ -#define PCS_MII_STATUS_EXTEND_CAP 0x0001 /* reads 0 */ -#define PCS_MII_STATUS_JABBER_DETECT 0x0002 /* reads 0 */ -#define PCS_MII_STATUS_LINK_STATUS 0x0004 /* 1 -> link up. - 0 -> link down. 0 is - latched so that 0 is - kept until read. read - 2x to determine if the - link has gone up again */ -#define PCS_MII_STATUS_AUTONEG_ABLE 0x0008 /* reads 1 (able to perform - auto-neg) */ -#define PCS_MII_STATUS_REMOTE_FAULT 0x0010 /* 1 -> remote fault detected - from received link code - word. only valid after - auto-neg completed */ -#define PCS_MII_STATUS_AUTONEG_COMP 0x0020 /* 1 -> auto-negotiation - completed - 0 -> auto-negotiation not - completed */ -#define PCS_MII_STATUS_EXTEND_STATUS 0x0100 /* reads as 1. used as an - indication that this is - a 1000 Base-X PHY. writes - to it are ignored */ - -/* used during auto-negotiation. - * DEFAULT: 0x00E0 - */ -#define REG_PCS_MII_ADVERT 0x9008 /* PCS MII advertisement - reg */ -#define PCS_MII_ADVERT_FD 0x0020 /* advertise full duplex - 1000 Base-X */ -#define PCS_MII_ADVERT_HD 0x0040 /* advertise half-duplex - 1000 Base-X */ -#define PCS_MII_ADVERT_SYM_PAUSE 0x0080 /* advertise PAUSE - symmetric capability */ -#define PCS_MII_ADVERT_ASYM_PAUSE 0x0100 /* advertises PAUSE - asymmetric capability */ -#define PCS_MII_ADVERT_RF_MASK 0x3000 /* remote fault. write bit13 - to optionally indicate to - link partner that chip is - going off-line. bit12 will - get set when signal - detect == FAIL and will - remain set until - successful negotiation */ -#define PCS_MII_ADVERT_ACK 0x4000 /* (ro) */ -#define PCS_MII_ADVERT_NEXT_PAGE 0x8000 /* (ro) forced 0x0 */ - -/* contents updated as a result of autonegotiation. layout and definitions - * identical to PCS_MII_ADVERT - */ -#define REG_PCS_MII_LPA 0x900C /* PCS MII link partner - ability reg */ -#define PCS_MII_LPA_FD PCS_MII_ADVERT_FD -#define PCS_MII_LPA_HD PCS_MII_ADVERT_HD -#define PCS_MII_LPA_SYM_PAUSE PCS_MII_ADVERT_SYM_PAUSE -#define PCS_MII_LPA_ASYM_PAUSE PCS_MII_ADVERT_ASYM_PAUSE -#define PCS_MII_LPA_RF_MASK PCS_MII_ADVERT_RF_MASK -#define PCS_MII_LPA_ACK PCS_MII_ADVERT_ACK -#define PCS_MII_LPA_NEXT_PAGE PCS_MII_ADVERT_NEXT_PAGE - -/* DEFAULT: 0x0 */ -#define REG_PCS_CFG 0x9010 /* PCS config reg */ -#define PCS_CFG_EN 0x01 /* enable PCS. must be - 0 when modifying - PCS_MII_ADVERT */ -#define PCS_CFG_SD_OVERRIDE 0x02 /* sets signal detect to - OK. bit is - non-resettable */ -#define PCS_CFG_SD_ACTIVE_LOW 0x04 /* changes interpretation - of optical signal to make - signal detect okay when - signal is low */ -#define PCS_CFG_JITTER_STUDY_MASK 0x18 /* used to make jitter - measurements. a single - code group is xmitted - regularly. - 0x0 = normal operation - 0x1 = high freq test - pattern, D21.5 - 0x2 = low freq test - pattern, K28.7 - 0x3 = reserved */ -#define PCS_CFG_10MS_TIMER_OVERRIDE 0x20 /* shortens 10-20ms auto- - negotiation timer to - a few cycles for test - purposes */ - -/* used for diagnostic purposes. bits 20-22 autoclear on read */ -#define REG_PCS_STATE_MACHINE 0x9014 /* (ro) PCS state machine - and diagnostic reg */ -#define PCS_SM_TX_STATE_MASK 0x0000000F /* 0 and 1 indicate - xmission of idle. - otherwise, xmission of - a packet */ -#define PCS_SM_RX_STATE_MASK 0x000000F0 /* 0 indicates reception - of idle. otherwise, - reception of packet */ -#define PCS_SM_WORD_SYNC_STATE_MASK 0x00000700 /* 0 indicates loss of - sync */ -#define PCS_SM_SEQ_DETECT_STATE_MASK 0x00001800 /* cycling through 0-3 - indicates reception of - Config codes. cycling - through 0-1 indicates - reception of idles */ -#define PCS_SM_LINK_STATE_MASK 0x0001E000 -#define SM_LINK_STATE_UP 0x00016000 /* link state is up */ - -#define PCS_SM_LOSS_LINK_C 0x00100000 /* loss of link due to - recept of Config - codes */ -#define PCS_SM_LOSS_LINK_SYNC 0x00200000 /* loss of link due to - loss of sync */ -#define PCS_SM_LOSS_SIGNAL_DETECT 0x00400000 /* signal detect goes - from OK to FAIL. bit29 - will also be set if - this is set */ -#define PCS_SM_NO_LINK_BREAKLINK 0x01000000 /* link not up due to - receipt of breaklink - C codes from partner. - C codes w/ 0 content - received triggering - start/restart of - autonegotiation. - should be sent for - no longer than 20ms */ -#define PCS_SM_NO_LINK_SERDES 0x02000000 /* serdes being - initialized. see serdes - state reg */ -#define PCS_SM_NO_LINK_C 0x04000000 /* C codes not stable or - not received */ -#define PCS_SM_NO_LINK_SYNC 0x08000000 /* word sync not - achieved */ -#define PCS_SM_NO_LINK_WAIT_C 0x10000000 /* waiting for C codes - w/ ack bit set */ -#define PCS_SM_NO_LINK_NO_IDLE 0x20000000 /* link partner continues - to send C codes - instead of idle - symbols or pkt data */ - -/* this register indicates interrupt changes in specific PCS MII status bits. - * PCS_INT may be masked at the ISR level. only a single bit is implemented - * for link status change. - */ -#define REG_PCS_INTR_STATUS 0x9018 /* PCS interrupt status */ -#define PCS_INTR_STATUS_LINK_CHANGE 0x04 /* link status has changed - since last read */ - -/* control which network interface is used. no more than one bit should - * be set. - * DEFAULT: none - */ -#define REG_PCS_DATAPATH_MODE 0x9050 /* datapath mode reg */ -#define PCS_DATAPATH_MODE_MII 0x00 /* PCS is not used and - MII/GMII is selected. - selection between MII and - GMII is controlled by - XIF_CFG */ -#define PCS_DATAPATH_MODE_SERDES 0x02 /* PCS is used via the - 10-bit interface */ - -/* input to serdes chip or serialink block */ -#define REG_PCS_SERDES_CTRL 0x9054 /* serdes control reg */ -#define PCS_SERDES_CTRL_LOOPBACK 0x01 /* enable loopback on - serdes interface */ -#define PCS_SERDES_CTRL_SYNCD_EN 0x02 /* enable sync carrier - detection. should be - 0x0 for normal - operation */ -#define PCS_SERDES_CTRL_LOCKREF 0x04 /* frequency-lock RBC[0:1] - to REFCLK when set. - when clear, receiver - clock locks to incoming - serial data */ - -/* multiplex test outputs into the PROM address (PA_3 through PA_0) pins. - * should be 0x0 for normal operations. - * 0b000 normal operation, PROM address[3:0] selected - * 0b001 rxdma req, rxdma ack, rxdma ready, rxdma read - * 0b010 rxmac req, rx ack, rx tag, rx clk shared - * 0b011 txmac req, tx ack, tx tag, tx retry req - * 0b100 tx tp3, tx tp2, tx tp1, tx tp0 - * 0b101 R period RX, R period TX, R period HP, R period BIM - * DEFAULT: 0x0 - */ -#define REG_PCS_SHARED_OUTPUT_SEL 0x9058 /* shared output select */ -#define PCS_SOS_PROM_ADDR_MASK 0x0007 - -/* used for diagnostics. this register indicates progress of the SERDES - * boot up. - * 0b00 undergoing reset - * 0b01 waiting 500us while lockrefn is asserted - * 0b10 waiting for comma detect - * 0b11 receive data is synchronized - * DEFAULT: 0x0 - */ -#define REG_PCS_SERDES_STATE 0x905C /* (ro) serdes state */ -#define PCS_SERDES_STATE_MASK 0x03 - -/* used for diagnostics. indicates number of packets transmitted or received. - * counters rollover w/out generating an interrupt. - * DEFAULT: 0x0 - */ -#define REG_PCS_PACKET_COUNT 0x9060 /* (ro) PCS packet counter */ -#define PCS_PACKET_COUNT_TX 0x000007FF /* pkts xmitted by PCS */ -#define PCS_PACKET_COUNT_RX 0x07FF0000 /* pkts recvd by PCS - whether they - encountered an error - or not */ - -/** LocalBus Devices. the following provides run-time access to the - * Cassini's PROM - ***/ -#define REG_EXPANSION_ROM_RUN_START 0x100000 /* expansion rom run time - access */ -#define REG_EXPANSION_ROM_RUN_END 0x17FFFF - -#define REG_SECOND_LOCALBUS_START 0x180000 /* secondary local bus - device */ -#define REG_SECOND_LOCALBUS_END 0x1FFFFF - -/* entropy device */ -#define REG_ENTROPY_START REG_SECOND_LOCALBUS_START -#define REG_ENTROPY_DATA (REG_ENTROPY_START + 0x00) -#define REG_ENTROPY_STATUS (REG_ENTROPY_START + 0x04) -#define ENTROPY_STATUS_DRDY 0x01 -#define ENTROPY_STATUS_BUSY 0x02 -#define ENTROPY_STATUS_CIPHER 0x04 -#define ENTROPY_STATUS_BYPASS_MASK 0x18 -#define REG_ENTROPY_MODE (REG_ENTROPY_START + 0x05) -#define ENTROPY_MODE_KEY_MASK 0x07 -#define ENTROPY_MODE_ENCRYPT 0x40 -#define REG_ENTROPY_RAND_REG (REG_ENTROPY_START + 0x06) -#define REG_ENTROPY_RESET (REG_ENTROPY_START + 0x07) -#define ENTROPY_RESET_DES_IO 0x01 -#define ENTROPY_RESET_STC_MODE 0x02 -#define ENTROPY_RESET_KEY_CACHE 0x04 -#define ENTROPY_RESET_IV 0x08 -#define REG_ENTROPY_IV (REG_ENTROPY_START + 0x08) -#define REG_ENTROPY_KEY0 (REG_ENTROPY_START + 0x10) -#define REG_ENTROPY_KEYN(x) (REG_ENTROPY_KEY0 + 4*(x)) - -/* phys of interest w/ their special mii registers */ -#define PHY_LUCENT_B0 0x00437421 -#define LUCENT_MII_REG 0x1F - -#define PHY_NS_DP83065 0x20005c78 -#define DP83065_MII_MEM 0x16 -#define DP83065_MII_REGD 0x1D -#define DP83065_MII_REGE 0x1E - -#define PHY_BROADCOM_5411 0x00206071 -#define PHY_BROADCOM_B0 0x00206050 -#define BROADCOM_MII_REG4 0x14 -#define BROADCOM_MII_REG5 0x15 -#define BROADCOM_MII_REG7 0x17 -#define BROADCOM_MII_REG8 0x18 - -#define CAS_MII_ANNPTR 0x07 -#define CAS_MII_ANNPRR 0x08 -#define CAS_MII_1000_CTRL 0x09 -#define CAS_MII_1000_STATUS 0x0A -#define CAS_MII_1000_EXTEND 0x0F - -#define CAS_BMSR_1000_EXTEND 0x0100 /* supports 1000Base-T extended status */ -/* - * if autoneg is disabled, here's the table: - * BMCR_SPEED100 = 100Mbps - * BMCR_SPEED1000 = 1000Mbps - * ~(BMCR_SPEED100 | BMCR_SPEED1000) = 10Mbps - */ -#define CAS_BMCR_SPEED1000 0x0040 /* Select 1000Mbps */ - -#define CAS_ADVERTISE_1000HALF 0x0100 -#define CAS_ADVERTISE_1000FULL 0x0200 -#define CAS_ADVERTISE_PAUSE 0x0400 -#define CAS_ADVERTISE_ASYM_PAUSE 0x0800 - -/* regular lpa register */ -#define CAS_LPA_PAUSE CAS_ADVERTISE_PAUSE -#define CAS_LPA_ASYM_PAUSE CAS_ADVERTISE_ASYM_PAUSE - -/* 1000_STATUS register */ -#define CAS_LPA_1000HALF 0x0400 -#define CAS_LPA_1000FULL 0x0800 - -#define CAS_EXTEND_1000XFULL 0x8000 -#define CAS_EXTEND_1000XHALF 0x4000 -#define CAS_EXTEND_1000TFULL 0x2000 -#define CAS_EXTEND_1000THALF 0x1000 - -/* cassini header parser firmware */ -typedef struct cas_hp_inst { - const char *note; - - u16 mask, val; - - u8 op; - u8 soff, snext; /* if match succeeds, new offset and match */ - u8 foff, fnext; /* if match fails, new offset and match */ - /* output info */ - u8 outop; /* output opcode */ - - u16 outarg; /* output argument */ - u8 outenab; /* output enable: 0 = not, 1 = if match - 2 = if !match, 3 = always */ - u8 outshift; /* barrel shift right, 4 bits */ - u16 outmask; -} cas_hp_inst_t; - -/* comparison */ -#define OP_EQ 0 /* packet == value */ -#define OP_LT 1 /* packet < value */ -#define OP_GT 2 /* packet > value */ -#define OP_NP 3 /* new packet */ - -/* output opcodes */ -#define CL_REG 0 -#define LD_FID 1 -#define LD_SEQ 2 -#define LD_CTL 3 -#define LD_SAP 4 -#define LD_R1 5 -#define LD_L3 6 -#define LD_SUM 7 -#define LD_HDR 8 -#define IM_FID 9 -#define IM_SEQ 10 -#define IM_SAP 11 -#define IM_R1 12 -#define IM_CTL 13 -#define LD_LEN 14 -#define ST_FLG 15 - -/* match setp #s for IP4TCP4 */ -#define S1_PCKT 0 -#define S1_VLAN 1 -#define S1_CFI 2 -#define S1_8023 3 -#define S1_LLC 4 -#define S1_LLCc 5 -#define S1_IPV4 6 -#define S1_IPV4c 7 -#define S1_IPV4F 8 -#define S1_TCP44 9 -#define S1_IPV6 10 -#define S1_IPV6L 11 -#define S1_IPV6c 12 -#define S1_TCP64 13 -#define S1_TCPSQ 14 -#define S1_TCPFG 15 -#define S1_TCPHL 16 -#define S1_TCPHc 17 -#define S1_CLNP 18 -#define S1_CLNP2 19 -#define S1_DROP 20 -#define S2_HTTP 21 -#define S1_ESP4 22 -#define S1_AH4 23 -#define S1_ESP6 24 -#define S1_AH6 25 - -#define CAS_PROG_IP46TCP4_PREAMBLE \ -{ "packet arrival?", 0xffff, 0x0000, OP_NP, 6, S1_VLAN, 0, S1_PCKT, \ - CL_REG, 0x3ff, 1, 0x0, 0x0000}, \ -{ "VLAN?", 0xffff, 0x8100, OP_EQ, 1, S1_CFI, 0, S1_8023, \ - IM_CTL, 0x00a, 3, 0x0, 0xffff}, \ -{ "CFI?", 0x1000, 0x1000, OP_EQ, 0, S1_DROP, 1, S1_8023, \ - CL_REG, 0x000, 0, 0x0, 0x0000}, \ -{ "8023?", 0xffff, 0x0600, OP_LT, 1, S1_LLC, 0, S1_IPV4, \ - CL_REG, 0x000, 0, 0x0, 0x0000}, \ -{ "LLC?", 0xffff, 0xaaaa, OP_EQ, 1, S1_LLCc, 0, S1_CLNP, \ - CL_REG, 0x000, 0, 0x0, 0x0000}, \ -{ "LLCc?", 0xff00, 0x0300, OP_EQ, 2, S1_IPV4, 0, S1_CLNP, \ - CL_REG, 0x000, 0, 0x0, 0x0000}, \ -{ "IPV4?", 0xffff, 0x0800, OP_EQ, 1, S1_IPV4c, 0, S1_IPV6, \ - LD_SAP, 0x100, 3, 0x0, 0xffff}, \ -{ "IPV4 cont?", 0xff00, 0x4500, OP_EQ, 3, S1_IPV4F, 0, S1_CLNP, \ - LD_SUM, 0x00a, 1, 0x0, 0x0000}, \ -{ "IPV4 frag?", 0x3fff, 0x0000, OP_EQ, 1, S1_TCP44, 0, S1_CLNP, \ - LD_LEN, 0x03e, 1, 0x0, 0xffff}, \ -{ "TCP44?", 0x00ff, 0x0006, OP_EQ, 7, S1_TCPSQ, 0, S1_CLNP, \ - LD_FID, 0x182, 1, 0x0, 0xffff}, /* FID IP4&TCP src+dst */ \ -{ "IPV6?", 0xffff, 0x86dd, OP_EQ, 1, S1_IPV6L, 0, S1_CLNP, \ - LD_SUM, 0x015, 1, 0x0, 0x0000}, \ -{ "IPV6 len", 0xf000, 0x6000, OP_EQ, 0, S1_IPV6c, 0, S1_CLNP, \ - IM_R1, 0x128, 1, 0x0, 0xffff}, \ -{ "IPV6 cont?", 0x0000, 0x0000, OP_EQ, 3, S1_TCP64, 0, S1_CLNP, \ - LD_FID, 0x484, 1, 0x0, 0xffff}, /* FID IP6&TCP src+dst */ \ -{ "TCP64?", 0xff00, 0x0600, OP_EQ, 18, S1_TCPSQ, 0, S1_CLNP, \ - LD_LEN, 0x03f, 1, 0x0, 0xffff} - -#ifdef USE_HP_IP46TCP4 -static cas_hp_inst_t cas_prog_ip46tcp4tab[] = { - CAS_PROG_IP46TCP4_PREAMBLE, - { "TCP seq", /* DADDR should point to dest port */ - 0x0000, 0x0000, OP_EQ, 0, S1_TCPFG, 4, S1_TCPFG, LD_SEQ, - 0x081, 3, 0x0, 0xffff}, /* Load TCP seq # */ - { "TCP control flags", 0x0000, 0x0000, OP_EQ, 0, S1_TCPHL, 0, - S1_TCPHL, ST_FLG, 0x045, 3, 0x0, 0x002f}, /* Load TCP flags */ - { "TCP length", 0x0000, 0x0000, OP_EQ, 0, S1_TCPHc, 0, - S1_TCPHc, LD_R1, 0x205, 3, 0xB, 0xf000}, - { "TCP length cont", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, - S1_PCKT, LD_HDR, 0x0ff, 3, 0x0, 0xffff}, - { "Cleanup", 0x0000, 0x0000, OP_EQ, 0, S1_CLNP2, 0, S1_CLNP2, - IM_CTL, 0x001, 3, 0x0, 0x0001}, - { "Cleanup 2", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - IM_CTL, 0x000, 0, 0x0, 0x0000}, - { "Drop packet", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - IM_CTL, 0x080, 3, 0x0, 0xffff}, - { NULL }, -}; -#ifdef HP_IP46TCP4_DEFAULT -#define CAS_HP_FIRMWARE cas_prog_ip46tcp4tab -#endif -#endif - -/* - * Alternate table load which excludes HTTP server traffic from reassembly. - * It is substantially similar to the basic table, with one extra state - * and a few extra compares. */ -#ifdef USE_HP_IP46TCP4NOHTTP -static cas_hp_inst_t cas_prog_ip46tcp4nohttptab[] = { - CAS_PROG_IP46TCP4_PREAMBLE, - { "TCP seq", /* DADDR should point to dest port */ - 0xFFFF, 0x0080, OP_EQ, 0, S2_HTTP, 0, S1_TCPFG, LD_SEQ, - 0x081, 3, 0x0, 0xffff} , /* Load TCP seq # */ - { "TCP control flags", 0xFFFF, 0x8080, OP_EQ, 0, S2_HTTP, 0, - S1_TCPHL, ST_FLG, 0x145, 2, 0x0, 0x002f, }, /* Load TCP flags */ - { "TCP length", 0x0000, 0x0000, OP_EQ, 0, S1_TCPHc, 0, S1_TCPHc, - LD_R1, 0x205, 3, 0xB, 0xf000}, - { "TCP length cont", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - LD_HDR, 0x0ff, 3, 0x0, 0xffff}, - { "Cleanup", 0x0000, 0x0000, OP_EQ, 0, S1_CLNP2, 0, S1_CLNP2, - IM_CTL, 0x001, 3, 0x0, 0x0001}, - { "Cleanup 2", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - CL_REG, 0x002, 3, 0x0, 0x0000}, - { "Drop packet", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - IM_CTL, 0x080, 3, 0x0, 0xffff}, - { "No HTTP", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - IM_CTL, 0x044, 3, 0x0, 0xffff}, - { NULL }, -}; -#ifdef HP_IP46TCP4NOHTTP_DEFAULT -#define CAS_HP_FIRMWARE cas_prog_ip46tcp4nohttptab -#endif -#endif - -/* match step #s for IP4FRAG */ -#define S3_IPV6c 11 -#define S3_TCP64 12 -#define S3_TCPSQ 13 -#define S3_TCPFG 14 -#define S3_TCPHL 15 -#define S3_TCPHc 16 -#define S3_FRAG 17 -#define S3_FOFF 18 -#define S3_CLNP 19 - -#ifdef USE_HP_IP4FRAG -static cas_hp_inst_t cas_prog_ip4fragtab[] = { - { "packet arrival?", 0xffff, 0x0000, OP_NP, 6, S1_VLAN, 0, S1_PCKT, - CL_REG, 0x3ff, 1, 0x0, 0x0000}, - { "VLAN?", 0xffff, 0x8100, OP_EQ, 1, S1_CFI, 0, S1_8023, - IM_CTL, 0x00a, 3, 0x0, 0xffff}, - { "CFI?", 0x1000, 0x1000, OP_EQ, 0, S3_CLNP, 1, S1_8023, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "8023?", 0xffff, 0x0600, OP_LT, 1, S1_LLC, 0, S1_IPV4, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "LLC?", 0xffff, 0xaaaa, OP_EQ, 1, S1_LLCc, 0, S3_CLNP, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "LLCc?",0xff00, 0x0300, OP_EQ, 2, S1_IPV4, 0, S3_CLNP, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "IPV4?", 0xffff, 0x0800, OP_EQ, 1, S1_IPV4c, 0, S1_IPV6, - LD_SAP, 0x100, 3, 0x0, 0xffff}, - { "IPV4 cont?", 0xff00, 0x4500, OP_EQ, 3, S1_IPV4F, 0, S3_CLNP, - LD_SUM, 0x00a, 1, 0x0, 0x0000}, - { "IPV4 frag?", 0x3fff, 0x0000, OP_EQ, 1, S1_TCP44, 0, S3_FRAG, - LD_LEN, 0x03e, 3, 0x0, 0xffff}, - { "TCP44?", 0x00ff, 0x0006, OP_EQ, 7, S3_TCPSQ, 0, S3_CLNP, - LD_FID, 0x182, 3, 0x0, 0xffff}, /* FID IP4&TCP src+dst */ - { "IPV6?", 0xffff, 0x86dd, OP_EQ, 1, S3_IPV6c, 0, S3_CLNP, - LD_SUM, 0x015, 1, 0x0, 0x0000}, - { "IPV6 cont?", 0xf000, 0x6000, OP_EQ, 3, S3_TCP64, 0, S3_CLNP, - LD_FID, 0x484, 1, 0x0, 0xffff}, /* FID IP6&TCP src+dst */ - { "TCP64?", 0xff00, 0x0600, OP_EQ, 18, S3_TCPSQ, 0, S3_CLNP, - LD_LEN, 0x03f, 1, 0x0, 0xffff}, - { "TCP seq", /* DADDR should point to dest port */ - 0x0000, 0x0000, OP_EQ, 0, S3_TCPFG, 4, S3_TCPFG, LD_SEQ, - 0x081, 3, 0x0, 0xffff}, /* Load TCP seq # */ - { "TCP control flags", 0x0000, 0x0000, OP_EQ, 0, S3_TCPHL, 0, - S3_TCPHL, ST_FLG, 0x045, 3, 0x0, 0x002f}, /* Load TCP flags */ - { "TCP length", 0x0000, 0x0000, OP_EQ, 0, S3_TCPHc, 0, S3_TCPHc, - LD_R1, 0x205, 3, 0xB, 0xf000}, - { "TCP length cont", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - LD_HDR, 0x0ff, 3, 0x0, 0xffff}, - { "IP4 Fragment", 0x0000, 0x0000, OP_EQ, 0, S3_FOFF, 0, S3_FOFF, - LD_FID, 0x103, 3, 0x0, 0xffff}, /* FID IP4 src+dst */ - { "IP4 frag offset", 0x0000, 0x0000, OP_EQ, 0, S3_FOFF, 0, S3_FOFF, - LD_SEQ, 0x040, 1, 0xD, 0xfff8}, - { "Cleanup", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - IM_CTL, 0x001, 3, 0x0, 0x0001}, - { NULL }, -}; -#ifdef HP_IP4FRAG_DEFAULT -#define CAS_HP_FIRMWARE cas_prog_ip4fragtab -#endif -#endif - -/* - * Alternate table which does batching without reassembly - */ -#ifdef USE_HP_IP46TCP4BATCH -static cas_hp_inst_t cas_prog_ip46tcp4batchtab[] = { - CAS_PROG_IP46TCP4_PREAMBLE, - { "TCP seq", /* DADDR should point to dest port */ - 0x0000, 0x0000, OP_EQ, 0, S1_TCPFG, 0, S1_TCPFG, LD_SEQ, - 0x081, 3, 0x0, 0xffff}, /* Load TCP seq # */ - { "TCP control flags", 0x0000, 0x0000, OP_EQ, 0, S1_TCPHL, 0, - S1_TCPHL, ST_FLG, 0x000, 3, 0x0, 0x0000}, /* Load TCP flags */ - { "TCP length", 0x0000, 0x0000, OP_EQ, 0, S1_TCPHc, 0, - S1_TCPHc, LD_R1, 0x205, 3, 0xB, 0xf000}, - { "TCP length cont", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, - S1_PCKT, IM_CTL, 0x040, 3, 0x0, 0xffff}, /* set batch bit */ - { "Cleanup", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - IM_CTL, 0x001, 3, 0x0, 0x0001}, - { "Drop packet", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, - S1_PCKT, IM_CTL, 0x080, 3, 0x0, 0xffff}, - { NULL }, -}; -#ifdef HP_IP46TCP4BATCH_DEFAULT -#define CAS_HP_FIRMWARE cas_prog_ip46tcp4batchtab -#endif -#endif - -/* Workaround for Cassini rev2 descriptor corruption problem. - * Does batching without reassembly, and sets the SAP to a known - * data pattern for all packets. - */ -#ifdef USE_HP_WORKAROUND -static cas_hp_inst_t cas_prog_workaroundtab[] = { - { "packet arrival?", 0xffff, 0x0000, OP_NP, 6, S1_VLAN, 0, - S1_PCKT, CL_REG, 0x3ff, 1, 0x0, 0x0000} , - { "VLAN?", 0xffff, 0x8100, OP_EQ, 1, S1_CFI, 0, S1_8023, - IM_CTL, 0x04a, 3, 0x0, 0xffff}, - { "CFI?", 0x1000, 0x1000, OP_EQ, 0, S1_CLNP, 1, S1_8023, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "8023?", 0xffff, 0x0600, OP_LT, 1, S1_LLC, 0, S1_IPV4, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "LLC?", 0xffff, 0xaaaa, OP_EQ, 1, S1_LLCc, 0, S1_CLNP, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "LLCc?", 0xff00, 0x0300, OP_EQ, 2, S1_IPV4, 0, S1_CLNP, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "IPV4?", 0xffff, 0x0800, OP_EQ, 1, S1_IPV4c, 0, S1_IPV6, - IM_SAP, 0x6AE, 3, 0x0, 0xffff}, - { "IPV4 cont?", 0xff00, 0x4500, OP_EQ, 3, S1_IPV4F, 0, S1_CLNP, - LD_SUM, 0x00a, 1, 0x0, 0x0000}, - { "IPV4 frag?", 0x3fff, 0x0000, OP_EQ, 1, S1_TCP44, 0, S1_CLNP, - LD_LEN, 0x03e, 1, 0x0, 0xffff}, - { "TCP44?", 0x00ff, 0x0006, OP_EQ, 7, S1_TCPSQ, 0, S1_CLNP, - LD_FID, 0x182, 3, 0x0, 0xffff}, /* FID IP4&TCP src+dst */ - { "IPV6?", 0xffff, 0x86dd, OP_EQ, 1, S1_IPV6L, 0, S1_CLNP, - LD_SUM, 0x015, 1, 0x0, 0x0000}, - { "IPV6 len", 0xf000, 0x6000, OP_EQ, 0, S1_IPV6c, 0, S1_CLNP, - IM_R1, 0x128, 1, 0x0, 0xffff}, - { "IPV6 cont?", 0x0000, 0x0000, OP_EQ, 3, S1_TCP64, 0, S1_CLNP, - LD_FID, 0x484, 1, 0x0, 0xffff}, /* FID IP6&TCP src+dst */ - { "TCP64?", 0xff00, 0x0600, OP_EQ, 18, S1_TCPSQ, 0, S1_CLNP, - LD_LEN, 0x03f, 1, 0x0, 0xffff}, - { "TCP seq", /* DADDR should point to dest port */ - 0x0000, 0x0000, OP_EQ, 0, S1_TCPFG, 4, S1_TCPFG, LD_SEQ, - 0x081, 3, 0x0, 0xffff}, /* Load TCP seq # */ - { "TCP control flags", 0x0000, 0x0000, OP_EQ, 0, S1_TCPHL, 0, - S1_TCPHL, ST_FLG, 0x045, 3, 0x0, 0x002f}, /* Load TCP flags */ - { "TCP length", 0x0000, 0x0000, OP_EQ, 0, S1_TCPHc, 0, S1_TCPHc, - LD_R1, 0x205, 3, 0xB, 0xf000}, - { "TCP length cont", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, - S1_PCKT, LD_HDR, 0x0ff, 3, 0x0, 0xffff}, - { "Cleanup", 0x0000, 0x0000, OP_EQ, 0, S1_CLNP2, 0, S1_CLNP2, - IM_SAP, 0x6AE, 3, 0x0, 0xffff} , - { "Cleanup 2", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - IM_CTL, 0x001, 3, 0x0, 0x0001}, - { NULL }, -}; -#ifdef HP_WORKAROUND_DEFAULT -#define CAS_HP_FIRMWARE cas_prog_workaroundtab -#endif -#endif - -#ifdef USE_HP_ENCRYPT -static cas_hp_inst_t cas_prog_encryptiontab[] = { - { "packet arrival?", 0xffff, 0x0000, OP_NP, 6, S1_VLAN, 0, - S1_PCKT, CL_REG, 0x3ff, 1, 0x0, 0x0000}, - { "VLAN?", 0xffff, 0x8100, OP_EQ, 1, S1_CFI, 0, S1_8023, - IM_CTL, 0x00a, 3, 0x0, 0xffff}, -#if 0 -//"CFI?", /* 02 FIND CFI and If FIND go to S1_DROP */ -//0x1000, 0x1000, OP_EQ, 0, S1_DROP, 1, S1_8023, CL_REG, 0x000, 0, 0x0, 0x00 - 00, -#endif - { "CFI?", /* FIND CFI and If FIND go to CleanUP1 (ignore and send to host) */ - 0x1000, 0x1000, OP_EQ, 0, S1_CLNP, 1, S1_8023, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "8023?", 0xffff, 0x0600, OP_LT, 1, S1_LLC, 0, S1_IPV4, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "LLC?", 0xffff, 0xaaaa, OP_EQ, 1, S1_LLCc, 0, S1_CLNP, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "LLCc?", 0xff00, 0x0300, OP_EQ, 2, S1_IPV4, 0, S1_CLNP, - CL_REG, 0x000, 0, 0x0, 0x0000}, - { "IPV4?", 0xffff, 0x0800, OP_EQ, 1, S1_IPV4c, 0, S1_IPV6, - LD_SAP, 0x100, 3, 0x0, 0xffff}, - { "IPV4 cont?", 0xff00, 0x4500, OP_EQ, 3, S1_IPV4F, 0, S1_CLNP, - LD_SUM, 0x00a, 1, 0x0, 0x0000}, - { "IPV4 frag?", 0x3fff, 0x0000, OP_EQ, 1, S1_TCP44, 0, S1_CLNP, - LD_LEN, 0x03e, 1, 0x0, 0xffff}, - { "TCP44?", 0x00ff, 0x0006, OP_EQ, 7, S1_TCPSQ, 0, S1_ESP4, - LD_FID, 0x182, 1, 0x0, 0xffff}, /* FID IP4&TCP src+dst */ - { "IPV6?", 0xffff, 0x86dd, OP_EQ, 1, S1_IPV6L, 0, S1_CLNP, - LD_SUM, 0x015, 1, 0x0, 0x0000}, - { "IPV6 len", 0xf000, 0x6000, OP_EQ, 0, S1_IPV6c, 0, S1_CLNP, - IM_R1, 0x128, 1, 0x0, 0xffff}, - { "IPV6 cont?", 0x0000, 0x0000, OP_EQ, 3, S1_TCP64, 0, S1_CLNP, - LD_FID, 0x484, 1, 0x0, 0xffff}, /* FID IP6&TCP src+dst */ - { "TCP64?", -#if 0 -//@@@0xff00, 0x0600, OP_EQ, 18, S1_TCPSQ, 0, S1_ESP6, LD_LEN, 0x03f, 1, 0x0, 0xffff, -#endif - 0xff00, 0x0600, OP_EQ, 12, S1_TCPSQ, 0, S1_ESP6, LD_LEN, - 0x03f, 1, 0x0, 0xffff}, - { "TCP seq", /* 14:DADDR should point to dest port */ - 0xFFFF, 0x0080, OP_EQ, 0, S2_HTTP, 0, S1_TCPFG, LD_SEQ, - 0x081, 3, 0x0, 0xffff}, /* Load TCP seq # */ - { "TCP control flags", 0xFFFF, 0x8080, OP_EQ, 0, S2_HTTP, 0, - S1_TCPHL, ST_FLG, 0x145, 2, 0x0, 0x002f}, /* Load TCP flags */ - { "TCP length", 0x0000, 0x0000, OP_EQ, 0, S1_TCPHc, 0, S1_TCPHc, - LD_R1, 0x205, 3, 0xB, 0xf000} , - { "TCP length cont", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, - S1_PCKT, LD_HDR, 0x0ff, 3, 0x0, 0xffff}, - { "Cleanup", 0x0000, 0x0000, OP_EQ, 0, S1_CLNP2, 0, S1_CLNP2, - IM_CTL, 0x001, 3, 0x0, 0x0001}, - { "Cleanup 2", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - CL_REG, 0x002, 3, 0x0, 0x0000}, - { "Drop packet", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - IM_CTL, 0x080, 3, 0x0, 0xffff}, - { "No HTTP", 0x0000, 0x0000, OP_EQ, 0, S1_PCKT, 0, S1_PCKT, - IM_CTL, 0x044, 3, 0x0, 0xffff}, - { "IPV4 ESP encrypted?", /* S1_ESP4 */ - 0x00ff, 0x0032, OP_EQ, 0, S1_CLNP2, 0, S1_AH4, IM_CTL, - 0x021, 1, 0x0, 0xffff}, - { "IPV4 AH encrypted?", /* S1_AH4 */ - 0x00ff, 0x0033, OP_EQ, 0, S1_CLNP2, 0, S1_CLNP, IM_CTL, - 0x021, 1, 0x0, 0xffff}, - { "IPV6 ESP encrypted?", /* S1_ESP6 */ -#if 0 -//@@@0x00ff, 0x0032, OP_EQ, 0, S1_CLNP2, 0, S1_AH6, IM_CTL, 0x021, 1, 0x0, 0xffff, -#endif - 0xff00, 0x3200, OP_EQ, 0, S1_CLNP2, 0, S1_AH6, IM_CTL, - 0x021, 1, 0x0, 0xffff}, - { "IPV6 AH encrypted?", /* S1_AH6 */ -#if 0 -//@@@0x00ff, 0x0033, OP_EQ, 0, S1_CLNP2, 0, S1_CLNP, IM_CTL, 0x021, 1, 0x0, 0xffff, -#endif - 0xff00, 0x3300, OP_EQ, 0, S1_CLNP2, 0, S1_CLNP, IM_CTL, - 0x021, 1, 0x0, 0xffff}, - { NULL }, -}; -#ifdef HP_ENCRYPT_DEFAULT -#define CAS_HP_FIRMWARE cas_prog_encryptiontab -#endif -#endif - -static cas_hp_inst_t cas_prog_null[] = { {NULL} }; -#ifdef HP_NULL_DEFAULT -#define CAS_HP_FIRMWARE cas_prog_null -#endif - -/* phy types */ -#define CAS_PHY_UNKNOWN 0x00 -#define CAS_PHY_SERDES 0x01 -#define CAS_PHY_MII_MDIO0 0x02 -#define CAS_PHY_MII_MDIO1 0x04 -#define CAS_PHY_MII(x) ((x) & (CAS_PHY_MII_MDIO0 | CAS_PHY_MII_MDIO1)) - -/* _RING_INDEX is the index for the ring sizes to be used. _RING_SIZE - * is the actual size. the default index for the various rings is - * 8. NOTE: there a bunch of alignment constraints for the rings. to - * deal with that, i just allocate rings to create the desired - * alignment. here are the constraints: - * RX DESC and COMP rings must be 8KB aligned - * TX DESC must be 2KB aligned. - * if you change the numbers, be cognizant of how the alignment will change - * in INIT_BLOCK as well. - */ - -#define DESC_RING_I_TO_S(x) (32*(1 << (x))) -#define COMP_RING_I_TO_S(x) (128*(1 << (x))) -#define TX_DESC_RING_INDEX 4 /* 512 = 8k */ -#define RX_DESC_RING_INDEX 4 /* 512 = 8k */ -#define RX_COMP_RING_INDEX 4 /* 2048 = 64k: should be 4x rx ring size */ - -#if (TX_DESC_RING_INDEX > 8) || (TX_DESC_RING_INDEX < 0) -#error TX_DESC_RING_INDEX must be between 0 and 8 -#endif - -#if (RX_DESC_RING_INDEX > 8) || (RX_DESC_RING_INDEX < 0) -#error RX_DESC_RING_INDEX must be between 0 and 8 -#endif - -#if (RX_COMP_RING_INDEX > 8) || (RX_COMP_RING_INDEX < 0) -#error RX_COMP_RING_INDEX must be between 0 and 8 -#endif - -#define N_TX_RINGS MAX_TX_RINGS /* for QoS */ -#define N_TX_RINGS_MASK MAX_TX_RINGS_MASK -#define N_RX_DESC_RINGS MAX_RX_DESC_RINGS /* 1 for ipsec */ -#define N_RX_COMP_RINGS 0x1 /* for mult. PCI interrupts */ - -/* number of flows that can go through re-assembly */ -#define N_RX_FLOWS 64 - -#define TX_DESC_RING_SIZE DESC_RING_I_TO_S(TX_DESC_RING_INDEX) -#define RX_DESC_RING_SIZE DESC_RING_I_TO_S(RX_DESC_RING_INDEX) -#define RX_COMP_RING_SIZE COMP_RING_I_TO_S(RX_COMP_RING_INDEX) -#define TX_DESC_RINGN_INDEX(x) TX_DESC_RING_INDEX -#define RX_DESC_RINGN_INDEX(x) RX_DESC_RING_INDEX -#define RX_COMP_RINGN_INDEX(x) RX_COMP_RING_INDEX -#define TX_DESC_RINGN_SIZE(x) TX_DESC_RING_SIZE -#define RX_DESC_RINGN_SIZE(x) RX_DESC_RING_SIZE -#define RX_COMP_RINGN_SIZE(x) RX_COMP_RING_SIZE - -/* convert values */ -#define CAS_BASE(x, y) (((y) << (x ## _SHIFT)) & (x ## _MASK)) -#define CAS_VAL(x, y) (((y) & (x ## _MASK)) >> (x ## _SHIFT)) -#define CAS_TX_RINGN_BASE(y) ((TX_DESC_RINGN_INDEX(y) << \ - TX_CFG_DESC_RINGN_SHIFT(y)) & \ - TX_CFG_DESC_RINGN_MASK(y)) - -/* min is 2k, but we can't do jumbo frames unless it's at least 8k */ -#define CAS_MIN_PAGE_SHIFT 11 /* 2048 */ -#define CAS_JUMBO_PAGE_SHIFT 13 /* 8192 */ -#define CAS_MAX_PAGE_SHIFT 14 /* 16384 */ - -#define TX_DESC_BUFLEN_MASK 0x0000000000003FFFULL /* buffer length in - bytes. 0 - 9256 */ -#define TX_DESC_BUFLEN_SHIFT 0 -#define TX_DESC_CSUM_START_MASK 0x00000000001F8000ULL /* checksum start. # - of bytes to be - skipped before - csum calc begins. - value must be - even */ -#define TX_DESC_CSUM_START_SHIFT 15 -#define TX_DESC_CSUM_STUFF_MASK 0x000000001FE00000ULL /* checksum stuff. - byte offset w/in - the pkt for the - 1st csum byte. - must be > 8 */ -#define TX_DESC_CSUM_STUFF_SHIFT 21 -#define TX_DESC_CSUM_EN 0x0000000020000000ULL /* enable checksum */ -#define TX_DESC_EOF 0x0000000040000000ULL /* end of frame */ -#define TX_DESC_SOF 0x0000000080000000ULL /* start of frame */ -#define TX_DESC_INTME 0x0000000100000000ULL /* interrupt me */ -#define TX_DESC_NO_CRC 0x0000000200000000ULL /* debugging only. - CRC will not be - inserted into - outgoing frame. */ -struct cas_tx_desc { - __le64 control; - __le64 buffer; -}; - -/* descriptor ring for free buffers contains page-sized buffers. the index - * value is not used by the hw in any way. it's just stored and returned in - * the completion ring. - */ -struct cas_rx_desc { - __le64 index; - __le64 buffer; -}; - -/* received packets are put on the completion ring. */ -/* word 1 */ -#define RX_COMP1_DATA_SIZE_MASK 0x0000000007FFE000ULL -#define RX_COMP1_DATA_SIZE_SHIFT 13 -#define RX_COMP1_DATA_OFF_MASK 0x000001FFF8000000ULL -#define RX_COMP1_DATA_OFF_SHIFT 27 -#define RX_COMP1_DATA_INDEX_MASK 0x007FFE0000000000ULL -#define RX_COMP1_DATA_INDEX_SHIFT 41 -#define RX_COMP1_SKIP_MASK 0x0180000000000000ULL -#define RX_COMP1_SKIP_SHIFT 55 -#define RX_COMP1_RELEASE_NEXT 0x0200000000000000ULL -#define RX_COMP1_SPLIT_PKT 0x0400000000000000ULL -#define RX_COMP1_RELEASE_FLOW 0x0800000000000000ULL -#define RX_COMP1_RELEASE_DATA 0x1000000000000000ULL -#define RX_COMP1_RELEASE_HDR 0x2000000000000000ULL -#define RX_COMP1_TYPE_MASK 0xC000000000000000ULL -#define RX_COMP1_TYPE_SHIFT 62 - -/* word 2 */ -#define RX_COMP2_NEXT_INDEX_MASK 0x00000007FFE00000ULL -#define RX_COMP2_NEXT_INDEX_SHIFT 21 -#define RX_COMP2_HDR_SIZE_MASK 0x00000FF800000000ULL -#define RX_COMP2_HDR_SIZE_SHIFT 35 -#define RX_COMP2_HDR_OFF_MASK 0x0003F00000000000ULL -#define RX_COMP2_HDR_OFF_SHIFT 44 -#define RX_COMP2_HDR_INDEX_MASK 0xFFFC000000000000ULL -#define RX_COMP2_HDR_INDEX_SHIFT 50 - -/* word 3 */ -#define RX_COMP3_SMALL_PKT 0x0000000000000001ULL -#define RX_COMP3_JUMBO_PKT 0x0000000000000002ULL -#define RX_COMP3_JUMBO_HDR_SPLIT_EN 0x0000000000000004ULL -#define RX_COMP3_CSUM_START_MASK 0x000000000007F000ULL -#define RX_COMP3_CSUM_START_SHIFT 12 -#define RX_COMP3_FLOWID_MASK 0x0000000001F80000ULL -#define RX_COMP3_FLOWID_SHIFT 19 -#define RX_COMP3_OPCODE_MASK 0x000000000E000000ULL -#define RX_COMP3_OPCODE_SHIFT 25 -#define RX_COMP3_FORCE_FLAG 0x0000000010000000ULL -#define RX_COMP3_NO_ASSIST 0x0000000020000000ULL -#define RX_COMP3_LOAD_BAL_MASK 0x000001F800000000ULL -#define RX_COMP3_LOAD_BAL_SHIFT 35 -#define RX_PLUS_COMP3_ENC_PKT 0x0000020000000000ULL /* cas+ */ -#define RX_COMP3_L3_HEAD_OFF_MASK 0x0000FE0000000000ULL /* cas */ -#define RX_COMP3_L3_HEAD_OFF_SHIFT 41 -#define RX_PLUS_COMP_L3_HEAD_OFF_MASK 0x0000FC0000000000ULL /* cas+ */ -#define RX_PLUS_COMP_L3_HEAD_OFF_SHIFT 42 -#define RX_COMP3_SAP_MASK 0xFFFF000000000000ULL -#define RX_COMP3_SAP_SHIFT 48 - -/* word 4 */ -#define RX_COMP4_TCP_CSUM_MASK 0x000000000000FFFFULL -#define RX_COMP4_TCP_CSUM_SHIFT 0 -#define RX_COMP4_PKT_LEN_MASK 0x000000003FFF0000ULL -#define RX_COMP4_PKT_LEN_SHIFT 16 -#define RX_COMP4_PERFECT_MATCH_MASK 0x00000003C0000000ULL -#define RX_COMP4_PERFECT_MATCH_SHIFT 30 -#define RX_COMP4_ZERO 0x0000080000000000ULL -#define RX_COMP4_HASH_VAL_MASK 0x0FFFF00000000000ULL -#define RX_COMP4_HASH_VAL_SHIFT 44 -#define RX_COMP4_HASH_PASS 0x1000000000000000ULL -#define RX_COMP4_BAD 0x4000000000000000ULL -#define RX_COMP4_LEN_MISMATCH 0x8000000000000000ULL - -/* we encode the following: ring/index/release. only 14 bits - * are usable. - * NOTE: the encoding is dependent upon RX_DESC_RING_SIZE and - * MAX_RX_DESC_RINGS. */ -#define RX_INDEX_NUM_MASK 0x0000000000000FFFULL -#define RX_INDEX_NUM_SHIFT 0 -#define RX_INDEX_RING_MASK 0x0000000000001000ULL -#define RX_INDEX_RING_SHIFT 12 -#define RX_INDEX_RELEASE 0x0000000000002000ULL - -struct cas_rx_comp { - __le64 word1; - __le64 word2; - __le64 word3; - __le64 word4; -}; - -enum link_state { - link_down = 0, /* No link, will retry */ - link_aneg, /* Autoneg in progress */ - link_force_try, /* Try Forced link speed */ - link_force_ret, /* Forced mode worked, retrying autoneg */ - link_force_ok, /* Stay in forced mode */ - link_up /* Link is up */ -}; - -typedef struct cas_page { - struct list_head list; - struct page *buffer; - dma_addr_t dma_addr; - int used; -} cas_page_t; - - -/* some alignment constraints: - * TX DESC, RX DESC, and RX COMP must each be 8K aligned. - * TX COMPWB must be 8-byte aligned. - * to accomplish this, here's what we do: - * - * INIT_BLOCK_RX_COMP = 64k (already aligned) - * INIT_BLOCK_RX_DESC = 8k - * INIT_BLOCK_TX = 8k - * INIT_BLOCK_RX1_DESC = 8k - * TX COMPWB - */ -#define INIT_BLOCK_TX (TX_DESC_RING_SIZE) -#define INIT_BLOCK_RX_DESC (RX_DESC_RING_SIZE) -#define INIT_BLOCK_RX_COMP (RX_COMP_RING_SIZE) - -struct cas_init_block { - struct cas_rx_comp rxcs[N_RX_COMP_RINGS][INIT_BLOCK_RX_COMP]; - struct cas_rx_desc rxds[N_RX_DESC_RINGS][INIT_BLOCK_RX_DESC]; - struct cas_tx_desc txds[N_TX_RINGS][INIT_BLOCK_TX]; - __le64 tx_compwb; -}; - -/* tiny buffers to deal with target abort issue. we allocate a bit - * over so that we don't have target abort issues with these buffers - * as well. - */ -#define TX_TINY_BUF_LEN 0x100 -#define TX_TINY_BUF_BLOCK ((INIT_BLOCK_TX + 1)*TX_TINY_BUF_LEN) - -struct cas_tiny_count { - int nbufs; - int used; -}; - -struct cas { - spinlock_t lock; /* for most bits */ - spinlock_t tx_lock[N_TX_RINGS]; /* tx bits */ - spinlock_t stat_lock[N_TX_RINGS + 1]; /* for stat gathering */ - spinlock_t rx_inuse_lock; /* rx inuse list */ - spinlock_t rx_spare_lock; /* rx spare list */ - - void __iomem *regs; - int tx_new[N_TX_RINGS], tx_old[N_TX_RINGS]; - int rx_old[N_RX_DESC_RINGS]; - int rx_cur[N_RX_COMP_RINGS], rx_new[N_RX_COMP_RINGS]; - int rx_last[N_RX_DESC_RINGS]; - - struct napi_struct napi; - - /* Set when chip is actually in operational state - * (ie. not power managed) */ - int hw_running; - int opened; - struct mutex pm_mutex; /* open/close/suspend/resume */ - - struct cas_init_block *init_block; - struct cas_tx_desc *init_txds[MAX_TX_RINGS]; - struct cas_rx_desc *init_rxds[MAX_RX_DESC_RINGS]; - struct cas_rx_comp *init_rxcs[MAX_RX_COMP_RINGS]; - - /* we use sk_buffs for tx and pages for rx. the rx skbuffs - * are there for flow re-assembly. */ - struct sk_buff *tx_skbs[N_TX_RINGS][TX_DESC_RING_SIZE]; - struct sk_buff_head rx_flows[N_RX_FLOWS]; - cas_page_t *rx_pages[N_RX_DESC_RINGS][RX_DESC_RING_SIZE]; - struct list_head rx_spare_list, rx_inuse_list; - int rx_spares_needed; - - /* for small packets when copying would be quicker than - mapping */ - struct cas_tiny_count tx_tiny_use[N_TX_RINGS][TX_DESC_RING_SIZE]; - u8 *tx_tiny_bufs[N_TX_RINGS]; - - u32 msg_enable; - - /* N_TX_RINGS must be >= N_RX_DESC_RINGS */ - struct net_device_stats net_stats[N_TX_RINGS + 1]; - - u32 pci_cfg[64 >> 2]; - u8 pci_revision; - - int phy_type; - int phy_addr; - u32 phy_id; -#define CAS_FLAG_1000MB_CAP 0x00000001 -#define CAS_FLAG_REG_PLUS 0x00000002 -#define CAS_FLAG_TARGET_ABORT 0x00000004 -#define CAS_FLAG_SATURN 0x00000008 -#define CAS_FLAG_RXD_POST_MASK 0x000000F0 -#define CAS_FLAG_RXD_POST_SHIFT 4 -#define CAS_FLAG_RXD_POST(x) ((1 << (CAS_FLAG_RXD_POST_SHIFT + (x))) & \ - CAS_FLAG_RXD_POST_MASK) -#define CAS_FLAG_ENTROPY_DEV 0x00000100 -#define CAS_FLAG_NO_HW_CSUM 0x00000200 - u32 cas_flags; - int packet_min; /* minimum packet size */ - int tx_fifo_size; - int rx_fifo_size; - int rx_pause_off; - int rx_pause_on; - int crc_size; /* 4 if half-duplex */ - - int pci_irq_INTC; - int min_frame_size; /* for tx fifo workaround */ - - /* page size allocation */ - int page_size; - int page_order; - int mtu_stride; - - u32 mac_rx_cfg; - - /* Autoneg & PHY control */ - int link_cntl; - int link_fcntl; - enum link_state lstate; - struct timer_list link_timer; - int timer_ticks; - struct work_struct reset_task; -#if 0 - atomic_t reset_task_pending; -#else - atomic_t reset_task_pending; - atomic_t reset_task_pending_mtu; - atomic_t reset_task_pending_spare; - atomic_t reset_task_pending_all; -#endif - - /* Link-down problem workaround */ -#define LINK_TRANSITION_UNKNOWN 0 -#define LINK_TRANSITION_ON_FAILURE 1 -#define LINK_TRANSITION_STILL_FAILED 2 -#define LINK_TRANSITION_LINK_UP 3 -#define LINK_TRANSITION_LINK_CONFIG 4 -#define LINK_TRANSITION_LINK_DOWN 5 -#define LINK_TRANSITION_REQUESTED_RESET 6 - int link_transition; - int link_transition_jiffies_valid; - unsigned long link_transition_jiffies; - - /* Tuning */ - u8 orig_cacheline_size; /* value when loaded */ -#define CAS_PREF_CACHELINE_SIZE 0x20 /* Minimum desired */ - - /* Diagnostic counters and state. */ - int casreg_len; /* reg-space size for dumping */ - u64 pause_entered; - u16 pause_last_time_recvd; - - dma_addr_t block_dvma, tx_tiny_dvma[N_TX_RINGS]; - struct pci_dev *pdev; - struct net_device *dev; -#if defined(CONFIG_OF) - struct device_node *of_node; -#endif - - /* Firmware Info */ - u16 fw_load_addr; - u32 fw_size; - u8 *fw_data; -}; - -#define TX_DESC_NEXT(r, x) (((x) + 1) & (TX_DESC_RINGN_SIZE(r) - 1)) -#define RX_DESC_ENTRY(r, x) ((x) & (RX_DESC_RINGN_SIZE(r) - 1)) -#define RX_COMP_ENTRY(r, x) ((x) & (RX_COMP_RINGN_SIZE(r) - 1)) - -#define TX_BUFF_COUNT(r, x, y) ((x) <= (y) ? ((y) - (x)) : \ - (TX_DESC_RINGN_SIZE(r) - (x) + (y))) - -#define TX_BUFFS_AVAIL(cp, i) ((cp)->tx_old[(i)] <= (cp)->tx_new[(i)] ? \ - (cp)->tx_old[(i)] + (TX_DESC_RINGN_SIZE(i) - 1) - (cp)->tx_new[(i)] : \ - (cp)->tx_old[(i)] - (cp)->tx_new[(i)] - 1) - -#define CAS_ALIGN(addr, align) \ - (((unsigned long) (addr) + ((align) - 1UL)) & ~((align) - 1)) - -#define RX_FIFO_SIZE 16384 -#define EXPANSION_ROM_SIZE 65536 - -#define CAS_MC_EXACT_MATCH_SIZE 15 -#define CAS_MC_HASH_SIZE 256 -#define CAS_MC_HASH_MAX (CAS_MC_EXACT_MATCH_SIZE + \ - CAS_MC_HASH_SIZE) - -#define TX_TARGET_ABORT_LEN 0x20 -#define RX_SWIVEL_OFF_VAL 0x2 -#define RX_AE_FREEN_VAL(x) (RX_DESC_RINGN_SIZE(x) >> 1) -#define RX_AE_COMP_VAL (RX_COMP_RING_SIZE >> 1) -#define RX_BLANK_INTR_PKT_VAL 0x05 -#define RX_BLANK_INTR_TIME_VAL 0x0F -#define HP_TCP_THRESH_VAL 1530 /* reduce to enable reassembly */ - -#define RX_SPARE_COUNT (RX_DESC_RING_SIZE >> 1) -#define RX_SPARE_RECOVER_VAL (RX_SPARE_COUNT >> 2) - -#endif /* _CASSINI_H */ From patchwork Fri Jan 6 22:00:15 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Anirudh Venkataramanan X-Patchwork-Id: 13091838 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id D4EC0C678DA for ; Fri, 6 Jan 2023 21:59:24 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236427AbjAFV7X (ORCPT ); Fri, 6 Jan 2023 16:59:23 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:42322 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234216AbjAFV7U (ORCPT ); Fri, 6 Jan 2023 16:59:20 -0500 Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id CBE53728AD; Fri, 6 Jan 2023 13:59:19 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1673042359; x=1704578359; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=TG7M7IO4ZvGwossk/dYz2Y4AzoE+L3tw+oWib10OPcY=; b=nPl9yFLbhZaXSjVC+bhLEfhaZQ34AQ2zowslFCo3vaSI2vecw/vstnfm 0OffP2CiZ6PsVinvya8f+zrQnNqTL7eR1GGbRCaVWdqxEIkcHG+zMrhbD kPsKi1uatCqgR0Jt74hQIWnURdbuR8/T/YJSoliaQ9qMeekWzcZSaQL4j WI2AzgRfp3WnXhUZ6soAwDyLqNtDR/Oy3v/Azf+MCczgJsZIlqfecnjyJ +2YEaRgYCeiYY/ZbWwv2lTcNFq8OcXe0MmwW3BMIaQEGLC5jxaunyj6I9 ijdK9hOYeEHcMMrXDCbXJSzkA/69TBsrziJmjL0IiqT4nZBBt+rvuFLh+ Q==; X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="387030705" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="387030705" Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Jan 2023 13:59:17 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="763652881" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="763652881" Received: from avenkata-desk0.sc.intel.com ([172.25.112.60]) by fmsmga002.fm.intel.com with ESMTP; 06 Jan 2023 13:59:16 -0800 From: Anirudh Venkataramanan To: netdev@vger.kernel.org Cc: linux-pci@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-mips@vger.kernel.org, linux-trace-kernel@vger.kernel.org, sparclinux@vger.kernel.org, Anirudh Venkataramanan , Leon Romanovsky Subject: [PATCH net-next 2/7] PCI: Remove PCI IDs used by the Sun Cassini driver Date: Fri, 6 Jan 2023 14:00:15 -0800 Message-Id: <20230106220020.1820147-3-anirudh.venkataramanan@intel.com> X-Mailer: git-send-email 2.37.2 In-Reply-To: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> References: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org The previous patch removed the Cassini driver (drivers/net/ethernet/sun). With this, PCI_DEVICE_ID_NS_SATURN and PCI_DEVICE_ID_SUN_CASSINI are unused. Remove them. Cc: Leon Romanovsky Signed-off-by: Anirudh Venkataramanan --- include/linux/pci_ids.h | 2 -- 1 file changed, 2 deletions(-) diff --git a/include/linux/pci_ids.h b/include/linux/pci_ids.h index b362d90..eca2340 100644 --- a/include/linux/pci_ids.h +++ b/include/linux/pci_ids.h @@ -433,7 +433,6 @@ #define PCI_DEVICE_ID_NS_CS5535_AUDIO 0x002e #define PCI_DEVICE_ID_NS_CS5535_USB 0x002f #define PCI_DEVICE_ID_NS_GX_VIDEO 0x0030 -#define PCI_DEVICE_ID_NS_SATURN 0x0035 #define PCI_DEVICE_ID_NS_SCx200_BRIDGE 0x0500 #define PCI_DEVICE_ID_NS_SCx200_SMI 0x0501 #define PCI_DEVICE_ID_NS_SCx200_IDE 0x0502 @@ -1047,7 +1046,6 @@ #define PCI_DEVICE_ID_SUN_SABRE 0xa000 #define PCI_DEVICE_ID_SUN_HUMMINGBIRD 0xa001 #define PCI_DEVICE_ID_SUN_TOMATILLO 0xa801 -#define PCI_DEVICE_ID_SUN_CASSINI 0xabba #define PCI_VENDOR_ID_NI 0x1093 #define PCI_DEVICE_ID_NI_PCI2322 0xd130 From patchwork Fri Jan 6 22:00:16 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Anirudh Venkataramanan X-Patchwork-Id: 13091839 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id B3235C6379F for ; Fri, 6 Jan 2023 21:59:26 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236359AbjAFV7Y (ORCPT ); Fri, 6 Jan 2023 16:59:24 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:42324 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S233616AbjAFV7T (ORCPT ); Fri, 6 Jan 2023 16:59:19 -0500 Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id B9081676DE; Fri, 6 Jan 2023 13:59:18 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1673042359; x=1704578359; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=YxmNblIsKqhfyJq7N8D9hjsKeTOrNyEjlTYDV33bZjo=; b=VXX7vg31fAd/2yeyvsuO2tyhbOQJHg914hhCBY9Eg765noFLqsG0A/YW xVa4i/hJFBXkzgoBoKVCq3yO4bJiWXmtZIlwNoOPCsEbCGXyTtD5ZR/NO /gZDFo8d++of32H43g4nSruw/rX3p0XwjSxUUCAEGy4roMiSrxB9uI1Ap mhnZsLbNpobq4LkSObZMNrhYaP5fFAkLq4eTjF0ywRYe2odObbm9aKSO9 1TUY4Ia3fbGc+io6AKw76f6mKANiWpASX9VY6w8Aohg550kDj4RL/hclk IukSkqYYrXe27uOf/0RL7dEY27b6BgwYwmvhZvTECgIHxzv+Ddyhdb6eK A==; X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="387030707" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="387030707" Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Jan 2023 13:59:18 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="763652885" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="763652885" Received: from avenkata-desk0.sc.intel.com ([172.25.112.60]) by fmsmga002.fm.intel.com with ESMTP; 06 Jan 2023 13:59:16 -0800 From: Anirudh Venkataramanan To: netdev@vger.kernel.org Cc: linux-pci@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-mips@vger.kernel.org, linux-trace-kernel@vger.kernel.org, sparclinux@vger.kernel.org, Anirudh Venkataramanan , Leon Romanovsky Subject: [PATCH net-next 3/7] powerpc: configs: Remove reference to CONFIG_CASSINI Date: Fri, 6 Jan 2023 14:00:16 -0800 Message-Id: <20230106220020.1820147-4-anirudh.venkataramanan@intel.com> X-Mailer: git-send-email 2.37.2 In-Reply-To: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> References: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org An earlier patch removed the Sun Cassnini driver. Remove references to CONFIG_CASSINI from the ppc6xx defconfig. Cc: Leon Romanovsky Signed-off-by: Anirudh Venkataramanan --- arch/powerpc/configs/ppc6xx_defconfig | 1 - 1 file changed, 1 deletion(-) diff --git a/arch/powerpc/configs/ppc6xx_defconfig b/arch/powerpc/configs/ppc6xx_defconfig index 1102582..0eba1de 100644 --- a/arch/powerpc/configs/ppc6xx_defconfig +++ b/arch/powerpc/configs/ppc6xx_defconfig @@ -487,7 +487,6 @@ CONFIG_PCMCIA_SMC91C92=m CONFIG_EPIC100=m CONFIG_HAPPYMEAL=m CONFIG_SUNGEM=m -CONFIG_CASSINI=m CONFIG_NIU=m CONFIG_TEHUTI=m CONFIG_TLAN=m From patchwork Fri Jan 6 22:00:17 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Anirudh Venkataramanan X-Patchwork-Id: 13091840 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 83945C678D8 for ; Fri, 6 Jan 2023 21:59:29 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236549AbjAFV70 (ORCPT ); Fri, 6 Jan 2023 16:59:26 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:42340 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235841AbjAFV7U (ORCPT ); Fri, 6 Jan 2023 16:59:20 -0500 Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 140726B5F0; Fri, 6 Jan 2023 13:59:20 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1673042360; x=1704578360; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=sQb3sEVMUrtQ11r87eT7TSDBLNdId+qUxBQzgXh/Yr4=; b=AJGaUUeLiVw+MSnV9EoFlzWtud8NDBlwuLKrtYNmM7XkSEKP7V/b6HJc ya6OxH062uVVl+SLJ+BeDZBhHg0LmHfQMMjZQPpGlEijjvodvl6soaswp FjXYgLQ4dEcjUv4pOx4guGGvFZ1D2AwaPnn37OkCncuCeGWKGdkPu6XME jLQEBGtLuKmroxafRqIAVLNUAR+2w48cubP8brLpLiHcMGsMZahHEEUfU GW7gC0P5BUmcrVdcxapFddAExhnaZ4GRHaVGI43k5r7IhjxlG8KAf4kOv +ZMB/jMWQGeIDwttaWmPJ33HJIWaBD3QsFD3WxoTl62kJko9IXol7hbSg Q==; X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="387030715" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="387030715" Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Jan 2023 13:59:18 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="763652888" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="763652888" Received: from avenkata-desk0.sc.intel.com ([172.25.112.60]) by fmsmga002.fm.intel.com with ESMTP; 06 Jan 2023 13:59:16 -0800 From: Anirudh Venkataramanan To: netdev@vger.kernel.org Cc: linux-pci@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-mips@vger.kernel.org, linux-trace-kernel@vger.kernel.org, sparclinux@vger.kernel.org, Anirudh Venkataramanan , Leon Romanovsky Subject: [PATCH net-next 4/7] mips: configs: Remove reference to CONFIG_CASSINI Date: Fri, 6 Jan 2023 14:00:17 -0800 Message-Id: <20230106220020.1820147-5-anirudh.venkataramanan@intel.com> X-Mailer: git-send-email 2.37.2 In-Reply-To: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> References: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org An earlier patch removed the Sun Cassnini driver. Remove references to CONFIG_CASSINI from the mtx1 defconfig Cc: Leon Romanovsky Signed-off-by: Anirudh Venkataramanan --- arch/mips/configs/mtx1_defconfig | 1 - 1 file changed, 1 deletion(-) diff --git a/arch/mips/configs/mtx1_defconfig b/arch/mips/configs/mtx1_defconfig index 89a1511..17d88a0 100644 --- a/arch/mips/configs/mtx1_defconfig +++ b/arch/mips/configs/mtx1_defconfig @@ -305,7 +305,6 @@ CONFIG_PCMCIA_SMC91C92=m CONFIG_EPIC100=m CONFIG_HAPPYMEAL=m CONFIG_SUNGEM=m -CONFIG_CASSINI=m CONFIG_TLAN=m CONFIG_VIA_RHINE=m CONFIG_VIA_VELOCITY=m From patchwork Fri Jan 6 22:00:18 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Anirudh Venkataramanan X-Patchwork-Id: 13091843 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id C5491C6379F for ; Fri, 6 Jan 2023 21:59:39 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236445AbjAFV7h (ORCPT ); Fri, 6 Jan 2023 16:59:37 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:42362 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S236459AbjAFV7Y (ORCPT ); Fri, 6 Jan 2023 16:59:24 -0500 Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 0A132676DE; Fri, 6 Jan 2023 13:59:21 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1673042361; x=1704578361; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=2ArlPq/P9KTReo8PRrrSolVsFQatvp06LbQjTmSHVJc=; b=EW7ZkKLNYW0BOMTMw493P6tOclgjj/1HiSrWex4W/nNgYtk/ngrzF3yS mBnr8QfazGeYsuW34GqQvXXypR/is1knPpByRsAUwEwRipTjddK60cDcW Yn+AIk3gZCMVa1Zf4P3mYf1TF/ysbj37qS32jfWBns+5qPlzSvUUPYUhF V2s3ag60M6n2UJacm0JeqJOgrv/Cdbr6JRBcheaf0JbDto1Rxlyf6yksi kDuiBoQfluiyCk9iTVisIY4RoC2/r+2XywAPoq3vhzliYog7avsc1q0dX Lw1SIoHxhEAm0cQ4rq9B2kPQGlEM++eRRcfjSalC9Uk/I6xyQK5O76iRd w==; X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="387030717" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="387030717" Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Jan 2023 13:59:18 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="763652892" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="763652892" Received: from avenkata-desk0.sc.intel.com ([172.25.112.60]) by fmsmga002.fm.intel.com with ESMTP; 06 Jan 2023 13:59:16 -0800 From: Anirudh Venkataramanan To: netdev@vger.kernel.org Cc: linux-pci@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-mips@vger.kernel.org, linux-trace-kernel@vger.kernel.org, sparclinux@vger.kernel.org, Anirudh Venkataramanan , Leon Romanovsky Subject: [PATCH net-next 5/7] ethernet: Remove the Sun LDOM vswitch and sunvnet drivers Date: Fri, 6 Jan 2023 14:00:18 -0800 Message-Id: <20230106220020.1820147-6-anirudh.venkataramanan@intel.com> X-Mailer: git-send-email 2.37.2 In-Reply-To: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> References: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org In a recent patch series that touched these drivers [1], it was suggested that these drivers should be removed completely. git logs suggest that there hasn't been any significant feature addition, improvement or fixes to user-visible bugs in a while. A web search didn't indicate any recent discussions or any evidence that there are users out there who care about these drivers, and so remove these drivers. Notes: checkpatch complains "WARNING: added, moved or deleted file(s), does MAINTAINERS need updating?". The files being removed don't have their own entries in the MAINTAINERS file, so there's nothing to remove. checkpatch also complains about the long lore link below. [1] https://lore.kernel.org/netdev/99629223-ac1b-0f82-50b8-ea307b3b0197@intel.com/T/#t Suggested-by: Leon Romanovsky Signed-off-by: Anirudh Venkataramanan --- drivers/net/ethernet/sun/Kconfig | 27 - drivers/net/ethernet/sun/Makefile | 3 - drivers/net/ethernet/sun/ldmvsw.c | 476 ------ drivers/net/ethernet/sun/sunvnet.c | 567 ------- drivers/net/ethernet/sun/sunvnet_common.c | 1813 --------------------- drivers/net/ethernet/sun/sunvnet_common.h | 157 -- 6 files changed, 3043 deletions(-) delete mode 100644 drivers/net/ethernet/sun/ldmvsw.c delete mode 100644 drivers/net/ethernet/sun/sunvnet.c delete mode 100644 drivers/net/ethernet/sun/sunvnet_common.c delete mode 100644 drivers/net/ethernet/sun/sunvnet_common.h diff --git a/drivers/net/ethernet/sun/Kconfig b/drivers/net/ethernet/sun/Kconfig index 57eb168..26472e4 100644 --- a/drivers/net/ethernet/sun/Kconfig +++ b/drivers/net/ethernet/sun/Kconfig @@ -62,33 +62,6 @@ config SUNGEM Support for the Sun GEM chip, aka Sun GigabitEthernet/P 2.0. See also . -config SUNVNET_COMMON - tristate "Common routines to support Sun Virtual Networking" - depends on SUN_LDOMS - depends on INET - default m - -config SUNVNET - tristate "Sun Virtual Network support" - default m - depends on SUN_LDOMS - depends on SUNVNET_COMMON - help - Support for virtual network devices under Sun Logical Domains. - -config LDMVSW - tristate "Sun4v LDoms Virtual Switch support" - default m - depends on SUN_LDOMS - depends on SUNVNET_COMMON - help - Support for virtual switch devices under Sun4v Logical Domains. - This driver adds a network interface for every vsw-port node - found in the machine description of a service domain. - Linux bridge/switch software can use these interfaces for - guest domain network interconnectivity or guest domain - connection to a physical network on a service domain. - config NIU tristate "Sun Neptune 10Gbit Ethernet support" depends on PCI diff --git a/drivers/net/ethernet/sun/Makefile b/drivers/net/ethernet/sun/Makefile index fd77810..35469d1 100644 --- a/drivers/net/ethernet/sun/Makefile +++ b/drivers/net/ethernet/sun/Makefile @@ -7,7 +7,4 @@ obj-$(CONFIG_HAPPYMEAL) += sunhme.o obj-$(CONFIG_SUNQE) += sunqe.o obj-$(CONFIG_SUNBMAC) += sunbmac.o obj-$(CONFIG_SUNGEM) += sungem.o -obj-$(CONFIG_SUNVNET_COMMON) += sunvnet_common.o -obj-$(CONFIG_SUNVNET) += sunvnet.o -obj-$(CONFIG_LDMVSW) += ldmvsw.o obj-$(CONFIG_NIU) += niu.o diff --git a/drivers/net/ethernet/sun/ldmvsw.c b/drivers/net/ethernet/sun/ldmvsw.c deleted file mode 100644 index 8addee6..00000000 --- a/drivers/net/ethernet/sun/ldmvsw.c +++ /dev/null @@ -1,476 +0,0 @@ -// SPDX-License-Identifier: GPL-2.0 -/* ldmvsw.c: Sun4v LDOM Virtual Switch Driver. - * - * Copyright (C) 2016-2017 Oracle. All rights reserved. - */ - -#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt - -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include - -#if defined(CONFIG_IPV6) -#include -#endif - -#include -#include -#include - -#include -#include - -/* This driver makes use of the common code in sunvnet_common.c */ -#include "sunvnet_common.h" - -/* Length of time before we decide the hardware is hung, - * and dev->tx_timeout() should be called to fix the problem. - */ -#define VSW_TX_TIMEOUT (10 * HZ) - -/* Static HW Addr used for the network interfaces representing vsw ports */ -static u8 vsw_port_hwaddr[ETH_ALEN] = {0xFE, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF}; - -#define DRV_MODULE_NAME "ldmvsw" -#define DRV_MODULE_VERSION "1.2" -#define DRV_MODULE_RELDATE "March 4, 2017" - -static char version[] = - DRV_MODULE_NAME " " DRV_MODULE_VERSION " (" DRV_MODULE_RELDATE ")"; -MODULE_AUTHOR("Oracle"); -MODULE_DESCRIPTION("Sun4v LDOM Virtual Switch Driver"); -MODULE_LICENSE("GPL"); -MODULE_VERSION(DRV_MODULE_VERSION); - -/* Ordered from largest major to lowest */ -static struct vio_version vsw_versions[] = { - { .major = 1, .minor = 8 }, - { .major = 1, .minor = 7 }, - { .major = 1, .minor = 6 }, - { .major = 1, .minor = 0 }, -}; - -static void vsw_get_drvinfo(struct net_device *dev, - struct ethtool_drvinfo *info) -{ - strscpy(info->driver, DRV_MODULE_NAME, sizeof(info->driver)); - strscpy(info->version, DRV_MODULE_VERSION, sizeof(info->version)); -} - -static u32 vsw_get_msglevel(struct net_device *dev) -{ - struct vnet_port *port = netdev_priv(dev); - - return port->vp->msg_enable; -} - -static void vsw_set_msglevel(struct net_device *dev, u32 value) -{ - struct vnet_port *port = netdev_priv(dev); - - port->vp->msg_enable = value; -} - -static const struct ethtool_ops vsw_ethtool_ops = { - .get_drvinfo = vsw_get_drvinfo, - .get_msglevel = vsw_get_msglevel, - .set_msglevel = vsw_set_msglevel, - .get_link = ethtool_op_get_link, -}; - -static LIST_HEAD(vnet_list); -static DEFINE_MUTEX(vnet_list_mutex); - -/* func arg to vnet_start_xmit_common() to get the proper tx port */ -static struct vnet_port *vsw_tx_port_find(struct sk_buff *skb, - struct net_device *dev) -{ - struct vnet_port *port = netdev_priv(dev); - - return port; -} - -static u16 vsw_select_queue(struct net_device *dev, struct sk_buff *skb, - struct net_device *sb_dev) -{ - struct vnet_port *port = netdev_priv(dev); - - if (!port) - return 0; - - return port->q_index; -} - -/* Wrappers to common functions */ -static netdev_tx_t vsw_start_xmit(struct sk_buff *skb, struct net_device *dev) -{ - return sunvnet_start_xmit_common(skb, dev, vsw_tx_port_find); -} - -static void vsw_set_rx_mode(struct net_device *dev) -{ - struct vnet_port *port = netdev_priv(dev); - - return sunvnet_set_rx_mode_common(dev, port->vp); -} - -int ldmvsw_open(struct net_device *dev) -{ - struct vnet_port *port = netdev_priv(dev); - struct vio_driver_state *vio = &port->vio; - - /* reset the channel */ - vio_link_state_change(vio, LDC_EVENT_RESET); - vnet_port_reset(port); - vio_port_up(vio); - - return 0; -} -EXPORT_SYMBOL_GPL(ldmvsw_open); - -#ifdef CONFIG_NET_POLL_CONTROLLER -static void vsw_poll_controller(struct net_device *dev) -{ - struct vnet_port *port = netdev_priv(dev); - - return sunvnet_poll_controller_common(dev, port->vp); -} -#endif - -static const struct net_device_ops vsw_ops = { - .ndo_open = ldmvsw_open, - .ndo_stop = sunvnet_close_common, - .ndo_set_rx_mode = vsw_set_rx_mode, - .ndo_set_mac_address = sunvnet_set_mac_addr_common, - .ndo_validate_addr = eth_validate_addr, - .ndo_tx_timeout = sunvnet_tx_timeout_common, - .ndo_start_xmit = vsw_start_xmit, - .ndo_select_queue = vsw_select_queue, -#ifdef CONFIG_NET_POLL_CONTROLLER - .ndo_poll_controller = vsw_poll_controller, -#endif -}; - -static const char *local_mac_prop = "local-mac-address"; -static const char *cfg_handle_prop = "cfg-handle"; - -static struct vnet *vsw_get_vnet(struct mdesc_handle *hp, - u64 port_node, - u64 *handle) -{ - struct vnet *vp; - struct vnet *iter; - const u64 *local_mac = NULL; - const u64 *cfghandle = NULL; - u64 a; - - /* Get the parent virtual-network-switch macaddr and cfghandle */ - mdesc_for_each_arc(a, hp, port_node, MDESC_ARC_TYPE_BACK) { - u64 target = mdesc_arc_target(hp, a); - const char *name; - - name = mdesc_get_property(hp, target, "name", NULL); - if (!name || strcmp(name, "virtual-network-switch")) - continue; - - local_mac = mdesc_get_property(hp, target, - local_mac_prop, NULL); - cfghandle = mdesc_get_property(hp, target, - cfg_handle_prop, NULL); - break; - } - if (!local_mac || !cfghandle) - return ERR_PTR(-ENODEV); - - /* find or create associated vnet */ - vp = NULL; - mutex_lock(&vnet_list_mutex); - list_for_each_entry(iter, &vnet_list, list) { - if (iter->local_mac == *local_mac) { - vp = iter; - break; - } - } - - if (!vp) { - vp = kzalloc(sizeof(*vp), GFP_KERNEL); - if (unlikely(!vp)) { - mutex_unlock(&vnet_list_mutex); - return ERR_PTR(-ENOMEM); - } - - spin_lock_init(&vp->lock); - INIT_LIST_HEAD(&vp->port_list); - INIT_LIST_HEAD(&vp->list); - vp->local_mac = *local_mac; - list_add(&vp->list, &vnet_list); - } - - mutex_unlock(&vnet_list_mutex); - - *handle = (u64)*cfghandle; - - return vp; -} - -static struct net_device *vsw_alloc_netdev(u8 hwaddr[], - struct vio_dev *vdev, - u64 handle, - u64 port_id) -{ - struct net_device *dev; - struct vnet_port *port; - - dev = alloc_etherdev_mqs(sizeof(*port), VNET_MAX_TXQS, 1); - if (!dev) - return ERR_PTR(-ENOMEM); - dev->needed_headroom = VNET_PACKET_SKIP + 8; - dev->needed_tailroom = 8; - - eth_hw_addr_set(dev, hwaddr); - ether_addr_copy(dev->perm_addr, dev->dev_addr); - - sprintf(dev->name, "vif%d.%d", (int)handle, (int)port_id); - - dev->netdev_ops = &vsw_ops; - dev->ethtool_ops = &vsw_ethtool_ops; - dev->watchdog_timeo = VSW_TX_TIMEOUT; - - dev->hw_features = NETIF_F_HW_CSUM | NETIF_F_SG; - dev->features = dev->hw_features; - - /* MTU range: 68 - 65535 */ - dev->min_mtu = ETH_MIN_MTU; - dev->max_mtu = VNET_MAX_MTU; - - SET_NETDEV_DEV(dev, &vdev->dev); - - return dev; -} - -static struct ldc_channel_config vsw_ldc_cfg = { - .event = sunvnet_event_common, - .mtu = 64, - .mode = LDC_MODE_UNRELIABLE, -}; - -static struct vio_driver_ops vsw_vio_ops = { - .send_attr = sunvnet_send_attr_common, - .handle_attr = sunvnet_handle_attr_common, - .handshake_complete = sunvnet_handshake_complete_common, -}; - -static const char *remote_macaddr_prop = "remote-mac-address"; -static const char *id_prop = "id"; - -static int vsw_port_probe(struct vio_dev *vdev, const struct vio_device_id *id) -{ - struct mdesc_handle *hp; - struct vnet_port *port; - unsigned long flags; - struct vnet *vp; - struct net_device *dev; - const u64 *rmac; - int len, i, err; - const u64 *port_id; - u64 handle; - - hp = mdesc_grab(); - - rmac = mdesc_get_property(hp, vdev->mp, remote_macaddr_prop, &len); - err = -ENODEV; - if (!rmac) { - pr_err("Port lacks %s property\n", remote_macaddr_prop); - mdesc_release(hp); - return err; - } - - port_id = mdesc_get_property(hp, vdev->mp, id_prop, NULL); - err = -ENODEV; - if (!port_id) { - pr_err("Port lacks %s property\n", id_prop); - mdesc_release(hp); - return err; - } - - /* Get (or create) the vnet associated with this port */ - vp = vsw_get_vnet(hp, vdev->mp, &handle); - if (IS_ERR(vp)) { - err = PTR_ERR(vp); - pr_err("Failed to get vnet for vsw-port\n"); - mdesc_release(hp); - return err; - } - - mdesc_release(hp); - - dev = vsw_alloc_netdev(vsw_port_hwaddr, vdev, handle, *port_id); - if (IS_ERR(dev)) { - err = PTR_ERR(dev); - pr_err("Failed to alloc netdev for vsw-port\n"); - return err; - } - - port = netdev_priv(dev); - - INIT_LIST_HEAD(&port->list); - - for (i = 0; i < ETH_ALEN; i++) - port->raddr[i] = (*rmac >> (5 - i) * 8) & 0xff; - - port->vp = vp; - port->dev = dev; - port->switch_port = 1; - port->tso = false; /* no tso in vsw, misbehaves in bridge */ - port->tsolen = 0; - - /* Mark the port as belonging to ldmvsw which directs the - * common code to use the net_device in the vnet_port - * rather than the net_device in the vnet (which is used - * by sunvnet). This bit is used by the VNET_PORT_TO_NET_DEVICE - * macro. - */ - port->vsw = 1; - - err = vio_driver_init(&port->vio, vdev, VDEV_NETWORK, - vsw_versions, ARRAY_SIZE(vsw_versions), - &vsw_vio_ops, dev->name); - if (err) - goto err_out_free_dev; - - err = vio_ldc_alloc(&port->vio, &vsw_ldc_cfg, port); - if (err) - goto err_out_free_dev; - - dev_set_drvdata(&vdev->dev, port); - - netif_napi_add(dev, &port->napi, sunvnet_poll_common); - - spin_lock_irqsave(&vp->lock, flags); - list_add_rcu(&port->list, &vp->port_list); - spin_unlock_irqrestore(&vp->lock, flags); - - timer_setup(&port->clean_timer, sunvnet_clean_timer_expire_common, 0); - - err = register_netdev(dev); - if (err) { - pr_err("Cannot register net device, aborting\n"); - goto err_out_del_timer; - } - - spin_lock_irqsave(&vp->lock, flags); - sunvnet_port_add_txq_common(port); - spin_unlock_irqrestore(&vp->lock, flags); - - napi_enable(&port->napi); - vio_port_up(&port->vio); - - /* assure no carrier until we receive an LDC_EVENT_UP, - * even if the vsw config script tries to force us up - */ - netif_carrier_off(dev); - - netdev_info(dev, "LDOM vsw-port %pM\n", dev->dev_addr); - - pr_info("%s: PORT ( remote-mac %pM%s )\n", dev->name, - port->raddr, " switch-port"); - - return 0; - -err_out_del_timer: - del_timer_sync(&port->clean_timer); - list_del_rcu(&port->list); - synchronize_rcu(); - netif_napi_del(&port->napi); - dev_set_drvdata(&vdev->dev, NULL); - vio_ldc_free(&port->vio); - -err_out_free_dev: - free_netdev(dev); - return err; -} - -static void vsw_port_remove(struct vio_dev *vdev) -{ - struct vnet_port *port = dev_get_drvdata(&vdev->dev); - unsigned long flags; - - if (port) { - del_timer_sync(&port->vio.timer); - del_timer_sync(&port->clean_timer); - - napi_disable(&port->napi); - unregister_netdev(port->dev); - - list_del_rcu(&port->list); - - synchronize_rcu(); - spin_lock_irqsave(&port->vp->lock, flags); - sunvnet_port_rm_txq_common(port); - spin_unlock_irqrestore(&port->vp->lock, flags); - netif_napi_del(&port->napi); - sunvnet_port_free_tx_bufs_common(port); - vio_ldc_free(&port->vio); - - dev_set_drvdata(&vdev->dev, NULL); - - free_netdev(port->dev); - } -} - -static void vsw_cleanup(void) -{ - struct vnet *vp; - - /* just need to free up the vnet list */ - mutex_lock(&vnet_list_mutex); - while (!list_empty(&vnet_list)) { - vp = list_first_entry(&vnet_list, struct vnet, list); - list_del(&vp->list); - /* vio_unregister_driver() should have cleaned up port_list */ - if (!list_empty(&vp->port_list)) - pr_err("Ports not removed by VIO subsystem!\n"); - kfree(vp); - } - mutex_unlock(&vnet_list_mutex); -} - -static const struct vio_device_id vsw_port_match[] = { - { - .type = "vsw-port", - }, - {}, -}; -MODULE_DEVICE_TABLE(vio, vsw_port_match); - -static struct vio_driver vsw_port_driver = { - .id_table = vsw_port_match, - .probe = vsw_port_probe, - .remove = vsw_port_remove, - .name = "vsw_port", -}; - -static int __init vsw_init(void) -{ - pr_info("%s\n", version); - return vio_register_driver(&vsw_port_driver); -} - -static void __exit vsw_exit(void) -{ - vio_unregister_driver(&vsw_port_driver); - vsw_cleanup(); -} - -module_init(vsw_init); -module_exit(vsw_exit); diff --git a/drivers/net/ethernet/sun/sunvnet.c b/drivers/net/ethernet/sun/sunvnet.c deleted file mode 100644 index fe86fbd..00000000 --- a/drivers/net/ethernet/sun/sunvnet.c +++ /dev/null @@ -1,567 +0,0 @@ -// SPDX-License-Identifier: GPL-2.0 -/* sunvnet.c: Sun LDOM Virtual Network Driver. - * - * Copyright (C) 2007, 2008 David S. Miller - * Copyright (C) 2016-2017 Oracle. All rights reserved. - */ - -#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt - -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include - -#if IS_ENABLED(CONFIG_IPV6) -#include -#endif - -#include -#include -#include - -#include -#include - -#include "sunvnet_common.h" - -/* length of time before we decide the hardware is borked, - * and dev->tx_timeout() should be called to fix the problem - */ -#define VNET_TX_TIMEOUT (5 * HZ) - -#define DRV_MODULE_NAME "sunvnet" -#define DRV_MODULE_VERSION "2.0" -#define DRV_MODULE_RELDATE "February 3, 2017" - -static char version[] = - DRV_MODULE_NAME " " DRV_MODULE_VERSION " (" DRV_MODULE_RELDATE ")"; -MODULE_AUTHOR("David S. Miller (davem@davemloft.net)"); -MODULE_DESCRIPTION("Sun LDOM virtual network driver"); -MODULE_LICENSE("GPL"); -MODULE_VERSION(DRV_MODULE_VERSION); - -/* Ordered from largest major to lowest */ -static struct vio_version vnet_versions[] = { - { .major = 1, .minor = 8 }, - { .major = 1, .minor = 7 }, - { .major = 1, .minor = 6 }, - { .major = 1, .minor = 0 }, -}; - -static void vnet_get_drvinfo(struct net_device *dev, - struct ethtool_drvinfo *info) -{ - strscpy(info->driver, DRV_MODULE_NAME, sizeof(info->driver)); - strscpy(info->version, DRV_MODULE_VERSION, sizeof(info->version)); -} - -static u32 vnet_get_msglevel(struct net_device *dev) -{ - struct vnet *vp = netdev_priv(dev); - - return vp->msg_enable; -} - -static void vnet_set_msglevel(struct net_device *dev, u32 value) -{ - struct vnet *vp = netdev_priv(dev); - - vp->msg_enable = value; -} - -static const struct { - const char string[ETH_GSTRING_LEN]; -} ethtool_stats_keys[] = { - { "rx_packets" }, - { "tx_packets" }, - { "rx_bytes" }, - { "tx_bytes" }, - { "rx_errors" }, - { "tx_errors" }, - { "rx_dropped" }, - { "tx_dropped" }, - { "multicast" }, - { "rx_length_errors" }, - { "rx_frame_errors" }, - { "rx_missed_errors" }, - { "tx_carrier_errors" }, - { "nports" }, -}; - -static int vnet_get_sset_count(struct net_device *dev, int sset) -{ - struct vnet *vp = (struct vnet *)netdev_priv(dev); - - switch (sset) { - case ETH_SS_STATS: - return ARRAY_SIZE(ethtool_stats_keys) - + (NUM_VNET_PORT_STATS * vp->nports); - default: - return -EOPNOTSUPP; - } -} - -static void vnet_get_strings(struct net_device *dev, u32 stringset, u8 *buf) -{ - struct vnet *vp = (struct vnet *)netdev_priv(dev); - struct vnet_port *port; - char *p = (char *)buf; - - switch (stringset) { - case ETH_SS_STATS: - memcpy(buf, ðtool_stats_keys, sizeof(ethtool_stats_keys)); - p += sizeof(ethtool_stats_keys); - - rcu_read_lock(); - list_for_each_entry_rcu(port, &vp->port_list, list) { - snprintf(p, ETH_GSTRING_LEN, "p%u.%s-%pM", - port->q_index, port->switch_port ? "s" : "q", - port->raddr); - p += ETH_GSTRING_LEN; - snprintf(p, ETH_GSTRING_LEN, "p%u.rx_packets", - port->q_index); - p += ETH_GSTRING_LEN; - snprintf(p, ETH_GSTRING_LEN, "p%u.tx_packets", - port->q_index); - p += ETH_GSTRING_LEN; - snprintf(p, ETH_GSTRING_LEN, "p%u.rx_bytes", - port->q_index); - p += ETH_GSTRING_LEN; - snprintf(p, ETH_GSTRING_LEN, "p%u.tx_bytes", - port->q_index); - p += ETH_GSTRING_LEN; - snprintf(p, ETH_GSTRING_LEN, "p%u.event_up", - port->q_index); - p += ETH_GSTRING_LEN; - snprintf(p, ETH_GSTRING_LEN, "p%u.event_reset", - port->q_index); - p += ETH_GSTRING_LEN; - } - rcu_read_unlock(); - break; - default: - WARN_ON(1); - break; - } -} - -static void vnet_get_ethtool_stats(struct net_device *dev, - struct ethtool_stats *estats, u64 *data) -{ - struct vnet *vp = (struct vnet *)netdev_priv(dev); - struct vnet_port *port; - int i = 0; - - data[i++] = dev->stats.rx_packets; - data[i++] = dev->stats.tx_packets; - data[i++] = dev->stats.rx_bytes; - data[i++] = dev->stats.tx_bytes; - data[i++] = dev->stats.rx_errors; - data[i++] = dev->stats.tx_errors; - data[i++] = dev->stats.rx_dropped; - data[i++] = dev->stats.tx_dropped; - data[i++] = dev->stats.multicast; - data[i++] = dev->stats.rx_length_errors; - data[i++] = dev->stats.rx_frame_errors; - data[i++] = dev->stats.rx_missed_errors; - data[i++] = dev->stats.tx_carrier_errors; - data[i++] = vp->nports; - - rcu_read_lock(); - list_for_each_entry_rcu(port, &vp->port_list, list) { - data[i++] = port->q_index; - data[i++] = port->stats.rx_packets; - data[i++] = port->stats.tx_packets; - data[i++] = port->stats.rx_bytes; - data[i++] = port->stats.tx_bytes; - data[i++] = port->stats.event_up; - data[i++] = port->stats.event_reset; - } - rcu_read_unlock(); -} - -static const struct ethtool_ops vnet_ethtool_ops = { - .get_drvinfo = vnet_get_drvinfo, - .get_msglevel = vnet_get_msglevel, - .set_msglevel = vnet_set_msglevel, - .get_link = ethtool_op_get_link, - .get_sset_count = vnet_get_sset_count, - .get_strings = vnet_get_strings, - .get_ethtool_stats = vnet_get_ethtool_stats, -}; - -static LIST_HEAD(vnet_list); -static DEFINE_MUTEX(vnet_list_mutex); - -static struct vnet_port *__tx_port_find(struct vnet *vp, struct sk_buff *skb) -{ - unsigned int hash = vnet_hashfn(skb->data); - struct hlist_head *hp = &vp->port_hash[hash]; - struct vnet_port *port; - - hlist_for_each_entry_rcu(port, hp, hash) { - if (!sunvnet_port_is_up_common(port)) - continue; - if (ether_addr_equal(port->raddr, skb->data)) - return port; - } - list_for_each_entry_rcu(port, &vp->port_list, list) { - if (!port->switch_port) - continue; - if (!sunvnet_port_is_up_common(port)) - continue; - return port; - } - return NULL; -} - -/* func arg to vnet_start_xmit_common() to get the proper tx port */ -static struct vnet_port *vnet_tx_port_find(struct sk_buff *skb, - struct net_device *dev) -{ - struct vnet *vp = netdev_priv(dev); - - return __tx_port_find(vp, skb); -} - -static u16 vnet_select_queue(struct net_device *dev, struct sk_buff *skb, - struct net_device *sb_dev) -{ - struct vnet *vp = netdev_priv(dev); - struct vnet_port *port = __tx_port_find(vp, skb); - - if (!port) - return 0; - - return port->q_index; -} - -/* Wrappers to common functions */ -static netdev_tx_t vnet_start_xmit(struct sk_buff *skb, struct net_device *dev) -{ - return sunvnet_start_xmit_common(skb, dev, vnet_tx_port_find); -} - -static void vnet_set_rx_mode(struct net_device *dev) -{ - struct vnet *vp = netdev_priv(dev); - - return sunvnet_set_rx_mode_common(dev, vp); -} - -#ifdef CONFIG_NET_POLL_CONTROLLER -static void vnet_poll_controller(struct net_device *dev) -{ - struct vnet *vp = netdev_priv(dev); - - return sunvnet_poll_controller_common(dev, vp); -} -#endif - -static const struct net_device_ops vnet_ops = { - .ndo_open = sunvnet_open_common, - .ndo_stop = sunvnet_close_common, - .ndo_set_rx_mode = vnet_set_rx_mode, - .ndo_set_mac_address = sunvnet_set_mac_addr_common, - .ndo_validate_addr = eth_validate_addr, - .ndo_tx_timeout = sunvnet_tx_timeout_common, - .ndo_start_xmit = vnet_start_xmit, - .ndo_select_queue = vnet_select_queue, -#ifdef CONFIG_NET_POLL_CONTROLLER - .ndo_poll_controller = vnet_poll_controller, -#endif -}; - -static struct vnet *vnet_new(const u64 *local_mac, - struct vio_dev *vdev) -{ - struct net_device *dev; - u8 addr[ETH_ALEN]; - struct vnet *vp; - int err, i; - - dev = alloc_etherdev_mqs(sizeof(*vp), VNET_MAX_TXQS, 1); - if (!dev) - return ERR_PTR(-ENOMEM); - dev->needed_headroom = VNET_PACKET_SKIP + 8; - dev->needed_tailroom = 8; - - for (i = 0; i < ETH_ALEN; i++) - addr[i] = (*local_mac >> (5 - i) * 8) & 0xff; - eth_hw_addr_set(dev, addr); - - vp = netdev_priv(dev); - - spin_lock_init(&vp->lock); - vp->dev = dev; - - INIT_LIST_HEAD(&vp->port_list); - for (i = 0; i < VNET_PORT_HASH_SIZE; i++) - INIT_HLIST_HEAD(&vp->port_hash[i]); - INIT_LIST_HEAD(&vp->list); - vp->local_mac = *local_mac; - - dev->netdev_ops = &vnet_ops; - dev->ethtool_ops = &vnet_ethtool_ops; - dev->watchdog_timeo = VNET_TX_TIMEOUT; - - dev->hw_features = NETIF_F_TSO | NETIF_F_GSO | NETIF_F_ALL_TSO | - NETIF_F_HW_CSUM | NETIF_F_SG; - dev->features = dev->hw_features; - - /* MTU range: 68 - 65535 */ - dev->min_mtu = ETH_MIN_MTU; - dev->max_mtu = VNET_MAX_MTU; - - SET_NETDEV_DEV(dev, &vdev->dev); - - err = register_netdev(dev); - if (err) { - pr_err("Cannot register net device, aborting\n"); - goto err_out_free_dev; - } - - netdev_info(dev, "Sun LDOM vnet %pM\n", dev->dev_addr); - - list_add(&vp->list, &vnet_list); - - return vp; - -err_out_free_dev: - free_netdev(dev); - - return ERR_PTR(err); -} - -static struct vnet *vnet_find_or_create(const u64 *local_mac, - struct vio_dev *vdev) -{ - struct vnet *iter, *vp; - - mutex_lock(&vnet_list_mutex); - vp = NULL; - list_for_each_entry(iter, &vnet_list, list) { - if (iter->local_mac == *local_mac) { - vp = iter; - break; - } - } - if (!vp) - vp = vnet_new(local_mac, vdev); - mutex_unlock(&vnet_list_mutex); - - return vp; -} - -static void vnet_cleanup(void) -{ - struct vnet *vp; - struct net_device *dev; - - mutex_lock(&vnet_list_mutex); - while (!list_empty(&vnet_list)) { - vp = list_first_entry(&vnet_list, struct vnet, list); - list_del(&vp->list); - dev = vp->dev; - /* vio_unregister_driver() should have cleaned up port_list */ - BUG_ON(!list_empty(&vp->port_list)); - unregister_netdev(dev); - free_netdev(dev); - } - mutex_unlock(&vnet_list_mutex); -} - -static const char *local_mac_prop = "local-mac-address"; - -static struct vnet *vnet_find_parent(struct mdesc_handle *hp, - u64 port_node, - struct vio_dev *vdev) -{ - const u64 *local_mac = NULL; - u64 a; - - mdesc_for_each_arc(a, hp, port_node, MDESC_ARC_TYPE_BACK) { - u64 target = mdesc_arc_target(hp, a); - const char *name; - - name = mdesc_get_property(hp, target, "name", NULL); - if (!name || strcmp(name, "network")) - continue; - - local_mac = mdesc_get_property(hp, target, - local_mac_prop, NULL); - if (local_mac) - break; - } - if (!local_mac) - return ERR_PTR(-ENODEV); - - return vnet_find_or_create(local_mac, vdev); -} - -static struct ldc_channel_config vnet_ldc_cfg = { - .event = sunvnet_event_common, - .mtu = 64, - .mode = LDC_MODE_UNRELIABLE, -}; - -static struct vio_driver_ops vnet_vio_ops = { - .send_attr = sunvnet_send_attr_common, - .handle_attr = sunvnet_handle_attr_common, - .handshake_complete = sunvnet_handshake_complete_common, -}; - -const char *remote_macaddr_prop = "remote-mac-address"; - -static int vnet_port_probe(struct vio_dev *vdev, const struct vio_device_id *id) -{ - struct mdesc_handle *hp; - struct vnet_port *port; - unsigned long flags; - struct vnet *vp; - const u64 *rmac; - int len, i, err, switch_port; - - hp = mdesc_grab(); - - vp = vnet_find_parent(hp, vdev->mp, vdev); - if (IS_ERR(vp)) { - pr_err("Cannot find port parent vnet\n"); - err = PTR_ERR(vp); - goto err_out_put_mdesc; - } - - rmac = mdesc_get_property(hp, vdev->mp, remote_macaddr_prop, &len); - err = -ENODEV; - if (!rmac) { - pr_err("Port lacks %s property\n", remote_macaddr_prop); - goto err_out_put_mdesc; - } - - port = kzalloc(sizeof(*port), GFP_KERNEL); - err = -ENOMEM; - if (!port) - goto err_out_put_mdesc; - - for (i = 0; i < ETH_ALEN; i++) - port->raddr[i] = (*rmac >> (5 - i) * 8) & 0xff; - - port->vp = vp; - - err = vio_driver_init(&port->vio, vdev, VDEV_NETWORK, - vnet_versions, ARRAY_SIZE(vnet_versions), - &vnet_vio_ops, vp->dev->name); - if (err) - goto err_out_free_port; - - err = vio_ldc_alloc(&port->vio, &vnet_ldc_cfg, port); - if (err) - goto err_out_free_port; - - netif_napi_add(port->vp->dev, &port->napi, sunvnet_poll_common); - - INIT_HLIST_NODE(&port->hash); - INIT_LIST_HEAD(&port->list); - - switch_port = 0; - if (mdesc_get_property(hp, vdev->mp, "switch-port", NULL)) - switch_port = 1; - port->switch_port = switch_port; - port->tso = true; - port->tsolen = 0; - - spin_lock_irqsave(&vp->lock, flags); - if (switch_port) - list_add_rcu(&port->list, &vp->port_list); - else - list_add_tail_rcu(&port->list, &vp->port_list); - hlist_add_head_rcu(&port->hash, - &vp->port_hash[vnet_hashfn(port->raddr)]); - sunvnet_port_add_txq_common(port); - spin_unlock_irqrestore(&vp->lock, flags); - - dev_set_drvdata(&vdev->dev, port); - - pr_info("%s: PORT ( remote-mac %pM%s )\n", - vp->dev->name, port->raddr, switch_port ? " switch-port" : ""); - - timer_setup(&port->clean_timer, sunvnet_clean_timer_expire_common, 0); - - napi_enable(&port->napi); - vio_port_up(&port->vio); - - mdesc_release(hp); - - return 0; - -err_out_free_port: - kfree(port); - -err_out_put_mdesc: - mdesc_release(hp); - return err; -} - -static void vnet_port_remove(struct vio_dev *vdev) -{ - struct vnet_port *port = dev_get_drvdata(&vdev->dev); - - if (port) { - del_timer_sync(&port->vio.timer); - - napi_disable(&port->napi); - - list_del_rcu(&port->list); - hlist_del_rcu(&port->hash); - - synchronize_rcu(); - timer_shutdown_sync(&port->clean_timer); - sunvnet_port_rm_txq_common(port); - netif_napi_del(&port->napi); - sunvnet_port_free_tx_bufs_common(port); - vio_ldc_free(&port->vio); - - dev_set_drvdata(&vdev->dev, NULL); - - kfree(port); - } -} - -static const struct vio_device_id vnet_port_match[] = { - { - .type = "vnet-port", - }, - {}, -}; -MODULE_DEVICE_TABLE(vio, vnet_port_match); - -static struct vio_driver vnet_port_driver = { - .id_table = vnet_port_match, - .probe = vnet_port_probe, - .remove = vnet_port_remove, - .name = "vnet_port", -}; - -static int __init vnet_init(void) -{ - pr_info("%s\n", version); - return vio_register_driver(&vnet_port_driver); -} - -static void __exit vnet_exit(void) -{ - vio_unregister_driver(&vnet_port_driver); - vnet_cleanup(); -} - -module_init(vnet_init); -module_exit(vnet_exit); diff --git a/drivers/net/ethernet/sun/sunvnet_common.c b/drivers/net/ethernet/sun/sunvnet_common.c deleted file mode 100644 index a6211b9..00000000 --- a/drivers/net/ethernet/sun/sunvnet_common.c +++ /dev/null @@ -1,1813 +0,0 @@ -// SPDX-License-Identifier: GPL-2.0 -/* sunvnet.c: Sun LDOM Virtual Network Driver. - * - * Copyright (C) 2007, 2008 David S. Miller - * Copyright (C) 2016-2017 Oracle. All rights reserved. - */ - -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#define CREATE_TRACE_POINTS -#include - -#if IS_ENABLED(CONFIG_IPV6) -#include -#endif - -#include -#include -#include - -#include -#include - -#include "sunvnet_common.h" - -/* Heuristic for the number of times to exponentially backoff and - * retry sending an LDC trigger when EAGAIN is encountered - */ -#define VNET_MAX_RETRIES 10 - -MODULE_AUTHOR("David S. Miller (davem@davemloft.net)"); -MODULE_DESCRIPTION("Sun LDOM virtual network support library"); -MODULE_LICENSE("GPL"); -MODULE_VERSION("1.1"); - -static int __vnet_tx_trigger(struct vnet_port *port, u32 start); - -static inline u32 vnet_tx_dring_avail(struct vio_dring_state *dr) -{ - return vio_dring_avail(dr, VNET_TX_RING_SIZE); -} - -static int vnet_handle_unknown(struct vnet_port *port, void *arg) -{ - struct vio_msg_tag *pkt = arg; - - pr_err("Received unknown msg [%02x:%02x:%04x:%08x]\n", - pkt->type, pkt->stype, pkt->stype_env, pkt->sid); - pr_err("Resetting connection\n"); - - ldc_disconnect(port->vio.lp); - - return -ECONNRESET; -} - -static int vnet_port_alloc_tx_ring(struct vnet_port *port); - -int sunvnet_send_attr_common(struct vio_driver_state *vio) -{ - struct vnet_port *port = to_vnet_port(vio); - struct net_device *dev = VNET_PORT_TO_NET_DEVICE(port); - struct vio_net_attr_info pkt; - int framelen = ETH_FRAME_LEN; - int i, err; - - err = vnet_port_alloc_tx_ring(to_vnet_port(vio)); - if (err) - return err; - - memset(&pkt, 0, sizeof(pkt)); - pkt.tag.type = VIO_TYPE_CTRL; - pkt.tag.stype = VIO_SUBTYPE_INFO; - pkt.tag.stype_env = VIO_ATTR_INFO; - pkt.tag.sid = vio_send_sid(vio); - if (vio_version_before(vio, 1, 2)) - pkt.xfer_mode = VIO_DRING_MODE; - else - pkt.xfer_mode = VIO_NEW_DRING_MODE; - pkt.addr_type = VNET_ADDR_ETHERMAC; - pkt.ack_freq = 0; - for (i = 0; i < 6; i++) - pkt.addr |= (u64)dev->dev_addr[i] << ((5 - i) * 8); - if (vio_version_after(vio, 1, 3)) { - if (port->rmtu) { - port->rmtu = min(VNET_MAXPACKET, port->rmtu); - pkt.mtu = port->rmtu; - } else { - port->rmtu = VNET_MAXPACKET; - pkt.mtu = port->rmtu; - } - if (vio_version_after_eq(vio, 1, 6)) - pkt.options = VIO_TX_DRING; - } else if (vio_version_before(vio, 1, 3)) { - pkt.mtu = framelen; - } else { /* v1.3 */ - pkt.mtu = framelen + VLAN_HLEN; - } - - pkt.cflags = 0; - if (vio_version_after_eq(vio, 1, 7) && port->tso) { - pkt.cflags |= VNET_LSO_IPV4_CAPAB; - if (!port->tsolen) - port->tsolen = VNET_MAXTSO; - pkt.ipv4_lso_maxlen = port->tsolen; - } - - pkt.plnk_updt = PHYSLINK_UPDATE_NONE; - - viodbg(HS, "SEND NET ATTR xmode[0x%x] atype[0x%x] addr[%llx] " - "ackfreq[%u] plnk_updt[0x%02x] opts[0x%02x] mtu[%llu] " - "cflags[0x%04x] lso_max[%u]\n", - pkt.xfer_mode, pkt.addr_type, - (unsigned long long)pkt.addr, - pkt.ack_freq, pkt.plnk_updt, pkt.options, - (unsigned long long)pkt.mtu, pkt.cflags, pkt.ipv4_lso_maxlen); - - return vio_ldc_send(vio, &pkt, sizeof(pkt)); -} -EXPORT_SYMBOL_GPL(sunvnet_send_attr_common); - -static int handle_attr_info(struct vio_driver_state *vio, - struct vio_net_attr_info *pkt) -{ - struct vnet_port *port = to_vnet_port(vio); - u64 localmtu; - u8 xfer_mode; - - viodbg(HS, "GOT NET ATTR xmode[0x%x] atype[0x%x] addr[%llx] " - "ackfreq[%u] plnk_updt[0x%02x] opts[0x%02x] mtu[%llu] " - " (rmtu[%llu]) cflags[0x%04x] lso_max[%u]\n", - pkt->xfer_mode, pkt->addr_type, - (unsigned long long)pkt->addr, - pkt->ack_freq, pkt->plnk_updt, pkt->options, - (unsigned long long)pkt->mtu, port->rmtu, pkt->cflags, - pkt->ipv4_lso_maxlen); - - pkt->tag.sid = vio_send_sid(vio); - - xfer_mode = pkt->xfer_mode; - /* for version < 1.2, VIO_DRING_MODE = 0x3 and no bitmask */ - if (vio_version_before(vio, 1, 2) && xfer_mode == VIO_DRING_MODE) - xfer_mode = VIO_NEW_DRING_MODE; - - /* MTU negotiation: - * < v1.3 - ETH_FRAME_LEN exactly - * > v1.3 - MIN(pkt.mtu, VNET_MAXPACKET, port->rmtu) and change - * pkt->mtu for ACK - * = v1.3 - ETH_FRAME_LEN + VLAN_HLEN exactly - */ - if (vio_version_before(vio, 1, 3)) { - localmtu = ETH_FRAME_LEN; - } else if (vio_version_after(vio, 1, 3)) { - localmtu = port->rmtu ? port->rmtu : VNET_MAXPACKET; - localmtu = min(pkt->mtu, localmtu); - pkt->mtu = localmtu; - } else { /* v1.3 */ - localmtu = ETH_FRAME_LEN + VLAN_HLEN; - } - port->rmtu = localmtu; - - /* LSO negotiation */ - if (vio_version_after_eq(vio, 1, 7)) - port->tso &= !!(pkt->cflags & VNET_LSO_IPV4_CAPAB); - else - port->tso = false; - if (port->tso) { - if (!port->tsolen) - port->tsolen = VNET_MAXTSO; - port->tsolen = min(port->tsolen, pkt->ipv4_lso_maxlen); - if (port->tsolen < VNET_MINTSO) { - port->tso = false; - port->tsolen = 0; - pkt->cflags &= ~VNET_LSO_IPV4_CAPAB; - } - pkt->ipv4_lso_maxlen = port->tsolen; - } else { - pkt->cflags &= ~VNET_LSO_IPV4_CAPAB; - pkt->ipv4_lso_maxlen = 0; - port->tsolen = 0; - } - - /* for version >= 1.6, ACK packet mode we support */ - if (vio_version_after_eq(vio, 1, 6)) { - pkt->xfer_mode = VIO_NEW_DRING_MODE; - pkt->options = VIO_TX_DRING; - } - - if (!(xfer_mode | VIO_NEW_DRING_MODE) || - pkt->addr_type != VNET_ADDR_ETHERMAC || - pkt->mtu != localmtu) { - viodbg(HS, "SEND NET ATTR NACK\n"); - - pkt->tag.stype = VIO_SUBTYPE_NACK; - - (void)vio_ldc_send(vio, pkt, sizeof(*pkt)); - - return -ECONNRESET; - } - - viodbg(HS, "SEND NET ATTR ACK xmode[0x%x] atype[0x%x] " - "addr[%llx] ackfreq[%u] plnk_updt[0x%02x] opts[0x%02x] " - "mtu[%llu] (rmtu[%llu]) cflags[0x%04x] lso_max[%u]\n", - pkt->xfer_mode, pkt->addr_type, - (unsigned long long)pkt->addr, - pkt->ack_freq, pkt->plnk_updt, pkt->options, - (unsigned long long)pkt->mtu, port->rmtu, pkt->cflags, - pkt->ipv4_lso_maxlen); - - pkt->tag.stype = VIO_SUBTYPE_ACK; - - return vio_ldc_send(vio, pkt, sizeof(*pkt)); -} - -static int handle_attr_ack(struct vio_driver_state *vio, - struct vio_net_attr_info *pkt) -{ - viodbg(HS, "GOT NET ATTR ACK\n"); - - return 0; -} - -static int handle_attr_nack(struct vio_driver_state *vio, - struct vio_net_attr_info *pkt) -{ - viodbg(HS, "GOT NET ATTR NACK\n"); - - return -ECONNRESET; -} - -int sunvnet_handle_attr_common(struct vio_driver_state *vio, void *arg) -{ - struct vio_net_attr_info *pkt = arg; - - switch (pkt->tag.stype) { - case VIO_SUBTYPE_INFO: - return handle_attr_info(vio, pkt); - - case VIO_SUBTYPE_ACK: - return handle_attr_ack(vio, pkt); - - case VIO_SUBTYPE_NACK: - return handle_attr_nack(vio, pkt); - - default: - return -ECONNRESET; - } -} -EXPORT_SYMBOL_GPL(sunvnet_handle_attr_common); - -void sunvnet_handshake_complete_common(struct vio_driver_state *vio) -{ - struct vio_dring_state *dr; - - dr = &vio->drings[VIO_DRIVER_RX_RING]; - dr->rcv_nxt = 1; - dr->snd_nxt = 1; - - dr = &vio->drings[VIO_DRIVER_TX_RING]; - dr->rcv_nxt = 1; - dr->snd_nxt = 1; -} -EXPORT_SYMBOL_GPL(sunvnet_handshake_complete_common); - -/* The hypervisor interface that implements copying to/from imported - * memory from another domain requires that copies are done to 8-byte - * aligned buffers, and that the lengths of such copies are also 8-byte - * multiples. - * - * So we align skb->data to an 8-byte multiple and pad-out the data - * area so we can round the copy length up to the next multiple of - * 8 for the copy. - * - * The transmitter puts the actual start of the packet 6 bytes into - * the buffer it sends over, so that the IP headers after the ethernet - * header are aligned properly. These 6 bytes are not in the descriptor - * length, they are simply implied. This offset is represented using - * the VNET_PACKET_SKIP macro. - */ -static struct sk_buff *alloc_and_align_skb(struct net_device *dev, - unsigned int len) -{ - struct sk_buff *skb; - unsigned long addr, off; - - skb = netdev_alloc_skb(dev, len + VNET_PACKET_SKIP + 8 + 8); - if (unlikely(!skb)) - return NULL; - - addr = (unsigned long)skb->data; - off = ((addr + 7UL) & ~7UL) - addr; - if (off) - skb_reserve(skb, off); - - return skb; -} - -static inline void vnet_fullcsum_ipv4(struct sk_buff *skb) -{ - struct iphdr *iph = ip_hdr(skb); - int offset = skb_transport_offset(skb); - - if (skb->protocol != htons(ETH_P_IP)) - return; - if (iph->protocol != IPPROTO_TCP && - iph->protocol != IPPROTO_UDP) - return; - skb->ip_summed = CHECKSUM_NONE; - skb->csum_level = 1; - skb->csum = 0; - if (iph->protocol == IPPROTO_TCP) { - struct tcphdr *ptcp = tcp_hdr(skb); - - ptcp->check = 0; - skb->csum = skb_checksum(skb, offset, skb->len - offset, 0); - ptcp->check = csum_tcpudp_magic(iph->saddr, iph->daddr, - skb->len - offset, IPPROTO_TCP, - skb->csum); - } else if (iph->protocol == IPPROTO_UDP) { - struct udphdr *pudp = udp_hdr(skb); - - pudp->check = 0; - skb->csum = skb_checksum(skb, offset, skb->len - offset, 0); - pudp->check = csum_tcpudp_magic(iph->saddr, iph->daddr, - skb->len - offset, IPPROTO_UDP, - skb->csum); - } -} - -#if IS_ENABLED(CONFIG_IPV6) -static inline void vnet_fullcsum_ipv6(struct sk_buff *skb) -{ - struct ipv6hdr *ip6h = ipv6_hdr(skb); - int offset = skb_transport_offset(skb); - - if (skb->protocol != htons(ETH_P_IPV6)) - return; - if (ip6h->nexthdr != IPPROTO_TCP && - ip6h->nexthdr != IPPROTO_UDP) - return; - skb->ip_summed = CHECKSUM_NONE; - skb->csum_level = 1; - skb->csum = 0; - if (ip6h->nexthdr == IPPROTO_TCP) { - struct tcphdr *ptcp = tcp_hdr(skb); - - ptcp->check = 0; - skb->csum = skb_checksum(skb, offset, skb->len - offset, 0); - ptcp->check = csum_ipv6_magic(&ip6h->saddr, &ip6h->daddr, - skb->len - offset, IPPROTO_TCP, - skb->csum); - } else if (ip6h->nexthdr == IPPROTO_UDP) { - struct udphdr *pudp = udp_hdr(skb); - - pudp->check = 0; - skb->csum = skb_checksum(skb, offset, skb->len - offset, 0); - pudp->check = csum_ipv6_magic(&ip6h->saddr, &ip6h->daddr, - skb->len - offset, IPPROTO_UDP, - skb->csum); - } -} -#endif - -static int vnet_rx_one(struct vnet_port *port, struct vio_net_desc *desc) -{ - struct net_device *dev = VNET_PORT_TO_NET_DEVICE(port); - unsigned int len = desc->size; - unsigned int copy_len; - struct sk_buff *skb; - int maxlen; - int err; - - err = -EMSGSIZE; - if (port->tso && port->tsolen > port->rmtu) - maxlen = port->tsolen; - else - maxlen = port->rmtu; - if (unlikely(len < ETH_ZLEN || len > maxlen)) { - dev->stats.rx_length_errors++; - goto out_dropped; - } - - skb = alloc_and_align_skb(dev, len); - err = -ENOMEM; - if (unlikely(!skb)) { - dev->stats.rx_missed_errors++; - goto out_dropped; - } - - copy_len = (len + VNET_PACKET_SKIP + 7U) & ~7U; - skb_put(skb, copy_len); - err = ldc_copy(port->vio.lp, LDC_COPY_IN, - skb->data, copy_len, 0, - desc->cookies, desc->ncookies); - if (unlikely(err < 0)) { - dev->stats.rx_frame_errors++; - goto out_free_skb; - } - - skb_pull(skb, VNET_PACKET_SKIP); - skb_trim(skb, len); - skb->protocol = eth_type_trans(skb, dev); - - if (vio_version_after_eq(&port->vio, 1, 8)) { - struct vio_net_dext *dext = vio_net_ext(desc); - - skb_reset_network_header(skb); - - if (dext->flags & VNET_PKT_HCK_IPV4_HDRCKSUM) { - if (skb->protocol == ETH_P_IP) { - struct iphdr *iph = ip_hdr(skb); - - iph->check = 0; - ip_send_check(iph); - } - } - if ((dext->flags & VNET_PKT_HCK_FULLCKSUM) && - skb->ip_summed == CHECKSUM_NONE) { - if (skb->protocol == htons(ETH_P_IP)) { - struct iphdr *iph = ip_hdr(skb); - int ihl = iph->ihl * 4; - - skb_set_transport_header(skb, ihl); - vnet_fullcsum_ipv4(skb); -#if IS_ENABLED(CONFIG_IPV6) - } else if (skb->protocol == htons(ETH_P_IPV6)) { - skb_set_transport_header(skb, - sizeof(struct ipv6hdr)); - vnet_fullcsum_ipv6(skb); -#endif - } - } - if (dext->flags & VNET_PKT_HCK_IPV4_HDRCKSUM_OK) { - skb->ip_summed = CHECKSUM_PARTIAL; - skb->csum_level = 0; - if (dext->flags & VNET_PKT_HCK_FULLCKSUM_OK) - skb->csum_level = 1; - } - } - - skb->ip_summed = port->switch_port ? CHECKSUM_NONE : CHECKSUM_PARTIAL; - - if (unlikely(is_multicast_ether_addr(eth_hdr(skb)->h_dest))) - dev->stats.multicast++; - dev->stats.rx_packets++; - dev->stats.rx_bytes += len; - port->stats.rx_packets++; - port->stats.rx_bytes += len; - napi_gro_receive(&port->napi, skb); - return 0; - -out_free_skb: - kfree_skb(skb); - -out_dropped: - dev->stats.rx_dropped++; - return err; -} - -static int vnet_send_ack(struct vnet_port *port, struct vio_dring_state *dr, - u32 start, u32 end, u8 vio_dring_state) -{ - struct vio_dring_data hdr = { - .tag = { - .type = VIO_TYPE_DATA, - .stype = VIO_SUBTYPE_ACK, - .stype_env = VIO_DRING_DATA, - .sid = vio_send_sid(&port->vio), - }, - .dring_ident = dr->ident, - .start_idx = start, - .end_idx = end, - .state = vio_dring_state, - }; - int err, delay; - int retries = 0; - - hdr.seq = dr->snd_nxt; - delay = 1; - do { - err = vio_ldc_send(&port->vio, &hdr, sizeof(hdr)); - if (err > 0) { - dr->snd_nxt++; - break; - } - udelay(delay); - if ((delay <<= 1) > 128) - delay = 128; - if (retries++ > VNET_MAX_RETRIES) { - pr_info("ECONNRESET %x:%x:%x:%x:%x:%x\n", - port->raddr[0], port->raddr[1], - port->raddr[2], port->raddr[3], - port->raddr[4], port->raddr[5]); - break; - } - } while (err == -EAGAIN); - - if (err <= 0 && vio_dring_state == VIO_DRING_STOPPED) { - port->stop_rx_idx = end; - port->stop_rx = true; - } else { - port->stop_rx_idx = 0; - port->stop_rx = false; - } - - return err; -} - -static struct vio_net_desc *get_rx_desc(struct vnet_port *port, - struct vio_dring_state *dr, - u32 index) -{ - struct vio_net_desc *desc = port->vio.desc_buf; - int err; - - err = ldc_get_dring_entry(port->vio.lp, desc, dr->entry_size, - (index * dr->entry_size), - dr->cookies, dr->ncookies); - if (err < 0) - return ERR_PTR(err); - - return desc; -} - -static int put_rx_desc(struct vnet_port *port, - struct vio_dring_state *dr, - struct vio_net_desc *desc, - u32 index) -{ - int err; - - err = ldc_put_dring_entry(port->vio.lp, desc, dr->entry_size, - (index * dr->entry_size), - dr->cookies, dr->ncookies); - if (err < 0) - return err; - - return 0; -} - -static int vnet_walk_rx_one(struct vnet_port *port, - struct vio_dring_state *dr, - u32 index, int *needs_ack) -{ - struct vio_net_desc *desc = get_rx_desc(port, dr, index); - struct vio_driver_state *vio = &port->vio; - int err; - - BUG_ON(!desc); - if (IS_ERR(desc)) - return PTR_ERR(desc); - - if (desc->hdr.state != VIO_DESC_READY) - return 1; - - dma_rmb(); - - viodbg(DATA, "vio_walk_rx_one desc[%02x:%02x:%08x:%08x:%llx:%llx]\n", - desc->hdr.state, desc->hdr.ack, - desc->size, desc->ncookies, - desc->cookies[0].cookie_addr, - desc->cookies[0].cookie_size); - - err = vnet_rx_one(port, desc); - if (err == -ECONNRESET) - return err; - trace_vnet_rx_one(port->vio._local_sid, port->vio._peer_sid, - index, desc->hdr.ack); - desc->hdr.state = VIO_DESC_DONE; - err = put_rx_desc(port, dr, desc, index); - if (err < 0) - return err; - *needs_ack = desc->hdr.ack; - return 0; -} - -static int vnet_walk_rx(struct vnet_port *port, struct vio_dring_state *dr, - u32 start, u32 end, int *npkts, int budget) -{ - struct vio_driver_state *vio = &port->vio; - int ack_start = -1, ack_end = -1; - bool send_ack = true; - - end = (end == (u32)-1) ? vio_dring_prev(dr, start) - : vio_dring_next(dr, end); - - viodbg(DATA, "vnet_walk_rx start[%08x] end[%08x]\n", start, end); - - while (start != end) { - int ack = 0, err = vnet_walk_rx_one(port, dr, start, &ack); - - if (err == -ECONNRESET) - return err; - if (err != 0) - break; - (*npkts)++; - if (ack_start == -1) - ack_start = start; - ack_end = start; - start = vio_dring_next(dr, start); - if (ack && start != end) { - err = vnet_send_ack(port, dr, ack_start, ack_end, - VIO_DRING_ACTIVE); - if (err == -ECONNRESET) - return err; - ack_start = -1; - } - if ((*npkts) >= budget) { - send_ack = false; - break; - } - } - if (unlikely(ack_start == -1)) { - ack_end = vio_dring_prev(dr, start); - ack_start = ack_end; - } - if (send_ack) { - port->napi_resume = false; - trace_vnet_tx_send_stopped_ack(port->vio._local_sid, - port->vio._peer_sid, - ack_end, *npkts); - return vnet_send_ack(port, dr, ack_start, ack_end, - VIO_DRING_STOPPED); - } else { - trace_vnet_tx_defer_stopped_ack(port->vio._local_sid, - port->vio._peer_sid, - ack_end, *npkts); - port->napi_resume = true; - port->napi_stop_idx = ack_end; - return 1; - } -} - -static int vnet_rx(struct vnet_port *port, void *msgbuf, int *npkts, - int budget) -{ - struct vio_dring_data *pkt = msgbuf; - struct vio_dring_state *dr = &port->vio.drings[VIO_DRIVER_RX_RING]; - struct vio_driver_state *vio = &port->vio; - - viodbg(DATA, "vnet_rx stype_env[%04x] seq[%016llx] rcv_nxt[%016llx]\n", - pkt->tag.stype_env, pkt->seq, dr->rcv_nxt); - - if (unlikely(pkt->tag.stype_env != VIO_DRING_DATA)) - return 0; - if (unlikely(pkt->seq != dr->rcv_nxt)) { - pr_err("RX out of sequence seq[0x%llx] rcv_nxt[0x%llx]\n", - pkt->seq, dr->rcv_nxt); - return 0; - } - - if (!port->napi_resume) - dr->rcv_nxt++; - - /* XXX Validate pkt->start_idx and pkt->end_idx XXX */ - - return vnet_walk_rx(port, dr, pkt->start_idx, pkt->end_idx, - npkts, budget); -} - -static int idx_is_pending(struct vio_dring_state *dr, u32 end) -{ - u32 idx = dr->cons; - int found = 0; - - while (idx != dr->prod) { - if (idx == end) { - found = 1; - break; - } - idx = vio_dring_next(dr, idx); - } - return found; -} - -static int vnet_ack(struct vnet_port *port, void *msgbuf) -{ - struct vio_dring_state *dr = &port->vio.drings[VIO_DRIVER_TX_RING]; - struct vio_dring_data *pkt = msgbuf; - struct net_device *dev; - u32 end; - struct vio_net_desc *desc; - struct netdev_queue *txq; - - if (unlikely(pkt->tag.stype_env != VIO_DRING_DATA)) - return 0; - - end = pkt->end_idx; - dev = VNET_PORT_TO_NET_DEVICE(port); - netif_tx_lock(dev); - if (unlikely(!idx_is_pending(dr, end))) { - netif_tx_unlock(dev); - return 0; - } - - /* sync for race conditions with vnet_start_xmit() and tell xmit it - * is time to send a trigger. - */ - trace_vnet_rx_stopped_ack(port->vio._local_sid, - port->vio._peer_sid, end); - dr->cons = vio_dring_next(dr, end); - desc = vio_dring_entry(dr, dr->cons); - if (desc->hdr.state == VIO_DESC_READY && !port->start_cons) { - /* vnet_start_xmit() just populated this dring but missed - * sending the "start" LDC message to the consumer. - * Send a "start" trigger on its behalf. - */ - if (__vnet_tx_trigger(port, dr->cons) > 0) - port->start_cons = false; - else - port->start_cons = true; - } else { - port->start_cons = true; - } - netif_tx_unlock(dev); - - txq = netdev_get_tx_queue(dev, port->q_index); - if (unlikely(netif_tx_queue_stopped(txq) && - vnet_tx_dring_avail(dr) >= VNET_TX_WAKEUP_THRESH(dr))) - return 1; - - return 0; -} - -static int vnet_nack(struct vnet_port *port, void *msgbuf) -{ - /* XXX just reset or similar XXX */ - return 0; -} - -static int handle_mcast(struct vnet_port *port, void *msgbuf) -{ - struct vio_net_mcast_info *pkt = msgbuf; - struct net_device *dev = VNET_PORT_TO_NET_DEVICE(port); - - if (pkt->tag.stype != VIO_SUBTYPE_ACK) - pr_err("%s: Got unexpected MCAST reply [%02x:%02x:%04x:%08x]\n", - dev->name, - pkt->tag.type, - pkt->tag.stype, - pkt->tag.stype_env, - pkt->tag.sid); - - return 0; -} - -/* If the queue is stopped, wake it up so that we'll - * send out another START message at the next TX. - */ -static void maybe_tx_wakeup(struct vnet_port *port) -{ - struct netdev_queue *txq; - - txq = netdev_get_tx_queue(VNET_PORT_TO_NET_DEVICE(port), - port->q_index); - __netif_tx_lock(txq, smp_processor_id()); - if (likely(netif_tx_queue_stopped(txq))) - netif_tx_wake_queue(txq); - __netif_tx_unlock(txq); -} - -bool sunvnet_port_is_up_common(struct vnet_port *vnet) -{ - struct vio_driver_state *vio = &vnet->vio; - - return !!(vio->hs_state & VIO_HS_COMPLETE); -} -EXPORT_SYMBOL_GPL(sunvnet_port_is_up_common); - -static int vnet_event_napi(struct vnet_port *port, int budget) -{ - struct net_device *dev = VNET_PORT_TO_NET_DEVICE(port); - struct vio_driver_state *vio = &port->vio; - int tx_wakeup, err; - int npkts = 0; - - /* we don't expect any other bits */ - BUG_ON(port->rx_event & ~(LDC_EVENT_DATA_READY | - LDC_EVENT_RESET | - LDC_EVENT_UP)); - - /* RESET takes precedent over any other event */ - if (port->rx_event & LDC_EVENT_RESET) { - /* a link went down */ - - if (port->vsw == 1) { - netif_tx_stop_all_queues(dev); - netif_carrier_off(dev); - } - - vio_link_state_change(vio, LDC_EVENT_RESET); - vnet_port_reset(port); - vio_port_up(vio); - - /* If the device is running but its tx queue was - * stopped (due to flow control), restart it. - * This is necessary since vnet_port_reset() - * clears the tx drings and thus we may never get - * back a VIO_TYPE_DATA ACK packet - which is - * the normal mechanism to restart the tx queue. - */ - if (netif_running(dev)) - maybe_tx_wakeup(port); - - port->rx_event = 0; - port->stats.event_reset++; - return 0; - } - - if (port->rx_event & LDC_EVENT_UP) { - /* a link came up */ - - if (port->vsw == 1) { - netif_carrier_on(port->dev); - netif_tx_start_all_queues(port->dev); - } - - vio_link_state_change(vio, LDC_EVENT_UP); - port->rx_event = 0; - port->stats.event_up++; - return 0; - } - - err = 0; - tx_wakeup = 0; - while (1) { - union { - struct vio_msg_tag tag; - u64 raw[8]; - } msgbuf; - - if (port->napi_resume) { - struct vio_dring_data *pkt = - (struct vio_dring_data *)&msgbuf; - struct vio_dring_state *dr = - &port->vio.drings[VIO_DRIVER_RX_RING]; - - pkt->tag.type = VIO_TYPE_DATA; - pkt->tag.stype = VIO_SUBTYPE_INFO; - pkt->tag.stype_env = VIO_DRING_DATA; - pkt->seq = dr->rcv_nxt; - pkt->start_idx = vio_dring_next(dr, - port->napi_stop_idx); - pkt->end_idx = -1; - } else { - err = ldc_read(vio->lp, &msgbuf, sizeof(msgbuf)); - if (unlikely(err < 0)) { - if (err == -ECONNRESET) - vio_conn_reset(vio); - break; - } - if (err == 0) - break; - viodbg(DATA, "TAG [%02x:%02x:%04x:%08x]\n", - msgbuf.tag.type, - msgbuf.tag.stype, - msgbuf.tag.stype_env, - msgbuf.tag.sid); - err = vio_validate_sid(vio, &msgbuf.tag); - if (err < 0) - break; - } - - if (likely(msgbuf.tag.type == VIO_TYPE_DATA)) { - if (msgbuf.tag.stype == VIO_SUBTYPE_INFO) { - if (!sunvnet_port_is_up_common(port)) { - /* failures like handshake_failure() - * may have cleaned up dring, but - * NAPI polling may bring us here. - */ - err = -ECONNRESET; - break; - } - err = vnet_rx(port, &msgbuf, &npkts, budget); - if (npkts >= budget) - break; - if (npkts == 0) - break; - } else if (msgbuf.tag.stype == VIO_SUBTYPE_ACK) { - err = vnet_ack(port, &msgbuf); - if (err > 0) - tx_wakeup |= err; - } else if (msgbuf.tag.stype == VIO_SUBTYPE_NACK) { - err = vnet_nack(port, &msgbuf); - } - } else if (msgbuf.tag.type == VIO_TYPE_CTRL) { - if (msgbuf.tag.stype_env == VNET_MCAST_INFO) - err = handle_mcast(port, &msgbuf); - else - err = vio_control_pkt_engine(vio, &msgbuf); - if (err) - break; - } else { - err = vnet_handle_unknown(port, &msgbuf); - } - if (err == -ECONNRESET) - break; - } - if (unlikely(tx_wakeup && err != -ECONNRESET)) - maybe_tx_wakeup(port); - return npkts; -} - -int sunvnet_poll_common(struct napi_struct *napi, int budget) -{ - struct vnet_port *port = container_of(napi, struct vnet_port, napi); - struct vio_driver_state *vio = &port->vio; - int processed = vnet_event_napi(port, budget); - - if (processed < budget) { - napi_complete_done(napi, processed); - port->rx_event &= ~LDC_EVENT_DATA_READY; - vio_set_intr(vio->vdev->rx_ino, HV_INTR_ENABLED); - } - return processed; -} -EXPORT_SYMBOL_GPL(sunvnet_poll_common); - -void sunvnet_event_common(void *arg, int event) -{ - struct vnet_port *port = arg; - struct vio_driver_state *vio = &port->vio; - - port->rx_event |= event; - vio_set_intr(vio->vdev->rx_ino, HV_INTR_DISABLED); - napi_schedule(&port->napi); -} -EXPORT_SYMBOL_GPL(sunvnet_event_common); - -static int __vnet_tx_trigger(struct vnet_port *port, u32 start) -{ - struct vio_dring_state *dr = &port->vio.drings[VIO_DRIVER_TX_RING]; - struct vio_dring_data hdr = { - .tag = { - .type = VIO_TYPE_DATA, - .stype = VIO_SUBTYPE_INFO, - .stype_env = VIO_DRING_DATA, - .sid = vio_send_sid(&port->vio), - }, - .dring_ident = dr->ident, - .start_idx = start, - .end_idx = (u32)-1, - }; - int err, delay; - int retries = 0; - - if (port->stop_rx) { - trace_vnet_tx_pending_stopped_ack(port->vio._local_sid, - port->vio._peer_sid, - port->stop_rx_idx, -1); - err = vnet_send_ack(port, - &port->vio.drings[VIO_DRIVER_RX_RING], - port->stop_rx_idx, -1, - VIO_DRING_STOPPED); - if (err <= 0) - return err; - } - - hdr.seq = dr->snd_nxt; - delay = 1; - do { - err = vio_ldc_send(&port->vio, &hdr, sizeof(hdr)); - if (err > 0) { - dr->snd_nxt++; - break; - } - udelay(delay); - if ((delay <<= 1) > 128) - delay = 128; - if (retries++ > VNET_MAX_RETRIES) - break; - } while (err == -EAGAIN); - trace_vnet_tx_trigger(port->vio._local_sid, - port->vio._peer_sid, start, err); - - return err; -} - -static struct sk_buff *vnet_clean_tx_ring(struct vnet_port *port, - unsigned *pending) -{ - struct vio_dring_state *dr = &port->vio.drings[VIO_DRIVER_TX_RING]; - struct sk_buff *skb = NULL; - int i, txi; - - *pending = 0; - - txi = dr->prod; - for (i = 0; i < VNET_TX_RING_SIZE; ++i) { - struct vio_net_desc *d; - - --txi; - if (txi < 0) - txi = VNET_TX_RING_SIZE - 1; - - d = vio_dring_entry(dr, txi); - - if (d->hdr.state == VIO_DESC_READY) { - (*pending)++; - continue; - } - if (port->tx_bufs[txi].skb) { - if (d->hdr.state != VIO_DESC_DONE) - pr_notice("invalid ring buffer state %d\n", - d->hdr.state); - BUG_ON(port->tx_bufs[txi].skb->next); - - port->tx_bufs[txi].skb->next = skb; - skb = port->tx_bufs[txi].skb; - port->tx_bufs[txi].skb = NULL; - - ldc_unmap(port->vio.lp, - port->tx_bufs[txi].cookies, - port->tx_bufs[txi].ncookies); - } else if (d->hdr.state == VIO_DESC_FREE) { - break; - } - d->hdr.state = VIO_DESC_FREE; - } - return skb; -} - -static inline void vnet_free_skbs(struct sk_buff *skb) -{ - struct sk_buff *next; - - while (skb) { - next = skb->next; - skb->next = NULL; - dev_kfree_skb(skb); - skb = next; - } -} - -void sunvnet_clean_timer_expire_common(struct timer_list *t) -{ - struct vnet_port *port = from_timer(port, t, clean_timer); - struct sk_buff *freeskbs; - unsigned pending; - - netif_tx_lock(VNET_PORT_TO_NET_DEVICE(port)); - freeskbs = vnet_clean_tx_ring(port, &pending); - netif_tx_unlock(VNET_PORT_TO_NET_DEVICE(port)); - - vnet_free_skbs(freeskbs); - - if (pending) - (void)mod_timer(&port->clean_timer, - jiffies + VNET_CLEAN_TIMEOUT); - else - del_timer(&port->clean_timer); -} -EXPORT_SYMBOL_GPL(sunvnet_clean_timer_expire_common); - -static inline int vnet_skb_map(struct ldc_channel *lp, struct sk_buff *skb, - struct ldc_trans_cookie *cookies, int ncookies, - unsigned int map_perm) -{ - int i, nc, err, blen; - - /* header */ - blen = skb_headlen(skb); - if (blen < ETH_ZLEN) - blen = ETH_ZLEN; - blen += VNET_PACKET_SKIP; - blen += 8 - (blen & 7); - - err = ldc_map_single(lp, skb->data - VNET_PACKET_SKIP, blen, cookies, - ncookies, map_perm); - if (err < 0) - return err; - nc = err; - - for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { - skb_frag_t *f = &skb_shinfo(skb)->frags[i]; - u8 *vaddr; - - if (nc < ncookies) { - vaddr = kmap_local_page(skb_frag_page(f)); - blen = skb_frag_size(f); - blen += 8 - (blen & 7); - err = ldc_map_single(lp, vaddr + skb_frag_off(f), - blen, cookies + nc, ncookies - nc, - map_perm); - kunmap_local(vaddr); - } else { - err = -EMSGSIZE; - } - - if (err < 0) { - ldc_unmap(lp, cookies, nc); - return err; - } - nc += err; - } - return nc; -} - -static inline struct sk_buff *vnet_skb_shape(struct sk_buff *skb, int ncookies) -{ - struct sk_buff *nskb; - int i, len, pad, docopy; - - len = skb->len; - pad = 0; - if (len < ETH_ZLEN) { - pad += ETH_ZLEN - skb->len; - len += pad; - } - len += VNET_PACKET_SKIP; - pad += 8 - (len & 7); - - /* make sure we have enough cookies and alignment in every frag */ - docopy = skb_shinfo(skb)->nr_frags >= ncookies; - for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { - skb_frag_t *f = &skb_shinfo(skb)->frags[i]; - - docopy |= skb_frag_off(f) & 7; - } - if (((unsigned long)skb->data & 7) != VNET_PACKET_SKIP || - skb_tailroom(skb) < pad || - skb_headroom(skb) < VNET_PACKET_SKIP || docopy) { - int start = 0, offset; - __wsum csum; - - len = skb->len > ETH_ZLEN ? skb->len : ETH_ZLEN; - nskb = alloc_and_align_skb(skb->dev, len); - if (!nskb) { - dev_kfree_skb(skb); - return NULL; - } - skb_reserve(nskb, VNET_PACKET_SKIP); - - nskb->protocol = skb->protocol; - offset = skb_mac_header(skb) - skb->data; - skb_set_mac_header(nskb, offset); - offset = skb_network_header(skb) - skb->data; - skb_set_network_header(nskb, offset); - offset = skb_transport_header(skb) - skb->data; - skb_set_transport_header(nskb, offset); - - offset = 0; - nskb->csum_offset = skb->csum_offset; - nskb->ip_summed = skb->ip_summed; - - if (skb->ip_summed == CHECKSUM_PARTIAL) - start = skb_checksum_start_offset(skb); - if (start) { - int offset = start + nskb->csum_offset; - - /* copy the headers, no csum here */ - if (skb_copy_bits(skb, 0, nskb->data, start)) { - dev_kfree_skb(nskb); - dev_kfree_skb(skb); - return NULL; - } - - /* copy the rest, with csum calculation */ - *(__sum16 *)(skb->data + offset) = 0; - csum = skb_copy_and_csum_bits(skb, start, - nskb->data + start, - skb->len - start); - - /* add in the header checksums */ - if (skb->protocol == htons(ETH_P_IP)) { - struct iphdr *iph = ip_hdr(nskb); - - if (iph->protocol == IPPROTO_TCP || - iph->protocol == IPPROTO_UDP) { - csum = csum_tcpudp_magic(iph->saddr, - iph->daddr, - skb->len - start, - iph->protocol, - csum); - } - } else if (skb->protocol == htons(ETH_P_IPV6)) { - struct ipv6hdr *ip6h = ipv6_hdr(nskb); - - if (ip6h->nexthdr == IPPROTO_TCP || - ip6h->nexthdr == IPPROTO_UDP) { - csum = csum_ipv6_magic(&ip6h->saddr, - &ip6h->daddr, - skb->len - start, - ip6h->nexthdr, - csum); - } - } - - /* save the final result */ - *(__sum16 *)(nskb->data + offset) = csum; - - nskb->ip_summed = CHECKSUM_NONE; - } else if (skb_copy_bits(skb, 0, nskb->data, skb->len)) { - dev_kfree_skb(nskb); - dev_kfree_skb(skb); - return NULL; - } - (void)skb_put(nskb, skb->len); - if (skb_is_gso(skb)) { - skb_shinfo(nskb)->gso_size = skb_shinfo(skb)->gso_size; - skb_shinfo(nskb)->gso_type = skb_shinfo(skb)->gso_type; - } - nskb->queue_mapping = skb->queue_mapping; - dev_kfree_skb(skb); - skb = nskb; - } - return skb; -} - -static netdev_tx_t -vnet_handle_offloads(struct vnet_port *port, struct sk_buff *skb, - struct vnet_port *(*vnet_tx_port) - (struct sk_buff *, struct net_device *)) -{ - struct net_device *dev = VNET_PORT_TO_NET_DEVICE(port); - struct vio_dring_state *dr = &port->vio.drings[VIO_DRIVER_TX_RING]; - struct sk_buff *segs, *curr, *next; - int maclen, datalen; - int status; - int gso_size, gso_type, gso_segs; - int hlen = skb_transport_header(skb) - skb_mac_header(skb); - int proto = IPPROTO_IP; - - if (skb->protocol == htons(ETH_P_IP)) - proto = ip_hdr(skb)->protocol; - else if (skb->protocol == htons(ETH_P_IPV6)) - proto = ipv6_hdr(skb)->nexthdr; - - if (proto == IPPROTO_TCP) { - hlen += tcp_hdr(skb)->doff * 4; - } else if (proto == IPPROTO_UDP) { - hlen += sizeof(struct udphdr); - } else { - pr_err("vnet_handle_offloads GSO with unknown transport " - "protocol %d tproto %d\n", skb->protocol, proto); - hlen = 128; /* XXX */ - } - datalen = port->tsolen - hlen; - - gso_size = skb_shinfo(skb)->gso_size; - gso_type = skb_shinfo(skb)->gso_type; - gso_segs = skb_shinfo(skb)->gso_segs; - - if (port->tso && gso_size < datalen) - gso_segs = DIV_ROUND_UP(skb->len - hlen, datalen); - - if (unlikely(vnet_tx_dring_avail(dr) < gso_segs)) { - struct netdev_queue *txq; - - txq = netdev_get_tx_queue(dev, port->q_index); - netif_tx_stop_queue(txq); - if (vnet_tx_dring_avail(dr) < skb_shinfo(skb)->gso_segs) - return NETDEV_TX_BUSY; - netif_tx_wake_queue(txq); - } - - maclen = skb_network_header(skb) - skb_mac_header(skb); - skb_pull(skb, maclen); - - if (port->tso && gso_size < datalen) { - if (skb_unclone(skb, GFP_ATOMIC)) - goto out_dropped; - - /* segment to TSO size */ - skb_shinfo(skb)->gso_size = datalen; - skb_shinfo(skb)->gso_segs = gso_segs; - } - segs = skb_gso_segment(skb, dev->features & ~NETIF_F_TSO); - if (IS_ERR(segs)) - goto out_dropped; - - skb_push(skb, maclen); - skb_reset_mac_header(skb); - - status = 0; - skb_list_walk_safe(segs, curr, next) { - skb_mark_not_on_list(curr); - if (port->tso && curr->len > dev->mtu) { - skb_shinfo(curr)->gso_size = gso_size; - skb_shinfo(curr)->gso_type = gso_type; - skb_shinfo(curr)->gso_segs = - DIV_ROUND_UP(curr->len - hlen, gso_size); - } else { - skb_shinfo(curr)->gso_size = 0; - } - - skb_push(curr, maclen); - skb_reset_mac_header(curr); - memcpy(skb_mac_header(curr), skb_mac_header(skb), - maclen); - curr->csum_start = skb_transport_header(curr) - curr->head; - if (ip_hdr(curr)->protocol == IPPROTO_TCP) - curr->csum_offset = offsetof(struct tcphdr, check); - else if (ip_hdr(curr)->protocol == IPPROTO_UDP) - curr->csum_offset = offsetof(struct udphdr, check); - - if (!(status & NETDEV_TX_MASK)) - status = sunvnet_start_xmit_common(curr, dev, - vnet_tx_port); - if (status & NETDEV_TX_MASK) - dev_kfree_skb_any(curr); - } - - if (!(status & NETDEV_TX_MASK)) - dev_kfree_skb_any(skb); - return status; -out_dropped: - dev->stats.tx_dropped++; - dev_kfree_skb_any(skb); - return NETDEV_TX_OK; -} - -netdev_tx_t -sunvnet_start_xmit_common(struct sk_buff *skb, struct net_device *dev, - struct vnet_port *(*vnet_tx_port) - (struct sk_buff *, struct net_device *)) -{ - struct vnet_port *port = NULL; - struct vio_dring_state *dr; - struct vio_net_desc *d; - unsigned int len; - struct sk_buff *freeskbs = NULL; - int i, err, txi; - unsigned pending = 0; - struct netdev_queue *txq; - - rcu_read_lock(); - port = vnet_tx_port(skb, dev); - if (unlikely(!port)) - goto out_dropped; - - if (skb_is_gso(skb) && skb->len > port->tsolen) { - err = vnet_handle_offloads(port, skb, vnet_tx_port); - rcu_read_unlock(); - return err; - } - - if (!skb_is_gso(skb) && skb->len > port->rmtu) { - unsigned long localmtu = port->rmtu - ETH_HLEN; - - if (vio_version_after_eq(&port->vio, 1, 3)) - localmtu -= VLAN_HLEN; - - if (skb->protocol == htons(ETH_P_IP)) - icmp_ndo_send(skb, ICMP_DEST_UNREACH, ICMP_FRAG_NEEDED, - htonl(localmtu)); -#if IS_ENABLED(CONFIG_IPV6) - else if (skb->protocol == htons(ETH_P_IPV6)) - icmpv6_ndo_send(skb, ICMPV6_PKT_TOOBIG, 0, localmtu); -#endif - goto out_dropped; - } - - skb = vnet_skb_shape(skb, 2); - - if (unlikely(!skb)) - goto out_dropped; - - if (skb->ip_summed == CHECKSUM_PARTIAL) { - if (skb->protocol == htons(ETH_P_IP)) - vnet_fullcsum_ipv4(skb); -#if IS_ENABLED(CONFIG_IPV6) - else if (skb->protocol == htons(ETH_P_IPV6)) - vnet_fullcsum_ipv6(skb); -#endif - } - - dr = &port->vio.drings[VIO_DRIVER_TX_RING]; - i = skb_get_queue_mapping(skb); - txq = netdev_get_tx_queue(dev, i); - if (unlikely(vnet_tx_dring_avail(dr) < 1)) { - if (!netif_tx_queue_stopped(txq)) { - netif_tx_stop_queue(txq); - - /* This is a hard error, log it. */ - netdev_err(dev, "BUG! Tx Ring full when queue awake!\n"); - dev->stats.tx_errors++; - } - rcu_read_unlock(); - return NETDEV_TX_BUSY; - } - - d = vio_dring_cur(dr); - - txi = dr->prod; - - freeskbs = vnet_clean_tx_ring(port, &pending); - - BUG_ON(port->tx_bufs[txi].skb); - - len = skb->len; - if (len < ETH_ZLEN) - len = ETH_ZLEN; - - err = vnet_skb_map(port->vio.lp, skb, port->tx_bufs[txi].cookies, 2, - (LDC_MAP_SHADOW | LDC_MAP_DIRECT | LDC_MAP_RW)); - if (err < 0) { - netdev_info(dev, "tx buffer map error %d\n", err); - goto out_dropped; - } - - port->tx_bufs[txi].skb = skb; - skb = NULL; - port->tx_bufs[txi].ncookies = err; - - /* We don't rely on the ACKs to free the skb in vnet_start_xmit(), - * thus it is safe to not set VIO_ACK_ENABLE for each transmission: - * the protocol itself does not require it as long as the peer - * sends a VIO_SUBTYPE_ACK for VIO_DRING_STOPPED. - * - * An ACK for every packet in the ring is expensive as the - * sending of LDC messages is slow and affects performance. - */ - d->hdr.ack = VIO_ACK_DISABLE; - d->size = len; - d->ncookies = port->tx_bufs[txi].ncookies; - for (i = 0; i < d->ncookies; i++) - d->cookies[i] = port->tx_bufs[txi].cookies[i]; - if (vio_version_after_eq(&port->vio, 1, 7)) { - struct vio_net_dext *dext = vio_net_ext(d); - - memset(dext, 0, sizeof(*dext)); - if (skb_is_gso(port->tx_bufs[txi].skb)) { - dext->ipv4_lso_mss = skb_shinfo(port->tx_bufs[txi].skb) - ->gso_size; - dext->flags |= VNET_PKT_IPV4_LSO; - } - if (vio_version_after_eq(&port->vio, 1, 8) && - !port->switch_port) { - dext->flags |= VNET_PKT_HCK_IPV4_HDRCKSUM_OK; - dext->flags |= VNET_PKT_HCK_FULLCKSUM_OK; - } - } - - /* This has to be a non-SMP write barrier because we are writing - * to memory which is shared with the peer LDOM. - */ - dma_wmb(); - - d->hdr.state = VIO_DESC_READY; - - /* Exactly one ldc "start" trigger (for dr->cons) needs to be sent - * to notify the consumer that some descriptors are READY. - * After that "start" trigger, no additional triggers are needed until - * a DRING_STOPPED is received from the consumer. The dr->cons field - * (set up by vnet_ack()) has the value of the next dring index - * that has not yet been ack-ed. We send a "start" trigger here - * if, and only if, start_cons is true (reset it afterward). Conversely, - * vnet_ack() should check if the dring corresponding to cons - * is marked READY, but start_cons was false. - * If so, vnet_ack() should send out the missed "start" trigger. - * - * Note that the dma_wmb() above makes sure the cookies et al. are - * not globally visible before the VIO_DESC_READY, and that the - * stores are ordered correctly by the compiler. The consumer will - * not proceed until the VIO_DESC_READY is visible assuring that - * the consumer does not observe anything related to descriptors - * out of order. The HV trap from the LDC start trigger is the - * producer to consumer announcement that work is available to the - * consumer - */ - if (!port->start_cons) { /* previous trigger suffices */ - trace_vnet_skip_tx_trigger(port->vio._local_sid, - port->vio._peer_sid, dr->cons); - goto ldc_start_done; - } - - err = __vnet_tx_trigger(port, dr->cons); - if (unlikely(err < 0)) { - netdev_info(dev, "TX trigger error %d\n", err); - d->hdr.state = VIO_DESC_FREE; - skb = port->tx_bufs[txi].skb; - port->tx_bufs[txi].skb = NULL; - dev->stats.tx_carrier_errors++; - goto out_dropped; - } - -ldc_start_done: - port->start_cons = false; - - dev->stats.tx_packets++; - dev->stats.tx_bytes += port->tx_bufs[txi].skb->len; - port->stats.tx_packets++; - port->stats.tx_bytes += port->tx_bufs[txi].skb->len; - - dr->prod = (dr->prod + 1) & (VNET_TX_RING_SIZE - 1); - if (unlikely(vnet_tx_dring_avail(dr) < 1)) { - netif_tx_stop_queue(txq); - smp_rmb(); - if (vnet_tx_dring_avail(dr) > VNET_TX_WAKEUP_THRESH(dr)) - netif_tx_wake_queue(txq); - } - - (void)mod_timer(&port->clean_timer, jiffies + VNET_CLEAN_TIMEOUT); - rcu_read_unlock(); - - vnet_free_skbs(freeskbs); - - return NETDEV_TX_OK; - -out_dropped: - if (pending) - (void)mod_timer(&port->clean_timer, - jiffies + VNET_CLEAN_TIMEOUT); - else if (port) - del_timer(&port->clean_timer); - rcu_read_unlock(); - dev_kfree_skb(skb); - vnet_free_skbs(freeskbs); - dev->stats.tx_dropped++; - return NETDEV_TX_OK; -} -EXPORT_SYMBOL_GPL(sunvnet_start_xmit_common); - -void sunvnet_tx_timeout_common(struct net_device *dev, unsigned int txqueue) -{ - /* XXX Implement me XXX */ -} -EXPORT_SYMBOL_GPL(sunvnet_tx_timeout_common); - -int sunvnet_open_common(struct net_device *dev) -{ - netif_carrier_on(dev); - netif_tx_start_all_queues(dev); - - return 0; -} -EXPORT_SYMBOL_GPL(sunvnet_open_common); - -int sunvnet_close_common(struct net_device *dev) -{ - netif_tx_stop_all_queues(dev); - netif_carrier_off(dev); - - return 0; -} -EXPORT_SYMBOL_GPL(sunvnet_close_common); - -static struct vnet_mcast_entry *__vnet_mc_find(struct vnet *vp, u8 *addr) -{ - struct vnet_mcast_entry *m; - - for (m = vp->mcast_list; m; m = m->next) { - if (ether_addr_equal(m->addr, addr)) - return m; - } - return NULL; -} - -static void __update_mc_list(struct vnet *vp, struct net_device *dev) -{ - struct netdev_hw_addr *ha; - - netdev_for_each_mc_addr(ha, dev) { - struct vnet_mcast_entry *m; - - m = __vnet_mc_find(vp, ha->addr); - if (m) { - m->hit = 1; - continue; - } - - if (!m) { - m = kzalloc(sizeof(*m), GFP_ATOMIC); - if (!m) - continue; - memcpy(m->addr, ha->addr, ETH_ALEN); - m->hit = 1; - - m->next = vp->mcast_list; - vp->mcast_list = m; - } - } -} - -static void __send_mc_list(struct vnet *vp, struct vnet_port *port) -{ - struct vio_net_mcast_info info; - struct vnet_mcast_entry *m, **pp; - int n_addrs; - - memset(&info, 0, sizeof(info)); - - info.tag.type = VIO_TYPE_CTRL; - info.tag.stype = VIO_SUBTYPE_INFO; - info.tag.stype_env = VNET_MCAST_INFO; - info.tag.sid = vio_send_sid(&port->vio); - info.set = 1; - - n_addrs = 0; - for (m = vp->mcast_list; m; m = m->next) { - if (m->sent) - continue; - m->sent = 1; - memcpy(&info.mcast_addr[n_addrs * ETH_ALEN], - m->addr, ETH_ALEN); - if (++n_addrs == VNET_NUM_MCAST) { - info.count = n_addrs; - - (void)vio_ldc_send(&port->vio, &info, - sizeof(info)); - n_addrs = 0; - } - } - if (n_addrs) { - info.count = n_addrs; - (void)vio_ldc_send(&port->vio, &info, sizeof(info)); - } - - info.set = 0; - - n_addrs = 0; - pp = &vp->mcast_list; - while ((m = *pp) != NULL) { - if (m->hit) { - m->hit = 0; - pp = &m->next; - continue; - } - - memcpy(&info.mcast_addr[n_addrs * ETH_ALEN], - m->addr, ETH_ALEN); - if (++n_addrs == VNET_NUM_MCAST) { - info.count = n_addrs; - (void)vio_ldc_send(&port->vio, &info, - sizeof(info)); - n_addrs = 0; - } - - *pp = m->next; - kfree(m); - } - if (n_addrs) { - info.count = n_addrs; - (void)vio_ldc_send(&port->vio, &info, sizeof(info)); - } -} - -void sunvnet_set_rx_mode_common(struct net_device *dev, struct vnet *vp) -{ - struct vnet_port *port; - - rcu_read_lock(); - list_for_each_entry_rcu(port, &vp->port_list, list) { - if (port->switch_port) { - __update_mc_list(vp, dev); - __send_mc_list(vp, port); - break; - } - } - rcu_read_unlock(); -} -EXPORT_SYMBOL_GPL(sunvnet_set_rx_mode_common); - -int sunvnet_set_mac_addr_common(struct net_device *dev, void *p) -{ - return -EINVAL; -} -EXPORT_SYMBOL_GPL(sunvnet_set_mac_addr_common); - -void sunvnet_port_free_tx_bufs_common(struct vnet_port *port) -{ - struct vio_dring_state *dr; - int i; - - dr = &port->vio.drings[VIO_DRIVER_TX_RING]; - - if (!dr->base) - return; - - for (i = 0; i < VNET_TX_RING_SIZE; i++) { - struct vio_net_desc *d; - void *skb = port->tx_bufs[i].skb; - - if (!skb) - continue; - - d = vio_dring_entry(dr, i); - - ldc_unmap(port->vio.lp, - port->tx_bufs[i].cookies, - port->tx_bufs[i].ncookies); - dev_kfree_skb(skb); - port->tx_bufs[i].skb = NULL; - d->hdr.state = VIO_DESC_FREE; - } - ldc_free_exp_dring(port->vio.lp, dr->base, - (dr->entry_size * dr->num_entries), - dr->cookies, dr->ncookies); - dr->base = NULL; - dr->entry_size = 0; - dr->num_entries = 0; - dr->pending = 0; - dr->ncookies = 0; -} -EXPORT_SYMBOL_GPL(sunvnet_port_free_tx_bufs_common); - -void vnet_port_reset(struct vnet_port *port) -{ - del_timer(&port->clean_timer); - sunvnet_port_free_tx_bufs_common(port); - port->rmtu = 0; - port->tso = (port->vsw == 0); /* no tso in vsw, misbehaves in bridge */ - port->tsolen = 0; -} -EXPORT_SYMBOL_GPL(vnet_port_reset); - -static int vnet_port_alloc_tx_ring(struct vnet_port *port) -{ - struct vio_dring_state *dr; - unsigned long len, elen; - int i, err, ncookies; - void *dring; - - dr = &port->vio.drings[VIO_DRIVER_TX_RING]; - - elen = sizeof(struct vio_net_desc) + - sizeof(struct ldc_trans_cookie) * 2; - if (vio_version_after_eq(&port->vio, 1, 7)) - elen += sizeof(struct vio_net_dext); - len = VNET_TX_RING_SIZE * elen; - - ncookies = VIO_MAX_RING_COOKIES; - dring = ldc_alloc_exp_dring(port->vio.lp, len, - dr->cookies, &ncookies, - (LDC_MAP_SHADOW | - LDC_MAP_DIRECT | - LDC_MAP_RW)); - if (IS_ERR(dring)) { - err = PTR_ERR(dring); - goto err_out; - } - - dr->base = dring; - dr->entry_size = elen; - dr->num_entries = VNET_TX_RING_SIZE; - dr->prod = 0; - dr->cons = 0; - port->start_cons = true; /* need an initial trigger */ - dr->pending = VNET_TX_RING_SIZE; - dr->ncookies = ncookies; - - for (i = 0; i < VNET_TX_RING_SIZE; ++i) { - struct vio_net_desc *d; - - d = vio_dring_entry(dr, i); - d->hdr.state = VIO_DESC_FREE; - } - return 0; - -err_out: - sunvnet_port_free_tx_bufs_common(port); - - return err; -} - -#ifdef CONFIG_NET_POLL_CONTROLLER -void sunvnet_poll_controller_common(struct net_device *dev, struct vnet *vp) -{ - struct vnet_port *port; - unsigned long flags; - - spin_lock_irqsave(&vp->lock, flags); - if (!list_empty(&vp->port_list)) { - port = list_entry(vp->port_list.next, struct vnet_port, list); - napi_schedule(&port->napi); - } - spin_unlock_irqrestore(&vp->lock, flags); -} -EXPORT_SYMBOL_GPL(sunvnet_poll_controller_common); -#endif - -void sunvnet_port_add_txq_common(struct vnet_port *port) -{ - struct vnet *vp = port->vp; - int smallest = 0; - int i; - - /* find the first least-used q - * When there are more ldoms than q's, we start to - * double up on ports per queue. - */ - for (i = 0; i < VNET_MAX_TXQS; i++) { - if (vp->q_used[i] == 0) { - smallest = i; - break; - } - if (vp->q_used[i] < vp->q_used[smallest]) - smallest = i; - } - - vp->nports++; - vp->q_used[smallest]++; - port->q_index = smallest; -} -EXPORT_SYMBOL_GPL(sunvnet_port_add_txq_common); - -void sunvnet_port_rm_txq_common(struct vnet_port *port) -{ - port->vp->nports--; - port->vp->q_used[port->q_index]--; - port->q_index = 0; -} -EXPORT_SYMBOL_GPL(sunvnet_port_rm_txq_common); diff --git a/drivers/net/ethernet/sun/sunvnet_common.h b/drivers/net/ethernet/sun/sunvnet_common.h deleted file mode 100644 index 5416a3c..00000000 --- a/drivers/net/ethernet/sun/sunvnet_common.h +++ /dev/null @@ -1,157 +0,0 @@ -/* SPDX-License-Identifier: GPL-2.0 */ -#ifndef _SUNVNETCOMMON_H -#define _SUNVNETCOMMON_H - -#include - -/* length of time (or less) we expect pending descriptors to be marked - * as VIO_DESC_DONE and skbs ready to be freed - */ -#define VNET_CLEAN_TIMEOUT ((HZ / 100) + 1) - -#define VNET_MAXPACKET (65535ULL + ETH_HLEN + VLAN_HLEN) -#define VNET_TX_RING_SIZE 512 -#define VNET_TX_WAKEUP_THRESH(dr) ((dr)->pending / 4) - -#define VNET_MINTSO 2048 /* VIO protocol's minimum TSO len */ -#define VNET_MAXTSO 65535 /* VIO protocol's maximum TSO len */ - -#define VNET_MAX_MTU 65535 - -/* VNET packets are sent in buffers with the first 6 bytes skipped - * so that after the ethernet header the IPv4/IPv6 headers are aligned - * properly. - */ -#define VNET_PACKET_SKIP 6 - -#define VNET_MAXCOOKIES (VNET_MAXPACKET / PAGE_SIZE + 1) - -#define VNET_MAX_TXQS 16 - -struct vnet_tx_entry { - struct sk_buff *skb; - unsigned int ncookies; - struct ldc_trans_cookie cookies[VNET_MAXCOOKIES]; -}; - -struct vnet; - -struct vnet_port_stats { - /* keep them all the same size */ - u32 rx_bytes; - u32 tx_bytes; - u32 rx_packets; - u32 tx_packets; - u32 event_up; - u32 event_reset; - u32 q_placeholder; -}; - -#define NUM_VNET_PORT_STATS (sizeof(struct vnet_port_stats) / sizeof(u32)) - -/* Structure to describe a vnet-port or vsw-port in the MD. - * If the vsw bit is set, this structure represents a vswitch - * port, and the net_device can be found from ->dev. If the - * vsw bit is not set, the net_device is available from ->vp->dev. - * See the VNET_PORT_TO_NET_DEVICE macro below. - */ -struct vnet_port { - struct vio_driver_state vio; - - struct vnet_port_stats stats; - - struct hlist_node hash; - u8 raddr[ETH_ALEN]; - unsigned switch_port:1; - unsigned tso:1; - unsigned vsw:1; - unsigned __pad:13; - - struct vnet *vp; - struct net_device *dev; - - struct vnet_tx_entry tx_bufs[VNET_TX_RING_SIZE]; - - struct list_head list; - - u32 stop_rx_idx; - bool stop_rx; - bool start_cons; - - struct timer_list clean_timer; - - u64 rmtu; - u16 tsolen; - - struct napi_struct napi; - u32 napi_stop_idx; - bool napi_resume; - int rx_event; - u16 q_index; -}; - -static inline struct vnet_port *to_vnet_port(struct vio_driver_state *vio) -{ - return container_of(vio, struct vnet_port, vio); -} - -#define VNET_PORT_HASH_SIZE 16 -#define VNET_PORT_HASH_MASK (VNET_PORT_HASH_SIZE - 1) - -static inline unsigned int vnet_hashfn(u8 *mac) -{ - unsigned int val = mac[4] ^ mac[5]; - - return val & (VNET_PORT_HASH_MASK); -} - -struct vnet_mcast_entry { - u8 addr[ETH_ALEN]; - u8 sent; - u8 hit; - struct vnet_mcast_entry *next; -}; - -struct vnet { - spinlock_t lock; /* Protects port_list and port_hash. */ - struct net_device *dev; - u32 msg_enable; - u8 q_used[VNET_MAX_TXQS]; - struct list_head port_list; - struct hlist_head port_hash[VNET_PORT_HASH_SIZE]; - struct vnet_mcast_entry *mcast_list; - struct list_head list; - u64 local_mac; - int nports; -}; - -/* Def used by common code to get the net_device from the proper location */ -#define VNET_PORT_TO_NET_DEVICE(__port) \ - ((__port)->vsw ? (__port)->dev : (__port)->vp->dev) - -/* Common funcs */ -void sunvnet_clean_timer_expire_common(struct timer_list *t); -int sunvnet_open_common(struct net_device *dev); -int sunvnet_close_common(struct net_device *dev); -void sunvnet_set_rx_mode_common(struct net_device *dev, struct vnet *vp); -int sunvnet_set_mac_addr_common(struct net_device *dev, void *p); -void sunvnet_tx_timeout_common(struct net_device *dev, unsigned int txqueue); -netdev_tx_t -sunvnet_start_xmit_common(struct sk_buff *skb, struct net_device *dev, - struct vnet_port *(*vnet_tx_port) - (struct sk_buff *, struct net_device *)); -#ifdef CONFIG_NET_POLL_CONTROLLER -void sunvnet_poll_controller_common(struct net_device *dev, struct vnet *vp); -#endif -void sunvnet_event_common(void *arg, int event); -int sunvnet_send_attr_common(struct vio_driver_state *vio); -int sunvnet_handle_attr_common(struct vio_driver_state *vio, void *arg); -void sunvnet_handshake_complete_common(struct vio_driver_state *vio); -int sunvnet_poll_common(struct napi_struct *napi, int budget); -void sunvnet_port_free_tx_bufs_common(struct vnet_port *port); -void vnet_port_reset(struct vnet_port *port); -bool sunvnet_port_is_up_common(struct vnet_port *vnet); -void sunvnet_port_add_txq_common(struct vnet_port *port); -void sunvnet_port_rm_txq_common(struct vnet_port *port); - -#endif /* _SUNVNETCOMMON_H */ From patchwork Fri Jan 6 22:00:19 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Anirudh Venkataramanan X-Patchwork-Id: 13091841 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id AFB80C678D9 for ; Fri, 6 Jan 2023 21:59:32 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236569AbjAFV73 (ORCPT ); Fri, 6 Jan 2023 16:59:29 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:42324 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235901AbjAFV7V (ORCPT ); Fri, 6 Jan 2023 16:59:21 -0500 Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 702FF6B5F1; Fri, 6 Jan 2023 13:59:20 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1673042360; x=1704578360; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=SSz6BzP2REm+D8MWs8FHchah/89P2zMYb+LTILGyOH0=; b=Lz0IRZY4A30z7DBfUdVognpVAMZY8c76zAH3/VaaDISgHFJ36lP+ZwH7 roBhmtYRZAn+/DUZ9cf/ml4doaGHs6LhdZjzFhorXITEaNanancI2hZ1J miOGyKf3HdQifzDL8I4KkM4qBosaLPRS8qf0n1/w9/BHNS2XQTXAst2m2 L+Jms2mEYhvZvyiMEZZ7HlVLde2bsL3+BEpKWTOLMDubGPEGXITO3et1o Z631Fkf3x2W+xB92Dg3uQBdu483JIgcc7oll408BKlrMw6527KrGOACS4 YMW1BLKlrYnzs6/hA6QXZaTkq4XWi2n6sR/OeFRBFz1mxF74UlFtakoO2 w==; X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="387030718" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="387030718" Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Jan 2023 13:59:18 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="763652895" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="763652895" Received: from avenkata-desk0.sc.intel.com ([172.25.112.60]) by fmsmga002.fm.intel.com with ESMTP; 06 Jan 2023 13:59:16 -0800 From: Anirudh Venkataramanan To: netdev@vger.kernel.org Cc: linux-pci@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-mips@vger.kernel.org, linux-trace-kernel@vger.kernel.org, sparclinux@vger.kernel.org, Anirudh Venkataramanan , Leon Romanovsky Subject: [PATCH net-next 6/7] sunvnet: Remove event tracing file Date: Fri, 6 Jan 2023 14:00:19 -0800 Message-Id: <20230106220020.1820147-7-anirudh.venkataramanan@intel.com> X-Mailer: git-send-email 2.37.2 In-Reply-To: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> References: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org An earlier patch removed the Sun LDOM vswitch and sunvnet drivers, and as a result, nothing includes sunvnet.h anymore. Remove it. Note: checkpatch complains "WARNING: added, moved or deleted file(s), does MAINTAINERS need updating?". The file being removed doesn't have its own entry in the MAINTAINERS file, so there's nothing to remove. Cc: Leon Romanovsky Signed-off-by: Anirudh Venkataramanan --- include/trace/events/sunvnet.h | 140 --------------------------------- 1 file changed, 140 deletions(-) delete mode 100644 include/trace/events/sunvnet.h diff --git a/include/trace/events/sunvnet.h b/include/trace/events/sunvnet.h deleted file mode 100644 index 8d444f1..00000000 --- a/include/trace/events/sunvnet.h +++ /dev/null @@ -1,140 +0,0 @@ -/* SPDX-License-Identifier: GPL-2.0 */ -#undef TRACE_SYSTEM -#define TRACE_SYSTEM sunvnet - -#if !defined(_TRACE_SUNVNET_H) || defined(TRACE_HEADER_MULTI_READ) -#define _TRACE_SUNVNET_H - -#include - -TRACE_EVENT(vnet_rx_one, - - TP_PROTO(int lsid, int rsid, int index, int needs_ack), - - TP_ARGS(lsid, rsid, index, needs_ack), - - TP_STRUCT__entry( - __field(int, lsid) - __field(int, rsid) - __field(int, index) - __field(int, needs_ack) - ), - - TP_fast_assign( - __entry->lsid = lsid; - __entry->rsid = rsid; - __entry->index = index; - __entry->needs_ack = needs_ack; - ), - - TP_printk("(%x:%x) walk_rx_one index %d; needs_ack %d", - __entry->lsid, __entry->rsid, - __entry->index, __entry->needs_ack) -); - -DECLARE_EVENT_CLASS(vnet_tx_stopped_ack_template, - - TP_PROTO(int lsid, int rsid, int ack_end, int npkts), - - TP_ARGS(lsid, rsid, ack_end, npkts), - - TP_STRUCT__entry( - __field(int, lsid) - __field(int, rsid) - __field(int, ack_end) - __field(int, npkts) - ), - - TP_fast_assign( - __entry->lsid = lsid; - __entry->rsid = rsid; - __entry->ack_end = ack_end; - __entry->npkts = npkts; - ), - - TP_printk("(%x:%x) stopped ack for %d; npkts %d", - __entry->lsid, __entry->rsid, - __entry->ack_end, __entry->npkts) -); -DEFINE_EVENT(vnet_tx_stopped_ack_template, vnet_tx_send_stopped_ack, - TP_PROTO(int lsid, int rsid, int ack_end, int npkts), - TP_ARGS(lsid, rsid, ack_end, npkts)); -DEFINE_EVENT(vnet_tx_stopped_ack_template, vnet_tx_defer_stopped_ack, - TP_PROTO(int lsid, int rsid, int ack_end, int npkts), - TP_ARGS(lsid, rsid, ack_end, npkts)); -DEFINE_EVENT(vnet_tx_stopped_ack_template, vnet_tx_pending_stopped_ack, - TP_PROTO(int lsid, int rsid, int ack_end, int npkts), - TP_ARGS(lsid, rsid, ack_end, npkts)); - -TRACE_EVENT(vnet_rx_stopped_ack, - - TP_PROTO(int lsid, int rsid, int end), - - TP_ARGS(lsid, rsid, end), - - TP_STRUCT__entry( - __field(int, lsid) - __field(int, rsid) - __field(int, end) - ), - - TP_fast_assign( - __entry->lsid = lsid; - __entry->rsid = rsid; - __entry->end = end; - ), - - TP_printk("(%x:%x) stopped ack for index %d", - __entry->lsid, __entry->rsid, __entry->end) -); - -TRACE_EVENT(vnet_tx_trigger, - - TP_PROTO(int lsid, int rsid, int start, int err), - - TP_ARGS(lsid, rsid, start, err), - - TP_STRUCT__entry( - __field(int, lsid) - __field(int, rsid) - __field(int, start) - __field(int, err) - ), - - TP_fast_assign( - __entry->lsid = lsid; - __entry->rsid = rsid; - __entry->start = start; - __entry->err = err; - ), - - TP_printk("(%x:%x) Tx trigger for %d sent with err %d %s", - __entry->lsid, __entry->rsid, __entry->start, - __entry->err, __entry->err > 0 ? "(ok)" : " ") -); - -TRACE_EVENT(vnet_skip_tx_trigger, - - TP_PROTO(int lsid, int rsid, int last), - - TP_ARGS(lsid, rsid, last), - - TP_STRUCT__entry( - __field(int, lsid) - __field(int, rsid) - __field(int, last) - ), - - TP_fast_assign( - __entry->lsid = lsid; - __entry->rsid = rsid; - __entry->last = last; - ), - - TP_printk("(%x:%x) Skip Tx trigger. Last trigger sent was %d", - __entry->lsid, __entry->rsid, __entry->last) -); -#endif /* _TRACE_SOCK_H */ - -/* This part must be outside protection */ -#include From patchwork Fri Jan 6 22:00:20 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Anirudh Venkataramanan X-Patchwork-Id: 13091842 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 588EDC6379F for ; Fri, 6 Jan 2023 21:59:35 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236568AbjAFV7c (ORCPT ); Fri, 6 Jan 2023 16:59:32 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:42322 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S236161AbjAFV7V (ORCPT ); Fri, 6 Jan 2023 16:59:21 -0500 Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 2AEB2728AD; Fri, 6 Jan 2023 13:59:21 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1673042361; x=1704578361; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=0AI0odRghP/P2eZTU9+vUvGrnUWUgK7vgCVx3xkIU9s=; b=I1PcuUKOF1qxryEjviOCk2Q/LT0NFLlqxVVmdatbpPrT7NPUjburRl5G V28RpTk/E0xXGgs3C/daHNlwUUuQbojodQF1Irc2c/iq0jUvMSjPQSclN 8/q5oN0jAHaCKyU7TmA/8cRDbNDpJ13ps/KRgpVLSHJyckPS7S0FF4Omk 6XHYXdo8c1b6Q3/GwPB8epW3SoW7xHho3bwLX+CZQy6qgZgdJzdGhrNyK ZRjq7sT+bV/XFfJ/wJg1mgWkyGL8i8IGqp4CXJTFa/93yiMgGq3xw/qCn ctHuaawCT6apjgSTO5eP60VHd6UtvqqeOvxUXpyhnHJKUp89qg18jtuXT Q==; X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="387030720" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="387030720" Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Jan 2023 13:59:18 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6500,9779,10582"; a="763652899" X-IronPort-AV: E=Sophos;i="5.96,306,1665471600"; d="scan'208";a="763652899" Received: from avenkata-desk0.sc.intel.com ([172.25.112.60]) by fmsmga002.fm.intel.com with ESMTP; 06 Jan 2023 13:59:17 -0800 From: Anirudh Venkataramanan To: netdev@vger.kernel.org Cc: linux-pci@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-mips@vger.kernel.org, linux-trace-kernel@vger.kernel.org, sparclinux@vger.kernel.org, Anirudh Venkataramanan , Leon Romanovsky Subject: [PATCH net-next 7/7] sparc: configs: Remove references to CONFIG_SUNVNET and CONFIG_LDMVSW Date: Fri, 6 Jan 2023 14:00:20 -0800 Message-Id: <20230106220020.1820147-8-anirudh.venkataramanan@intel.com> X-Mailer: git-send-email 2.37.2 In-Reply-To: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> References: <20230106220020.1820147-1-anirudh.venkataramanan@intel.com> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org An earlier patch removed the Sun LDOM vswitch and sunvnet drivers. Remove references to CONFIG_SUNVNET and CONFIG_LDMVSW from the sparc64 defconfig. Cc: Leon Romanovsky Signed-off-by: Anirudh Venkataramanan --- arch/sparc/configs/sparc64_defconfig | 2 -- 1 file changed, 2 deletions(-) diff --git a/arch/sparc/configs/sparc64_defconfig b/arch/sparc/configs/sparc64_defconfig index 1809909..a2c76e8 100644 --- a/arch/sparc/configs/sparc64_defconfig +++ b/arch/sparc/configs/sparc64_defconfig @@ -95,8 +95,6 @@ CONFIG_MII=m CONFIG_SUNLANCE=m CONFIG_HAPPYMEAL=y CONFIG_SUNGEM=m -CONFIG_SUNVNET=m -CONFIG_LDMVSW=m CONFIG_NET_PCI=y CONFIG_E1000=m CONFIG_E1000E=m