/* * This file is part of the Chelsio T4 Ethernet driver for Linux. * * Copyright (c) 2003-2014 Chelsio Communications, Inc. All rights reserved. * * This software is available to you under a choice of one of two * licenses. You may choose to be licensed under the terms of the GNU * General Public License (GPL) Version 2, available from the file * COPYING in the main directory of this source tree, or the * OpenIB.org BSD license below: * * Redistribution and use in source and binary forms, with or * without modification, are permitted provided that the following * conditions are met: * * - Redistributions of source code must retain the above * copyright notice, this list of conditions and the following * disclaimer. * * - Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials * provided with the distribution. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "cxgb4.h" #include "t4_regs.h" #include "t4_msg.h" #include "t4fw_api.h" #include "cxgb4_dcb.h" #include "l2t.h" #include <../drivers/net/bonding/bonding.h> #ifdef DRV_VERSION #undef DRV_VERSION #endif #define DRV_VERSION "2.0.0-ko" #define DRV_DESC "Chelsio T4/T5 Network Driver" /* * Max interrupt hold-off timer value in us. Queues fall back to this value * under extreme memory pressure so it's largish to give the system time to * recover. */ #define MAX_SGE_TIMERVAL 200U enum { /* * Physical Function provisioning constants. */ PFRES_NVI = 4, /* # of Virtual Interfaces */ PFRES_NETHCTRL = 128, /* # of EQs used for ETH or CTRL Qs */ PFRES_NIQFLINT = 128, /* # of ingress Qs/w Free List(s)/intr */ PFRES_NEQ = 256, /* # of egress queues */ PFRES_NIQ = 0, /* # of ingress queues */ PFRES_TC = 0, /* PCI-E traffic class */ PFRES_NEXACTF = 128, /* # of exact MPS filters */ PFRES_R_CAPS = FW_CMD_CAP_PF, PFRES_WX_CAPS = FW_CMD_CAP_PF, #ifdef CONFIG_PCI_IOV /* * Virtual Function provisioning constants. We need two extra Ingress * Queues with Interrupt capability to serve as the VF's Firmware * Event Queue and Forwarded Interrupt Queue (when using MSI mode) -- * neither will have Free Lists associated with them). For each * Ethernet/Control Egress Queue and for each Free List, we need an * Egress Context. */ VFRES_NPORTS = 1, /* # of "ports" per VF */ VFRES_NQSETS = 2, /* # of "Queue Sets" per VF */ VFRES_NVI = VFRES_NPORTS, /* # of Virtual Interfaces */ VFRES_NETHCTRL = VFRES_NQSETS, /* # of EQs used for ETH or CTRL Qs */ VFRES_NIQFLINT = VFRES_NQSETS+2,/* # of ingress Qs/w Free List(s)/intr */ VFRES_NEQ = VFRES_NQSETS*2, /* # of egress queues */ VFRES_NIQ = 0, /* # of non-fl/int ingress queues */ VFRES_TC = 0, /* PCI-E traffic class */ VFRES_NEXACTF = 16, /* # of exact MPS filters */ VFRES_R_CAPS = FW_CMD_CAP_DMAQ|FW_CMD_CAP_VF|FW_CMD_CAP_PORT, VFRES_WX_CAPS = FW_CMD_CAP_DMAQ|FW_CMD_CAP_VF, #endif }; /* * Provide a Port Access Rights Mask for the specified PF/VF. This is very * static and likely not to be useful in the long run. We really need to * implement some form of persistent configuration which the firmware * controls. */ static unsigned int pfvfres_pmask(struct adapter *adapter, unsigned int pf, unsigned int vf) { unsigned int portn, portvec; /* * Give PF's access to all of the ports. */ if (vf == 0) return FW_PFVF_CMD_PMASK_MASK; /* * For VFs, we'll assign them access to the ports based purely on the * PF. We assign active ports in order, wrapping around if there are * fewer active ports than PFs: e.g. active port[pf % nports]. * Unfortunately the adapter's port_info structs haven't been * initialized yet so we have to compute this. */ if (adapter->params.nports == 0) return 0; portn = pf % adapter->params.nports; portvec = adapter->params.portvec; for (;;) { /* * Isolate the lowest set bit in the port vector. If we're at * the port number that we want, return that as the pmask. * otherwise mask that bit out of the port vector and * decrement our port number ... */ unsigned int pmask = portvec ^ (portvec & (portvec-1)); if (portn == 0) return pmask; portn--; portvec &= ~pmask; } /*NOTREACHED*/ } enum { MAX_TXQ_ENTRIES = 16384, MAX_CTRL_TXQ_ENTRIES = 1024, MAX_RSPQ_ENTRIES = 16384, MAX_RX_BUFFERS = 16384, MIN_TXQ_ENTRIES = 32, MIN_CTRL_TXQ_ENTRIES = 32, MIN_RSPQ_ENTRIES = 128, MIN_FL_ENTRIES = 16 }; /* Host shadow copy of ingress filter entry. This is in host native format * and doesn't match the ordering or bit order, etc. of the hardware of the * firmware command. The use of bit-field structure elements is purely to * remind ourselves of the field size limitations and save memory in the case * where the filter table is large. */ struct filter_entry { /* Administrative fields for filter. */ u32 valid:1; /* filter allocated and valid */ u32 locked:1; /* filter is administratively locked */ u32 pending:1; /* filter action is pending firmware reply */ u32 smtidx:8; /* Source MAC Table index for smac */ struct l2t_entry *l2t; /* Layer Two Table entry for dmac */ /* The filter itself. Most of this is a straight copy of information * provided by the extended ioctl(). Some fields are translated to * internal forms -- for instance the Ingress Queue ID passed in from * the ioctl() is translated into the Absolute Ingress Queue ID. */ struct ch_filter_specification fs; }; #define DFLT_MSG_ENABLE (NETIF_MSG_DRV | NETIF_MSG_PROBE | NETIF_MSG_LINK | \ NETIF_MSG_TIMER | NETIF_MSG_IFDOWN | NETIF_MSG_IFUP |\ NETIF_MSG_RX_ERR | NETIF_MSG_TX_ERR) #define CH_DEVICE(devid, data) { PCI_VDEVICE(CHELSIO, devid), (data) } static const struct pci_device_id cxgb4_pci_tbl[] = { CH_DEVICE(0xa000, 0), /* PE10K */ CH_DEVICE(0x4001, -1), CH_DEVICE(0x4002, -1), CH_DEVICE(0x4003, -1), CH_DEVICE(0x4004, -1), CH_DEVICE(0x4005, -1), CH_DEVICE(0x4006, -1), CH_DEVICE(0x4007, -1), CH_DEVICE(0x4008, -1), CH_DEVICE(0x4009, -1), CH_DEVICE(0x400a, -1), CH_DEVICE(0x400d, -1), CH_DEVICE(0x400e, -1), CH_DEVICE(0x4080, -1), CH_DEVICE(0x4081, -1), CH_DEVICE(0x4082, -1), CH_DEVICE(0x4083, -1), CH_DEVICE(0x4084, -1), CH_DEVICE(0x4085, -1), CH_DEVICE(0x4086, -1), CH_DEVICE(0x4087, -1), CH_DEVICE(0x4088, -1), CH_DEVICE(0x4401, 4), CH_DEVICE(0x4402, 4), CH_DEVICE(0x4403, 4), CH_DEVICE(0x4404, 4), CH_DEVICE(0x4405, 4), CH_DEVICE(0x4406, 4), CH_DEVICE(0x4407, 4), CH_DEVICE(0x4408, 4), CH_DEVICE(0x4409, 4), CH_DEVICE(0x440a, 4), CH_DEVICE(0x440d, 4), CH_DEVICE(0x440e, 4), CH_DEVICE(0x4480, 4), CH_DEVICE(0x4481, 4), CH_DEVICE(0x4482, 4), CH_DEVICE(0x4483, 4), CH_DEVICE(0x4484, 4), CH_DEVICE(0x4485, 4), CH_DEVICE(0x4486, 4), CH_DEVICE(0x4487, 4), CH_DEVICE(0x4488, 4), CH_DEVICE(0x5001, 4), CH_DEVICE(0x5002, 4), CH_DEVICE(0x5003, 4), CH_DEVICE(0x5004, 4), CH_DEVICE(0x5005, 4), CH_DEVICE(0x5006, 4), CH_DEVICE(0x5007, 4), CH_DEVICE(0x5008, 4), CH_DEVICE(0x5009, 4), CH_DEVICE(0x500A, 4), CH_DEVICE(0x500B, 4), CH_DEVICE(0x500C, 4), CH_DEVICE(0x500D, 4), CH_DEVICE(0x500E, 4), CH_DEVICE(0x500F, 4), CH_DEVICE(0x5010, 4), CH_DEVICE(0x5011, 4), CH_DEVICE(0x5012, 4), CH_DEVICE(0x5013, 4), CH_DEVICE(0x5014, 4), CH_DEVICE(0x5015, 4), CH_DEVICE(0x5080, 4), CH_DEVICE(0x5081, 4), CH_DEVICE(0x5082, 4), CH_DEVICE(0x5083, 4), CH_DEVICE(0x5084, 4), CH_DEVICE(0x5085, 4), CH_DEVICE(0x5086, 4), CH_DEVICE(0x5401, 4), CH_DEVICE(0x5402, 4), CH_DEVICE(0x5403, 4), CH_DEVICE(0x5404, 4), CH_DEVICE(0x5405, 4), CH_DEVICE(0x5406, 4), CH_DEVICE(0x5407, 4), CH_DEVICE(0x5408, 4), CH_DEVICE(0x5409, 4), CH_DEVICE(0x540A, 4), CH_DEVICE(0x540B, 4), CH_DEVICE(0x540C, 4), CH_DEVICE(0x540D, 4), CH_DEVICE(0x540E, 4), CH_DEVICE(0x540F, 4), CH_DEVICE(0x5410, 4), CH_DEVICE(0x5411, 4), CH_DEVICE(0x5412, 4), CH_DEVICE(0x5413, 4), CH_DEVICE(0x5414, 4), CH_DEVICE(0x5415, 4), CH_DEVICE(0x5480, 4), CH_DEVICE(0x5481, 4), CH_DEVICE(0x5482, 4), CH_DEVICE(0x5483, 4), CH_DEVICE(0x5484, 4), CH_DEVICE(0x5485, 4), CH_DEVICE(0x5486, 4), { 0, } }; #define FW4_FNAME "cxgb4/t4fw.bin" #define FW5_FNAME "cxgb4/t5fw.bin" #define FW4_CFNAME "cxgb4/t4-config.txt" #define FW5_CFNAME "cxgb4/t5-config.txt" MODULE_DESCRIPTION(DRV_DESC); MODULE_AUTHOR("Chelsio Communications"); MODULE_LICENSE("Dual BSD/GPL"); MODULE_VERSION(DRV_VERSION); MODULE_DEVICE_TABLE(pci, cxgb4_pci_tbl); MODULE_FIRMWARE(FW4_FNAME); MODULE_FIRMWARE(FW5_FNAME); /* * Normally we're willing to become the firmware's Master PF but will be happy * if another PF has already become the Master and initialized the adapter. * Setting "force_init" will cause this driver to forcibly establish itself as * the Master PF and initialize the adapter. */ static uint force_init; module_param(force_init, uint, 0644); MODULE_PARM_DESC(force_init, "Forcibly become Master PF and initialize adapter"); /* * Normally if the firmware we connect to has Configuration File support, we * use that and only fall back to the old Driver-based initialization if the * Configuration File fails for some reason. If force_old_init is set, then * we'll always use the old Driver-based initialization sequence. */ static uint force_old_init; module_param(force_old_init, uint, 0644); MODULE_PARM_DESC(force_old_init, "Force old initialization sequence"); static int dflt_msg_enable = DFLT_MSG_ENABLE; module_param(dflt_msg_enable, int, 0644); MODULE_PARM_DESC(dflt_msg_enable, "Chelsio T4 default message enable bitmap"); /* * The driver uses the best interrupt scheme available on a platform in the * order MSI-X, MSI, legacy INTx interrupts. This parameter determines which * of these schemes the driver may consider as follows: * * msi = 2: choose from among all three options * msi = 1: only consider MSI and INTx interrupts * msi = 0: force INTx interrupts */ static int msi = 2; module_param(msi, int, 0644); MODULE_PARM_DESC(msi, "whether to use INTx (0), MSI (1) or MSI-X (2)"); /* * Queue interrupt hold-off timer values. Queues default to the first of these * upon creation. */ static unsigned int intr_holdoff[SGE_NTIMERS - 1] = { 5, 10, 20, 50, 100 }; module_param_array(intr_holdoff, uint, NULL, 0644); MODULE_PARM_DESC(intr_holdoff, "values for queue interrupt hold-off timers " "0..4 in microseconds"); static unsigned int intr_cnt[SGE_NCOUNTERS - 1] = { 4, 8, 16 }; module_param_array(intr_cnt, uint, NULL, 0644); MODULE_PARM_DESC(intr_cnt, "thresholds 1..3 for queue interrupt packet counters"); /* * Normally we tell the chip to deliver Ingress Packets into our DMA buffers * offset by 2 bytes in order to have the IP headers line up on 4-byte * boundaries. This is a requirement for many architectures which will throw * a machine check fault if an attempt is made to access one of the 4-byte IP * header fields on a non-4-byte boundary. And it's a major performance issue * even on some architectures which allow it like some implementations of the * x86 ISA. However, some architectures don't mind this and for some very * edge-case performance sensitive applications (like forwarding large volumes * of small packets), setting this DMA offset to 0 will decrease the number of * PCI-E Bus transfers enough to measurably affect performance. */ static int rx_dma_offset = 2; static bool vf_acls; #ifdef CONFIG_PCI_IOV module_param(vf_acls, bool, 0644); MODULE_PARM_DESC(vf_acls, "if set enable virtualization L2 ACL enforcement"); /* Configure the number of PCI-E Virtual Function which are to be instantiated * on SR-IOV Capable Physical Functions. */ static unsigned int num_vf[NUM_OF_PF_WITH_SRIOV]; module_param_array(num_vf, uint, NULL, 0644); MODULE_PARM_DESC(num_vf, "number of VFs for each of PFs 0-3"); #endif /* TX Queue select used to determine what algorithm to use for selecting TX * queue. Select between the kernel provided function (select_queue=0) or user * cxgb_select_queue function (select_queue=1) * * Default: select_queue=0 */ static int select_queue; module_param(select_queue, int, 0644); MODULE_PARM_DESC(select_queue, "Select between kernel provided method of selecting or driver method of selecting TX queue. Default is kernel method."); /* * The filter TCAM has a fixed portion and a variable portion. The fixed * portion can match on source/destination IP IPv4/IPv6 addresses and TCP/UDP * ports. The variable portion is 36 bits which can include things like Exact * Match MAC Index (9 bits), Ether Type (16 bits), IP Protocol (8 bits), * [Inner] VLAN Tag (17 bits), etc. which, if all were somehow selected, would * far exceed the 36-bit budget for this "compressed" header portion of the * filter. Thus, we have a scarce resource which must be carefully managed. * * By default we set this up to mostly match the set of filter matching * capabilities of T3 but with accommodations for some of T4's more * interesting features: * * { IP Fragment (1), MPS Match Type (3), IP Protocol (8), * [Inner] VLAN (17), Port (3), FCoE (1) } */ enum { TP_VLAN_PRI_MAP_DEFAULT = HW_TPL_FR_MT_PR_IV_P_FC, TP_VLAN_PRI_MAP_FIRST = FCOE_SHIFT, TP_VLAN_PRI_MAP_LAST = FRAGMENTATION_SHIFT, }; static unsigned int tp_vlan_pri_map = TP_VLAN_PRI_MAP_DEFAULT; module_param(tp_vlan_pri_map, uint, 0644); MODULE_PARM_DESC(tp_vlan_pri_map, "global compressed filter configuration"); static struct dentry *cxgb4_debugfs_root; static LIST_HEAD(adapter_list); static DEFINE_MUTEX(uld_mutex); /* Adapter list to be accessed from atomic context */ static LIST_HEAD(adap_rcu_list); static DEFINE_SPINLOCK(adap_rcu_lock); static struct cxgb4_uld_info ulds[CXGB4_ULD_MAX]; static const char *uld_str[] = { "RDMA", "iSCSI" }; static void link_report(struct net_device *dev) { if (!netif_carrier_ok(dev)) netdev_info(dev, "link down\n"); else { static const char *fc[] = { "no", "Rx", "Tx", "Tx/Rx" }; const char *s = "10Mbps"; const struct port_info *p = netdev_priv(dev); switch (p->link_cfg.speed) { case 10000: s = "10Gbps"; break; case 1000: s = "1000Mbps"; break; case 100: s = "100Mbps"; break; case 40000: s = "40Gbps"; break; } netdev_info(dev, "link up, %s, full-duplex, %s PAUSE\n", s, fc[p->link_cfg.fc]); } } #ifdef CONFIG_CHELSIO_T4_DCB /* Set up/tear down Data Center Bridging Priority mapping for a net device. */ static void dcb_tx_queue_prio_enable(struct net_device *dev, int enable) { struct port_info *pi = netdev_priv(dev); struct adapter *adap = pi->adapter; struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset]; int i; /* We use a simple mapping of Port TX Queue Index to DCB * Priority when we're enabling DCB. */ for (i = 0; i < pi->nqsets; i++, txq++) { u32 name, value; int err; name = (FW_PARAMS_MNEM(FW_PARAMS_MNEM_DMAQ) | FW_PARAMS_PARAM_X(FW_PARAMS_PARAM_DMAQ_EQ_DCBPRIO_ETH) | FW_PARAMS_PARAM_YZ(txq->q.cntxt_id)); value = enable ? i : 0xffffffff; /* Since we can be called while atomic (from "interrupt * level") we need to issue the Set Parameters Commannd * without sleeping (timeout < 0). */ err = t4_set_params_nosleep(adap, adap->mbox, adap->fn, 0, 1, &name, &value); if (err) dev_err(adap->pdev_dev, "Can't %s DCB Priority on port %d, TX Queue %d: err=%d\n", enable ? "set" : "unset", pi->port_id, i, -err); else txq->dcb_prio = value; } } #endif /* CONFIG_CHELSIO_T4_DCB */ void t4_os_link_changed(struct adapter *adapter, int port_id, int link_stat) { struct net_device *dev = adapter->port[port_id]; /* Skip changes from disabled ports. */ if (netif_running(dev) && link_stat != netif_carrier_ok(dev)) { if (link_stat) netif_carrier_on(dev); else { #ifdef CONFIG_CHELSIO_T4_DCB cxgb4_dcb_state_init(dev); dcb_tx_queue_prio_enable(dev, false); #endif /* CONFIG_CHELSIO_T4_DCB */ netif_carrier_off(dev); } link_report(dev); } } void t4_os_portmod_changed(const struct adapter *adap, int port_id) { static const char *mod_str[] = { NULL, "LR", "SR", "ER", "passive DA", "active DA", "LRM" }; const struct net_device *dev = adap->port[port_id]; const struct port_info *pi = netdev_priv(dev); if (pi->mod_type == FW_PORT_MOD_TYPE_NONE) netdev_info(dev, "port module unplugged\n"); else if (pi->mod_type < ARRAY_SIZE(mod_str)) netdev_info(dev, "%s module inserted\n", mod_str[pi->mod_type]); } /* * Configure the exact and hash address filters to handle a port's multicast * and secondary unicast MAC addresses. */ static int set_addr_filters(const struct net_device *dev, bool sleep) { u64 mhash = 0; u64 uhash = 0; bool free = true; u16 filt_idx[7]; const u8 *addr[7]; int ret, naddr = 0; const struct netdev_hw_addr *ha; int uc_cnt = netdev_uc_count(dev); int mc_cnt = netdev_mc_count(dev); const struct port_info *pi = netdev_priv(dev); unsigned int mb = pi->adapter->fn; /* first do the secondary unicast addresses */ netdev_for_each_uc_addr(ha, dev) { addr[naddr++] = ha->addr; if (--uc_cnt == 0 || naddr >= ARRAY_SIZE(addr)) { ret = t4_alloc_mac_filt(pi->adapter, mb, pi->viid, free, naddr, addr, filt_idx, &uhash, sleep); if (ret < 0) return ret; free = false; naddr = 0; } } /* next set up the multicast addresses */ netdev_for_each_mc_addr(ha, dev) { addr[naddr++] = ha->addr; if (--mc_cnt == 0 || naddr >= ARRAY_SIZE(addr)) { ret = t4_alloc_mac_filt(pi->adapter, mb, pi->viid, free, naddr, addr, filt_idx, &mhash, sleep); if (ret < 0) return ret; free = false; naddr = 0; } } return t4_set_addr_hash(pi->adapter, mb, pi->viid, uhash != 0, uhash | mhash, sleep); } int dbfifo_int_thresh = 10; /* 10 == 640 entry threshold */ module_param(dbfifo_int_thresh, int, 0644); MODULE_PARM_DESC(dbfifo_int_thresh, "doorbell fifo interrupt threshold"); /* * usecs to sleep while draining the dbfifo */ static int dbfifo_drain_delay = 1000; module_param(dbfifo_drain_delay, int, 0644); MODULE_PARM_DESC(dbfifo_drain_delay, "usecs to sleep while draining the dbfifo"); /* * Set Rx properties of a port, such as promiscruity, address filters, and MTU. * If @mtu is -1 it is left unchanged. */ static int set_rxmode(struct net_device *dev, int mtu, bool sleep_ok) { int ret; struct port_info *pi = netdev_priv(dev); ret = set_addr_filters(dev, sleep_ok); if (ret == 0) ret = t4_set_rxmode(pi->adapter, pi->adapter->fn, pi->viid, mtu, (dev->flags & IFF_PROMISC) ? 1 : 0, (dev->flags & IFF_ALLMULTI) ? 1 : 0, 1, -1, sleep_ok); return ret; } /** * link_start - enable a port * @dev: the port to enable * * Performs the MAC and PHY actions needed to enable a port. */ static int link_start(struct net_device *dev) { int ret; struct port_info *pi = netdev_priv(dev); unsigned int mb = pi->adapter->fn; /* * We do not set address filters and promiscuity here, the stack does * that step explicitly. */ ret = t4_set_rxmode(pi->adapter, mb, pi->viid, dev->mtu, -1, -1, -1, !!(dev->features & NETIF_F_HW_VLAN_CTAG_RX), true); if (ret == 0) { ret = t4_change_mac(pi->adapter, mb, pi->viid, pi->xact_addr_filt, dev->dev_addr, true, true); if (ret >= 0) { pi->xact_addr_filt = ret; ret = 0; } } if (ret == 0) ret = t4_link_start(pi->adapter, mb, pi->tx_chan, &pi->link_cfg); if (ret == 0) { local_bh_disable(); ret = t4_enable_vi_params(pi->adapter, mb, pi->viid, true, true, CXGB4_DCB_ENABLED); local_bh_enable(); } return ret; } int cxgb4_dcb_enabled(const struct net_device *dev) { #ifdef CONFIG_CHELSIO_T4_DCB struct port_info *pi = netdev_priv(dev); return pi->dcb.state == CXGB4_DCB_STATE_FW_ALLSYNCED; #else return 0; #endif } EXPORT_SYMBOL(cxgb4_dcb_enabled); #ifdef CONFIG_CHELSIO_T4_DCB /* Handle a Data Center Bridging update message from the firmware. */ static void dcb_rpl(struct adapter *adap, const struct fw_port_cmd *pcmd) { int port = FW_PORT_CMD_PORTID_GET(ntohl(pcmd->op_to_portid)); struct net_device *dev = adap->port[port]; int old_dcb_enabled = cxgb4_dcb_enabled(dev); int new_dcb_enabled; cxgb4_dcb_handle_fw_update(adap, pcmd); new_dcb_enabled = cxgb4_dcb_enabled(dev); /* If the DCB has become enabled or disabled on the port then we're * going to need to set up/tear down DCB Priority parameters for the * TX Queues associated with the port. */ if (new_dcb_enabled != old_dcb_enabled) dcb_tx_queue_prio_enable(dev, new_dcb_enabled); } #endif /* CONFIG_CHELSIO_T4_DCB */ /* Clear a filter and release any of its resources that we own. This also * clears the filter's "pending" status. */ static void clear_filter(struct adapter *adap, struct filter_entry *f) { /* If the new or old filter have loopback rewriteing rules then we'll * need to free any existing Layer Two Table (L2T) entries of the old * filter rule. The firmware will handle freeing up any Source MAC * Table (SMT) entries used for rewriting Source MAC Addresses in * loopback rules. */ if (f->l2t) cxgb4_l2t_release(f->l2t); /* The zeroing of the filter rule below clears the filter valid, * pending, locked flags, l2t pointer, etc. so it's all we need for * this operation. */ memset(f, 0, sizeof(*f)); } /* Handle a filter write/deletion reply. */ static void filter_rpl(struct adapter *adap, const struct cpl_set_tcb_rpl *rpl) { unsigned int idx = GET_TID(rpl); unsigned int nidx = idx - adap->tids.ftid_base; unsigned int ret; struct filter_entry *f; if (idx >= adap->tids.ftid_base && nidx < (adap->tids.nftids + adap->tids.nsftids)) { idx = nidx; ret = GET_TCB_COOKIE(rpl->cookie); f = &adap->tids.ftid_tab[idx]; if (ret == FW_FILTER_WR_FLT_DELETED) { /* Clear the filter when we get confirmation from the * hardware that the filter has been deleted. */ clear_filter(adap, f); } else if (ret == FW_FILTER_WR_SMT_TBL_FULL) { dev_err(adap->pdev_dev, "filter %u setup failed due to full SMT\n", idx); clear_filter(adap, f); } else if (ret == FW_FILTER_WR_FLT_ADDED) { f->smtidx = (be64_to_cpu(rpl->oldval) >> 24) & 0xff; f->pending = 0; /* asynchronous setup completed */ f->valid = 1; } else { /* Something went wrong. Issue a warning about the * problem and clear everything out. */ dev_err(adap->pdev_dev, "filter %u setup failed with error %u\n", idx, ret); clear_filter(adap, f); } } } /* Response queue handler for the FW event queue. */ static int fwevtq_handler(struct sge_rspq *q, const __be64 *rsp, const struct pkt_gl *gl) { u8 opcode = ((const struct rss_header *)rsp)->opcode; rsp++; /* skip RSS header */ /* FW can send EGR_UPDATEs encapsulated in a CPL_FW4_MSG. */ if (unlikely(opcode == CPL_FW4_MSG && ((const struct cpl_fw4_msg *)rsp)->type == FW_TYPE_RSSCPL)) { rsp++; opcode = ((const struct rss_header *)rsp)->opcode; rsp++; if (opcode != CPL_SGE_EGR_UPDATE) { dev_err(q->adap->pdev_dev, "unexpected FW4/CPL %#x on FW event queue\n" , opcode); goto out; } } if (likely(opcode == CPL_SGE_EGR_UPDATE)) { const struct cpl_sge_egr_update *p = (void *)rsp; unsigned int qid = EGR_QID(ntohl(p->opcode_qid)); struct sge_txq *txq; txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start]; txq->restarts++; if ((u8 *)txq < (u8 *)q->adap->sge.ofldtxq) { struct sge_eth_txq *eq; eq = container_of(txq, struct sge_eth_txq, q); netif_tx_wake_queue(eq->txq); } else { struct sge_ofld_txq *oq; oq = container_of(txq, struct sge_ofld_txq, q); tasklet_schedule(&oq->qresume_tsk); } } else if (opcode == CPL_FW6_MSG || opcode == CPL_FW4_MSG) { const struct cpl_fw6_msg *p = (void *)rsp; #ifdef CONFIG_CHELSIO_T4_DCB const struct fw_port_cmd *pcmd = (const void *)p->data; unsigned int cmd = FW_CMD_OP_GET(ntohl(pcmd->op_to_portid)); unsigned int action = FW_PORT_CMD_ACTION_GET(ntohl(pcmd->action_to_len16)); if (cmd == FW_PORT_CMD && action == FW_PORT_ACTION_GET_PORT_INFO) { int port = FW_PORT_CMD_PORTID_GET( be32_to_cpu(pcmd->op_to_portid)); struct net_device *dev = q->adap->port[port]; int state_input = ((pcmd->u.info.dcbxdis_pkd & FW_PORT_CMD_DCBXDIS) ? CXGB4_DCB_INPUT_FW_DISABLED : CXGB4_DCB_INPUT_FW_ENABLED); cxgb4_dcb_state_fsm(dev, state_input); } if (cmd == FW_PORT_CMD && action == FW_PORT_ACTION_L2_DCB_CFG) dcb_rpl(q->adap, pcmd); else #endif if (p->type == 0) t4_handle_fw_rpl(q->adap, p->data); } else if (opcode == CPL_L2T_WRITE_RPL) { const struct cpl_l2t_write_rpl *p = (void *)rsp; do_l2t_write_rpl(q->adap, p); } else if (opcode == CPL_SET_TCB_RPL) { const struct cpl_set_tcb_rpl *p = (void *)rsp; filter_rpl(q->adap, p); } else dev_err(q->adap->pdev_dev, "unexpected CPL %#x on FW event queue\n", opcode); out: return 0; } /** * uldrx_handler - response queue handler for ULD queues * @q: the response queue that received the packet * @rsp: the response queue descriptor holding the offload message * @gl: the gather list of packet fragments * * Deliver an ingress offload packet to a ULD. All processing is done by * the ULD, we just maintain statistics. */ static int uldrx_handler(struct sge_rspq *q, const __be64 *rsp, const struct pkt_gl *gl) { struct sge_ofld_rxq *rxq = container_of(q, struct sge_ofld_rxq, rspq); /* FW can send CPLs encapsulated in a CPL_FW4_MSG. */ if (((const struct rss_header *)rsp)->opcode == CPL_FW4_MSG && ((const struct cpl_fw4_msg *)(rsp + 1))->type == FW_TYPE_RSSCPL) rsp += 2; if (ulds[q->uld].rx_handler(q->adap->uld_handle[q->uld], rsp, gl)) { rxq->stats.nomem++; return -1; } if (gl == NULL) rxq->stats.imm++; else if (gl == CXGB4_MSG_AN) rxq->stats.an++; else rxq->stats.pkts++; return 0; } static void disable_msi(struct adapter *adapter) { if (adapter->flags & USING_MSIX) { pci_disable_msix(adapter->pdev); adapter->flags &= ~USING_MSIX; } else if (adapter->flags & USING_MSI) { pci_disable_msi(adapter->pdev); adapter->flags &= ~USING_MSI; } } /* * Interrupt handler for non-data events used with MSI-X. */ static irqreturn_t t4_nondata_intr(int irq, void *cookie) { struct adapter *adap = cookie; u32 v = t4_read_reg(adap, MYPF_REG(PL_PF_INT_CAUSE)); if (v & PFSW) { adap->swintr = 1; t4_write_reg(adap, MYPF_REG(PL_PF_INT_CAUSE), v); } t4_slow_intr_handler(adap); return IRQ_HANDLED; } /* * Name the MSI-X interrupts. */ static void name_msix_vecs(struct adapter *adap) { int i, j, msi_idx = 2, n = sizeof(adap->msix_info[0].desc); /* non-data interrupts */ snprintf(adap->msix_info[0].desc, n, "%s", adap->port[0]->name); /* FW events */ snprintf(adap->msix_info[1].desc, n, "%s-FWeventq", adap->port[0]->name); /* Ethernet queues */ for_each_port(adap, j) { struct net_device *d = adap->port[j]; const struct port_info *pi = netdev_priv(d); for (i = 0; i < pi->nqsets; i++, msi_idx++) snprintf(adap->msix_info[msi_idx].desc, n, "%s-Rx%d", d->name, i); } /* offload queues */ for_each_ofldrxq(&adap->sge, i) snprintf(adap->msix_info[msi_idx++].desc, n, "%s-ofld%d", adap->port[0]->name, i); for_each_rdmarxq(&adap->sge, i) snprintf(adap->msix_info[msi_idx++].desc, n, "%s-rdma%d", adap->port[0]->name, i); for_each_rdmaciq(&adap->sge, i) snprintf(adap->msix_info[msi_idx++].desc, n, "%s-rdma-ciq%d", adap->port[0]->name, i); } static int request_msix_queue_irqs(struct adapter *adap) { struct sge *s = &adap->sge; int err, ethqidx, ofldqidx = 0, rdmaqidx = 0, rdmaciqqidx = 0; int msi_index = 2; err = request_irq(adap->msix_info[1].vec, t4_sge_intr_msix, 0, adap->msix_info[1].desc, &s->fw_evtq); if (err) return err; for_each_ethrxq(s, ethqidx) { err = request_irq(adap->msix_info[msi_index].vec, t4_sge_intr_msix, 0, adap->msix_info[msi_index].desc, &s->ethrxq[ethqidx].rspq); if (err) goto unwind; msi_index++; } for_each_ofldrxq(s, ofldqidx) { err = request_irq(adap->msix_info[msi_index].vec, t4_sge_intr_msix, 0, adap->msix_info[msi_index].desc, &s->ofldrxq[ofldqidx].rspq); if (err) goto unwind; msi_index++; } for_each_rdmarxq(s, rdmaqidx) { err = request_irq(adap->msix_info[msi_index].vec, t4_sge_intr_msix, 0, adap->msix_info[msi_index].desc, &s->rdmarxq[rdmaqidx].rspq); if (err) goto unwind; msi_index++; } for_each_rdmaciq(s, rdmaciqqidx) { err = request_irq(adap->msix_info[msi_index].vec, t4_sge_intr_msix, 0, adap->msix_info[msi_index].desc, &s->rdmaciq[rdmaciqqidx].rspq); if (err) goto unwind; msi_index++; } return 0; unwind: while (--rdmaciqqidx >= 0) free_irq(adap->msix_info[--msi_index].vec, &s->rdmaciq[rdmaciqqidx].rspq); while (--rdmaqidx >= 0) free_irq(adap->msix_info[--msi_index].vec, &s->rdmarxq[rdmaqidx].rspq); while (--ofldqidx >= 0) free_irq(adap->msix_info[--msi_index].vec, &s->ofldrxq[ofldqidx].rspq); while (--ethqidx >= 0) free_irq(adap->msix_info[--msi_index].vec, &s->ethrxq[ethqidx].rspq); free_irq(adap->msix_info[1].vec, &s->fw_evtq); return err; } static void free_msix_queue_irqs(struct adapter *adap) { int i, msi_index = 2; struct sge *s = &adap->sge; free_irq(adap->msix_info[1].vec, &s->fw_evtq); for_each_ethrxq(s, i) free_irq(adap->msix_info[msi_index++].vec, &s->ethrxq[i].rspq); for_each_ofldrxq(s, i) free_irq(adap->msix_info[msi_index++].vec, &s->ofldrxq[i].rspq); for_each_rdmarxq(s, i) free_irq(adap->msix_info[msi_index++].vec, &s->rdmarxq[i].rspq); for_each_rdmaciq(s, i) free_irq(adap->msix_info[msi_index++].vec, &s->rdmaciq[i].rspq); } /** * write_rss - write the RSS table for a given port * @pi: the port * @queues: array of queue indices for RSS * * Sets up the portion of the HW RSS table for the port's VI to distribute * packets to the Rx queues in @queues. */ static int write_rss(const struct port_info *pi, const u16 *queues) { u16 *rss; int i, err; const struct sge_eth_rxq *q = &pi->adapter->sge.ethrxq[pi->first_qset]; rss = kmalloc(pi->rss_size * sizeof(u16), GFP_KERNEL); if (!rss) return -ENOMEM; /* map the queue indices to queue ids */ for (i = 0; i < pi->rss_size; i++, queues++) rss[i] = q[*queues].rspq.abs_id; err = t4_config_rss_range(pi->adapter, pi->adapter->fn, pi->viid, 0, pi->rss_size, rss, pi->rss_size); kfree(rss); return err; } /** * setup_rss - configure RSS * @adap: the adapter * * Sets up RSS for each port. */ static int setup_rss(struct adapter *adap) { int i, err; for_each_port(adap, i) { const struct port_info *pi = adap2pinfo(adap, i); err = write_rss(pi, pi->rss); if (err) return err; } return 0; } /* * Return the channel of the ingress queue with the given qid. */ static unsigned int rxq_to_chan(const struct sge *p, unsigned int qid) { qid -= p->ingr_start; return netdev2pinfo(p->ingr_map[qid]->netdev)->tx_chan; } /* * Wait until all NAPI handlers are descheduled. */ static void quiesce_rx(struct adapter *adap) { int i; for (i = 0; i < ARRAY_SIZE(adap->sge.ingr_map); i++) { struct sge_rspq *q = adap->sge.ingr_map[i]; if (q && q->handler) napi_disable(&q->napi); } } /* * Enable NAPI scheduling and interrupt generation for all Rx queues. */ static void enable_rx(struct adapter *adap) { int i; for (i = 0; i < ARRAY_SIZE(adap->sge.ingr_map); i++) { struct sge_rspq *q = adap->sge.ingr_map[i]; if (!q) continue; if (q->handler) napi_enable(&q->napi); /* 0-increment GTS to start the timer and enable interrupts */ t4_write_reg(adap, MYPF_REG(SGE_PF_GTS), SEINTARM(q->intr_params) | INGRESSQID(q->cntxt_id)); } } /** * setup_sge_queues - configure SGE Tx/Rx/response queues * @adap: the adapter * * Determines how many sets of SGE queues to use and initializes them. * We support multiple queue sets per port if we have MSI-X, otherwise * just one queue set per port. */ static int setup_sge_queues(struct adapter *adap) { int err, msi_idx, i, j; struct sge *s = &adap->sge; bitmap_zero(s->starving_fl, MAX_EGRQ); bitmap_zero(s->txq_maperr, MAX_EGRQ); if (adap->flags & USING_MSIX) msi_idx = 1; /* vector 0 is for non-queue interrupts */ else { err = t4_sge_alloc_rxq(adap, &s->intrq, false, adap->port[0], 0, NULL, NULL); if (err) return err; msi_idx = -((int)s->intrq.abs_id + 1); } err = t4_sge_alloc_rxq(adap, &s->fw_evtq, true, adap->port[0], msi_idx, NULL, fwevtq_handler); if (err) { freeout: t4_free_sge_resources(adap); return err; } for_each_port(adap, i) { struct net_device *dev = adap->port[i]; struct port_info *pi = netdev_priv(dev); struct sge_eth_rxq *q = &s->ethrxq[pi->first_qset]; struct sge_eth_txq *t = &s->ethtxq[pi->first_qset]; for (j = 0; j < pi->nqsets; j++, q++) { if (msi_idx > 0) msi_idx++; err = t4_sge_alloc_rxq(adap, &q->rspq, false, dev, msi_idx, &q->fl, t4_ethrx_handler); if (err) goto freeout; q->rspq.idx = j; memset(&q->stats, 0, sizeof(q->stats)); } for (j = 0; j < pi->nqsets; j++, t++) { err = t4_sge_alloc_eth_txq(adap, t, dev, netdev_get_tx_queue(dev, j), s->fw_evtq.cntxt_id); if (err) goto freeout; } } j = s->ofldqsets / adap->params.nports; /* ofld queues per channel */ for_each_ofldrxq(s, i) { struct sge_ofld_rxq *q = &s->ofldrxq[i]; struct net_device *dev = adap->port[i / j]; if (msi_idx > 0) msi_idx++; err = t4_sge_alloc_rxq(adap, &q->rspq, false, dev, msi_idx, q->fl.size ? &q->fl : NULL, uldrx_handler); if (err) goto freeout; memset(&q->stats, 0, sizeof(q->stats)); s->ofld_rxq[i] = q->rspq.abs_id; err = t4_sge_alloc_ofld_txq(adap, &s->ofldtxq[i], dev, s->fw_evtq.cntxt_id); if (err) goto freeout; } for_each_rdmarxq(s, i) { struct sge_ofld_rxq *q = &s->rdmarxq[i]; if (msi_idx > 0) msi_idx++; err = t4_sge_alloc_rxq(adap, &q->rspq, false, adap->port[i], msi_idx, q->fl.size ? &q->fl : NULL, uldrx_handler); if (err) goto freeout; memset(&q->stats, 0, sizeof(q->stats)); s->rdma_rxq[i] = q->rspq.abs_id; } for_each_rdmaciq(s, i) { struct sge_ofld_rxq *q = &s->rdmaciq[i]; if (msi_idx > 0) msi_idx++; err = t4_sge_alloc_rxq(adap, &q->rspq, false, adap->port[i], msi_idx, q->fl.size ? &q->fl : NULL, uldrx_handler); if (err) goto freeout; memset(&q->stats, 0, sizeof(q->stats)); s->rdma_ciq[i] = q->rspq.abs_id; } for_each_port(adap, i) { /* * Note that ->rdmarxq[i].rspq.cntxt_id below is 0 if we don't * have RDMA queues, and that's the right value. */ err = t4_sge_alloc_ctrl_txq(adap, &s->ctrlq[i], adap->port[i], s->fw_evtq.cntxt_id, s->rdmarxq[i].rspq.cntxt_id); if (err) goto freeout; } t4_write_reg(adap, is_t4(adap->params.chip) ? MPS_TRC_RSS_CONTROL : MPS_T5_TRC_RSS_CONTROL, RSSCONTROL(netdev2pinfo(adap->port[0])->tx_chan) | QUEUENUMBER(s->ethrxq[0].rspq.abs_id)); return 0; } /* * Allocate a chunk of memory using kmalloc or, if that fails, vmalloc. * The allocated memory is cleared. */ void *t4_alloc_mem(size_t size) { void *p = kzalloc(size, GFP_KERNEL | __GFP_NOWARN); if (!p) p = vzalloc(size); return p; } /* * Free memory allocated through alloc_mem(). */ static void t4_free_mem(void *addr) { if (is_vmalloc_addr(addr)) vfree(addr); else kfree(addr); } /* Send a Work Request to write the filter at a specified index. We construct * a Firmware Filter Work Request to have the work done and put the indicated * filter into "pending" mode which will prevent any further actions against * it till we get a reply from the firmware on the completion status of the * request. */ static int set_filter_wr(struct adapter *adapter, int fidx) { struct filter_entry *f = &adapter->tids.ftid_tab[fidx]; struct sk_buff *skb; struct fw_filter_wr *fwr; unsigned int ftid; /* If the new filter requires loopback Destination MAC and/or VLAN * rewriting then we need to allocate a Layer 2 Table (L2T) entry for * the filter. */ if (f->fs.newdmac || f->fs.newvlan) { /* allocate L2T entry for new filter */ f->l2t = t4_l2t_alloc_switching(adapter->l2t); if (f->l2t == NULL) return -EAGAIN; if (t4_l2t_set_switching(adapter, f->l2t, f->fs.vlan, f->fs.eport, f->fs.dmac)) { cxgb4_l2t_release(f->l2t); f->l2t = NULL; return -ENOMEM; } } ftid = adapter->tids.ftid_base + fidx; skb = alloc_skb(sizeof(*fwr), GFP_KERNEL | __GFP_NOFAIL); fwr = (struct fw_filter_wr *)__skb_put(skb, sizeof(*fwr)); memset(fwr, 0, sizeof(*fwr)); /* It would be nice to put most of the following in t4_hw.c but most * of the work is translating the cxgbtool ch_filter_specification * into the Work Request and the definition of that structure is * currently in cxgbtool.h which isn't appropriate to pull into the * common code. We may eventually try to come up with a more neutral * filter specification structure but for now it's easiest to simply * put this fairly direct code in line ... */ fwr->op_pkd = htonl(FW_WR_OP(FW_FILTER_WR)); fwr->len16_pkd = htonl(FW_WR_LEN16(sizeof(*fwr)/16)); fwr->tid_to_iq = htonl(V_FW_FILTER_WR_TID(ftid) | V_FW_FILTER_WR_RQTYPE(f->fs.type) | V_FW_FILTER_WR_NOREPLY(0) | V_FW_FILTER_WR_IQ(f->fs.iq)); fwr->del_filter_to_l2tix = htonl(V_FW_FILTER_WR_RPTTID(f->fs.rpttid) | V_FW_FILTER_WR_DROP(f->fs.action == FILTER_DROP) | V_FW_FILTER_WR_DIRSTEER(f->fs.dirsteer) | V_FW_FILTER_WR_MASKHASH(f->fs.maskhash) | V_FW_FILTER_WR_DIRSTEERHASH(f->fs.dirsteerhash) | V_FW_FILTER_WR_LPBK(f->fs.action == FILTER_SWITCH) | V_FW_FILTER_WR_DMAC(f->fs.newdmac) | V_FW_FILTER_WR_SMAC(f->fs.newsmac) | V_FW_FILTER_WR_INSVLAN(f->fs.newvlan == VLAN_INSERT || f->fs.newvlan == VLAN_REWRITE) | V_FW_FILTER_WR_RMVLAN(f->fs.newvlan == VLAN_REMOVE || f->fs.newvlan == VLAN_REWRITE) | V_FW_FILTER_WR_HITCNTS(f->fs.hitcnts) | V_FW_FILTER_WR_TXCHAN(f->fs.eport) | V_FW_FILTER_WR_PRIO(f->fs.prio) | V_FW_FILTER_WR_L2TIX(f->l2t ? f->l2t->idx : 0)); fwr->ethtype = htons(f->fs.val.ethtype); fwr->ethtypem = htons(f->fs.mask.ethtype); fwr->frag_to_ovlan_vldm = (V_FW_FILTER_WR_FRAG(f->fs.val.frag) | V_FW_FILTER_WR_FRAGM(f->fs.mask.frag) | V_FW_FILTER_WR_IVLAN_VLD(f->fs.val.ivlan_vld) | V_FW_FILTER_WR_OVLAN_VLD(f->fs.val.ovlan_vld) | V_FW_FILTER_WR_IVLAN_VLDM(f->fs.mask.ivlan_vld) | V_FW_FILTER_WR_OVLAN_VLDM(f->fs.mask.ovlan_vld)); fwr->smac_sel = 0; fwr->rx_chan_rx_rpl_iq = htons(V_FW_FILTER_WR_RX_CHAN(0) | V_FW_FILTER_WR_RX_RPL_IQ(adapter->sge.fw_evtq.abs_id)); fwr->maci_to_matchtypem = htonl(V_FW_FILTER_WR_MACI(f->fs.val.macidx) | V_FW_FILTER_WR_MACIM(f->fs.mask.macidx) | V_FW_FILTER_WR_FCOE(f->fs.val.fcoe) | V_FW_FILTER_WR_FCOEM(f->fs.mask.fcoe) | V_FW_FILTER_WR_PORT(f->fs.val.iport) | V_FW_FILTER_WR_PORTM(f->fs.mask.iport) | V_FW_FILTER_WR_MATCHTYPE(f->fs.val.matchtype) | V_FW_FILTER_WR_MATCHTYPEM(f->fs.mask.matchtype)); fwr->ptcl = f->fs.val.proto; fwr->ptclm = f->fs.mask.proto; fwr->ttyp = f->fs.val.tos; fwr->ttypm = f->fs.mask.tos; fwr->ivlan = htons(f->fs.val.ivlan); fwr->ivlanm = htons(f->fs.mask.ivlan); fwr->ovlan = htons(f->fs.val.ovlan); fwr->ovlanm = htons(f->fs.mask.ovlan); memcpy(fwr->lip, f->fs.val.lip, sizeof(fwr->lip)); memcpy(fwr->lipm, f->fs.mask.lip, sizeof(fwr->lipm)); memcpy(fwr->fip, f->fs.val.fip, sizeof(fwr->fip)); memcpy(fwr->fipm, f->fs.mask.fip, sizeof(fwr->fipm)); fwr->lp = htons(f->fs.val.lport); fwr->lpm = htons(f->fs.mask.lport); fwr->fp = htons(f->fs.val.fport); fwr->fpm = htons(f->fs.mask.fport); if (f->fs.newsmac) memcpy(fwr->sma, f->fs.smac, sizeof(fwr->sma)); /* Mark the filter as "pending" and ship off the Filter Work Request. * When we get the Work Request Reply we'll clear the pending status. */ f->pending = 1; set_wr_txq(skb, CPL_PRIORITY_CONTROL, f->fs.val.iport & 0x3); t4_ofld_send(adapter, skb); return 0; } /* Delete the filter at a specified index. */ static int del_filter_wr(struct adapter *adapter, int fidx) { struct filter_entry *f = &adapter->tids.ftid_tab[fidx]; struct sk_buff *skb; struct fw_filter_wr *fwr; unsigned int len, ftid; len = sizeof(*fwr); ftid = adapter->tids.ftid_base + fidx; skb = alloc_skb(len, GFP_KERNEL | __GFP_NOFAIL); fwr = (struct fw_filter_wr *)__skb_put(skb, len); t4_mk_filtdelwr(ftid, fwr, adapter->sge.fw_evtq.abs_id); /* Mark the filter as "pending" and ship off the Filter Work Request. * When we get the Work Request Reply we'll clear the pending status. */ f->pending = 1; t4_mgmt_tx(adapter, skb); return 0; } static u16 cxgb_select_queue(struct net_device *dev, struct sk_buff *skb, void *accel_priv, select_queue_fallback_t fallback) { int txq; #ifdef CONFIG_CHELSIO_T4_DCB /* If a Data Center Bridging has been successfully negotiated on this * link then we'll use the skb's priority to map it to a TX Queue. * The skb's priority is determined via the VLAN Tag Priority Code * Point field. */ if (cxgb4_dcb_enabled(dev)) { u16 vlan_tci; int err; err = vlan_get_tag(skb, &vlan_tci); if (unlikely(err)) { if (net_ratelimit()) netdev_warn(dev, "TX Packet without VLAN Tag on DCB Link\n"); txq = 0; } else { txq = (vlan_tci & VLAN_PRIO_MASK) >> VLAN_PRIO_SHIFT; } return txq; } #endif /* CONFIG_CHELSIO_T4_DCB */ if (select_queue) { txq = (skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : smp_processor_id()); while (unlikely(txq >= dev->real_num_tx_queues)) txq -= dev->real_num_tx_queues; return txq; } return fallback(dev, skb) % dev->real_num_tx_queues; } static inline int is_offload(const struct adapter *adap) { return adap->params.offload; } /* * Implementation of ethtool operations. */ static u32 get_msglevel(struct net_device *dev) { return netdev2adap(dev)->msg_enable; } static void set_msglevel(struct net_device *dev, u32 val) { netdev2adap(dev)->msg_enable = val; } static char stats_strings[][ETH_GSTRING_LEN] = { "TxOctetsOK ", "TxFramesOK ", "TxBroadcastFrames ", "TxMulticastFrames ", "TxUnicastFrames ", "TxErrorFrames ", "TxFrames64 ", "TxFrames65To127 ", "TxFrames128To255 ", "TxFrames256To511 ", "TxFrames512To1023 ", "TxFrames1024To1518 ", "TxFrames1519ToMax ", "TxFramesDropped ", "TxPauseFrames ", "TxPPP0Frames ", "TxPPP1Frames ", "TxPPP2Frames ", "TxPPP3Frames ", "TxPPP4Frames ", "TxPPP5Frames ", "TxPPP6Frames ", "TxPPP7Frames ", "RxOctetsOK ", "RxFramesOK ", "RxBroadcastFrames ", "RxMulticastFrames ", "RxUnicastFrames ", "RxFramesTooLong ", "RxJabberErrors ", "RxFCSErrors ", "RxLengthErrors ", "RxSymbolErrors ", "RxRuntFrames ", "RxFrames64 ", "RxFrames65To127 ", "RxFrames128To255 ", "RxFrames256To511 ", "RxFrames512To1023 ", "RxFrames1024To1518 ", "RxFrames1519ToMax ", "RxPauseFrames ", "RxPPP0Frames ", "RxPPP1Frames ", "RxPPP2Frames ", "RxPPP3Frames ", "RxPPP4Frames ", "RxPPP5Frames ", "RxPPP6Frames ", "RxPPP7Frames ", "RxBG0FramesDropped ", "RxBG1FramesDropped ", "RxBG2FramesDropped ", "RxBG3FramesDropped ", "RxBG0FramesTrunc ", "RxBG1FramesTrunc ", "RxBG2FramesTrunc ", "RxBG3FramesTrunc ", "TSO ", "TxCsumOffload ", "RxCsumGood ", "VLANextractions ", "VLANinsertions ", "GROpackets ", "GROmerged ", "WriteCoalSuccess ", "WriteCoalFail ", }; static int get_sset_count(struct net_device *dev, int sset) { switch (sset) { case ETH_SS_STATS: return ARRAY_SIZE(stats_strings); default: return -EOPNOTSUPP; } } #define T4_REGMAP_SIZE (160 * 1024) #define T5_REGMAP_SIZE (332 * 1024) static int get_regs_len(struct net_device *dev) { struct adapter *adap = netdev2adap(dev); if (is_t4(adap->params.chip)) return T4_REGMAP_SIZE; else return T5_REGMAP_SIZE; } static int get_eeprom_len(struct net_device *dev) { return EEPROMSIZE; } static void get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info) { struct adapter *adapter = netdev2adap(dev); strlcpy(info->driver, KBUILD_MODNAME, sizeof(info->driver)); strlcpy(info->version, DRV_VERSION, sizeof(info->version)); strlcpy(info->bus_info, pci_name(adapter->pdev), sizeof(info->bus_info)); if (adapter->params.fw_vers) snprintf(info->fw_version, sizeof(info->fw_version), "%u.%u.%u.%u, TP %u.%u.%u.%u", FW_HDR_FW_VER_MAJOR_GET(adapter->params.fw_vers), FW_HDR_FW_VER_MINOR_GET(adapter->params.fw_vers), FW_HDR_FW_VER_MICRO_GET(adapter->params.fw_vers), FW_HDR_FW_VER_BUILD_GET(adapter->params.fw_vers), FW_HDR_FW_VER_MAJOR_GET(adapter->params.tp_vers), FW_HDR_FW_VER_MINOR_GET(adapter->params.tp_vers), FW_HDR_FW_VER_MICRO_GET(adapter->params.tp_vers), FW_HDR_FW_VER_BUILD_GET(adapter->params.tp_vers)); } static void get_strings(struct net_device *dev, u32 stringset, u8 *data) { if (stringset == ETH_SS_STATS) memcpy(data, stats_strings, sizeof(stats_strings)); } /* * port stats maintained per queue of the port. They should be in the same * order as in stats_strings above. */ struct queue_port_stats { u64 tso; u64 tx_csum; u64 rx_csum; u64 vlan_ex; u64 vlan_ins; u64 gro_pkts; u64 gro_merged; }; static void collect_sge_port_stats(const struct adapter *adap, const struct port_info *p, struct queue_port_stats *s) { int i; const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset]; const struct sge_eth_rxq *rx = &adap->sge.ethrxq[p->first_qset]; memset(s, 0, sizeof(*s)); for (i = 0; i < p->nqsets; i++, rx++, tx++) { s->tso += tx->tso; s->tx_csum += tx->tx_cso; s->rx_csum += rx->stats.rx_cso; s->vlan_ex += rx->stats.vlan_ex; s->vlan_ins += tx->vlan_ins; s->gro_pkts += rx->stats.lro_pkts; s->gro_merged += rx->stats.lro_merged; } } static void get_stats(struct net_device *dev, struct ethtool_stats *stats, u64 *data) { struct port_info *pi = netdev_priv(dev); struct adapter *adapter = pi->adapter; u32 val1, val2; t4_get_port_stats(adapter, pi->tx_chan, (struct port_stats *)data); data += sizeof(struct port_stats) / sizeof(u64); collect_sge_port_stats(adapter, pi, (struct queue_port_stats *)data); data += sizeof(struct queue_port_stats) / sizeof(u64); if (!is_t4(adapter->params.chip)) { t4_write_reg(adapter, SGE_STAT_CFG, STATSOURCE_T5(7)); val1 = t4_read_reg(adapter, SGE_STAT_TOTAL); val2 = t4_read_reg(adapter, SGE_STAT_MATCH); *data = val1 - val2; data++; *data = val2; data++; } else { memset(data, 0, 2 * sizeof(u64)); *data += 2; } } /* * Return a version number to identify the type of adapter. The scheme is: * - bits 0..9: chip version * - bits 10..15: chip revision * - bits 16..23: register dump version */ static inline unsigned int mk_adap_vers(const struct adapter *ap) { return CHELSIO_CHIP_VERSION(ap->params.chip) | (CHELSIO_CHIP_RELEASE(ap->params.chip) << 10) | (1 << 16); } static void reg_block_dump(struct adapter *ap, void *buf, unsigned int start, unsigned int end) { u32 *p = buf + start; for ( ; start <= end; start += sizeof(u32)) *p++ = t4_read_reg(ap, start); } static void get_regs(struct net_device *dev, struct ethtool_regs *regs, void *buf) { static const unsigned int t4_reg_ranges[] = { 0x1008, 0x1108, 0x1180, 0x11b4, 0x11fc, 0x123c, 0x1300, 0x173c, 0x1800, 0x18fc, 0x3000, 0x30d8, 0x30e0, 0x5924, 0x5960, 0x59d4, 0x5a00, 0x5af8, 0x6000, 0x6098, 0x6100, 0x6150, 0x6200, 0x6208, 0x6240, 0x6248, 0x6280, 0x6338, 0x6370, 0x638c, 0x6400, 0x643c, 0x6500, 0x6524, 0x6a00, 0x6a38, 0x6a60, 0x6a78, 0x6b00, 0x6b84, 0x6bf0, 0x6c84, 0x6cf0, 0x6d84, 0x6df0, 0x6e84, 0x6ef0, 0x6f84, 0x6ff0, 0x7084, 0x70f0, 0x7184, 0x71f0, 0x7284, 0x72f0, 0x7384, 0x73f0, 0x7450, 0x7500, 0x7530, 0x7600, 0x761c, 0x7680, 0x76cc, 0x7700, 0x7798, 0x77c0, 0x77fc, 0x7900, 0x79fc, 0x7b00, 0x7c38, 0x7d00, 0x7efc, 0x8dc0, 0x8e1c, 0x8e30, 0x8e78, 0x8ea0, 0x8f6c, 0x8fc0, 0x9074, 0x90fc, 0x90fc, 0x9400, 0x9458, 0x9600, 0x96bc, 0x9800, 0x9808, 0x9820, 0x983c, 0x9850, 0x9864, 0x9c00, 0x9c6c, 0x9c80, 0x9cec, 0x9d00, 0x9d6c, 0x9d80, 0x9dec, 0x9e00, 0x9e6c, 0x9e80, 0x9eec, 0x9f00, 0x9f6c, 0x9f80, 0x9fec, 0xd004, 0xd03c, 0xdfc0, 0xdfe0, 0xe000, 0xea7c, 0xf000, 0x11110, 0x11118, 0x11190, 0x19040, 0x1906c, 0x19078, 0x19080, 0x1908c, 0x19124, 0x19150, 0x191b0, 0x191d0, 0x191e8, 0x19238, 0x1924c, 0x193f8, 0x19474, 0x19490, 0x194f8, 0x19800, 0x19f30, 0x1a000, 0x1a06c, 0x1a0b0, 0x1a120, 0x1a128, 0x1a138, 0x1a190, 0x1a1c4, 0x1a1fc, 0x1a1fc, 0x1e040, 0x1e04c, 0x1e284, 0x1e28c, 0x1e2c0, 0x1e2c0, 0x1e2e0, 0x1e2e0, 0x1e300, 0x1e384, 0x1e3c0, 0x1e3c8, 0x1e440, 0x1e44c, 0x1e684, 0x1e68c, 0x1e6c0, 0x1e6c0, 0x1e6e0, 0x1e6e0, 0x1e700, 0x1e784, 0x1e7c0, 0x1e7c8, 0x1e840, 0x1e84c, 0x1ea84, 0x1ea8c, 0x1eac0, 0x1eac0, 0x1eae0, 0x1eae0, 0x1eb00, 0x1eb84, 0x1ebc0, 0x1ebc8, 0x1ec40, 0x1ec4c, 0x1ee84, 0x1ee8c, 0x1eec0, 0x1eec0, 0x1eee0, 0x1eee0, 0x1ef00, 0x1ef84, 0x1efc0, 0x1efc8, 0x1f040, 0x1f04c, 0x1f284, 0x1f28c, 0x1f2c0, 0x1f2c0, 0x1f2e0, 0x1f2e0, 0x1f300, 0x1f384, 0x1f3c0, 0x1f3c8, 0x1f440, 0x1f44c, 0x1f684, 0x1f68c, 0x1f6c0, 0x1f6c0, 0x1f6e0, 0x1f6e0, 0x1f700, 0x1f784, 0x1f7c0, 0x1f7c8, 0x1f840, 0x1f84c, 0x1fa84, 0x1fa8c, 0x1fac0, 0x1fac0, 0x1fae0, 0x1fae0, 0x1fb00, 0x1fb84, 0x1fbc0, 0x1fbc8, 0x1fc40, 0x1fc4c, 0x1fe84, 0x1fe8c, 0x1fec0, 0x1fec0, 0x1fee0, 0x1fee0, 0x1ff00, 0x1ff84, 0x1ffc0, 0x1ffc8, 0x20000, 0x2002c, 0x20100, 0x2013c, 0x20190, 0x201c8, 0x20200, 0x20318, 0x20400, 0x20528, 0x20540, 0x20614, 0x21000, 0x21040, 0x2104c, 0x21060, 0x210c0, 0x210ec, 0x21200, 0x21268, 0x21270, 0x21284, 0x212fc, 0x21388, 0x21400, 0x21404, 0x21500, 0x21518, 0x2152c, 0x2153c, 0x21550, 0x21554, 0x21600, 0x21600, 0x21608, 0x21628, 0x21630, 0x2163c, 0x21700, 0x2171c, 0x21780, 0x2178c, 0x21800, 0x21c38, 0x21c80, 0x21d7c, 0x21e00, 0x21e04, 0x22000, 0x2202c, 0x22100, 0x2213c, 0x22190, 0x221c8, 0x22200, 0x22318, 0x22400, 0x22528, 0x22540, 0x22614, 0x23000, 0x23040, 0x2304c, 0x23060, 0x230c0, 0x230ec, 0x23200, 0x23268, 0x23270, 0x23284, 0x232fc, 0x23388, 0x23400, 0x23404, 0x23500, 0x23518, 0x2352c, 0x2353c, 0x23550, 0x23554, 0x23600, 0x23600, 0x23608, 0x23628, 0x23630, 0x2363c, 0x23700, 0x2371c, 0x23780, 0x2378c, 0x23800, 0x23c38, 0x23c80, 0x23d7c, 0x23e00, 0x23e04, 0x24000, 0x2402c, 0x24100, 0x2413c, 0x24190, 0x241c8, 0x24200, 0x24318, 0x24400, 0x24528, 0x24540, 0x24614, 0x25000, 0x25040, 0x2504c, 0x25060, 0x250c0, 0x250ec, 0x25200, 0x25268, 0x25270, 0x25284, 0x252fc, 0x25388, 0x25400, 0x25404, 0x25500, 0x25518, 0x2552c, 0x2553c, 0x25550, 0x25554, 0x25600, 0x25600, 0x25608, 0x25628, 0x25630, 0x2563c, 0x25700, 0x2571c, 0x25780, 0x2578c, 0x25800, 0x25c38, 0x25c80, 0x25d7c, 0x25e00, 0x25e04, 0x26000, 0x2602c, 0x26100, 0x2613c, 0x26190, 0x261c8, 0x26200, 0x26318, 0x26400, 0x26528, 0x26540, 0x26614, 0x27000, 0x27040, 0x2704c, 0x27060, 0x270c0, 0x270ec, 0x27200, 0x27268, 0x27270, 0x27284, 0x272fc, 0x27388, 0x27400, 0x27404, 0x27500, 0x27518, 0x2752c, 0x2753c, 0x27550, 0x27554, 0x27600, 0x27600, 0x27608, 0x27628, 0x27630, 0x2763c, 0x27700, 0x2771c, 0x27780, 0x2778c, 0x27800, 0x27c38, 0x27c80, 0x27d7c, 0x27e00, 0x27e04 }; static const unsigned int t5_reg_ranges[] = { 0x1008, 0x1148, 0x1180, 0x11b4, 0x11fc, 0x123c, 0x1280, 0x173c, 0x1800, 0x18fc, 0x3000, 0x3028, 0x3060, 0x30d8, 0x30e0, 0x30fc, 0x3140, 0x357c, 0x35a8, 0x35cc, 0x35ec, 0x35ec, 0x3600, 0x5624, 0x56cc, 0x575c, 0x580c, 0x5814, 0x5890, 0x58bc, 0x5940, 0x59dc, 0x59fc, 0x5a18, 0x5a60, 0x5a9c, 0x5b9c, 0x5bfc, 0x6000, 0x6040, 0x6058, 0x614c, 0x7700, 0x7798, 0x77c0, 0x78fc, 0x7b00, 0x7c54, 0x7d00, 0x7efc, 0x8dc0, 0x8de0, 0x8df8, 0x8e84, 0x8ea0, 0x8f84, 0x8fc0, 0x90f8, 0x9400, 0x9470, 0x9600, 0x96f4, 0x9800, 0x9808, 0x9820, 0x983c, 0x9850, 0x9864, 0x9c00, 0x9c6c, 0x9c80, 0x9cec, 0x9d00, 0x9d6c, 0x9d80, 0x9dec, 0x9e00, 0x9e6c, 0x9e80, 0x9eec, 0x9f00, 0x9f6c, 0x9f80, 0xa020, 0xd004, 0xd03c, 0xdfc0, 0xdfe0, 0xe000, 0x11088, 0x1109c, 0x11110, 0x11118, 0x1117c, 0x11190, 0x11204, 0x19040, 0x1906c, 0x19078, 0x19080, 0x1908c, 0x19124, 0x19150, 0x191b0, 0x191d0, 0x191e8, 0x19238, 0x19290, 0x193f8, 0x19474, 0x19490, 0x194cc, 0x194f0, 0x194f8, 0x19c00, 0x19c60, 0x19c94, 0x19e10, 0x19e50, 0x19f34, 0x19f40, 0x19f50, 0x19f90, 0x19fe4, 0x1a000, 0x1a06c, 0x1a0b0, 0x1a120, 0x1a128, 0x1a138, 0x1a190, 0x1a1c4, 0x1a1fc, 0x1a1fc, 0x1e008, 0x1e00c, 0x1e040, 0x1e04c, 0x1e284, 0x1e290, 0x1e2c0, 0x1e2c0, 0x1e2e0, 0x1e2e0, 0x1e300, 0x1e384, 0x1e3c0, 0x1e3c8, 0x1e408, 0x1e40c, 0x1e440, 0x1e44c, 0x1e684, 0x1e690, 0x1e6c0, 0x1e6c0, 0x1e6e0, 0x1e6e0, 0x1e700, 0x1e784, 0x1e7c0, 0x1e7c8, 0x1e808, 0x1e80c, 0x1e840, 0x1e84c, 0x1ea84, 0x1ea90, 0x1eac0, 0x1eac0, 0x1eae0, 0x1eae0, 0x1eb00, 0x1eb84, 0x1ebc0, 0x1ebc8, 0x1ec08, 0x1ec0c, 0x1ec40, 0x1ec4c, 0x1ee84, 0x1ee90, 0x1eec0, 0x1eec0, 0x1eee0, 0x1eee0, 0x1ef00, 0x1ef84, 0x1efc0, 0x1efc8, 0x1f008, 0x1f00c, 0x1f040, 0x1f04c, 0x1f284, 0x1f290, 0x1f2c0, 0x1f2c0, 0x1f2e0, 0x1f2e0, 0x1f300, 0x1f384, 0x1f3c0, 0x1f3c8, 0x1f408, 0x1f40c, 0x1f440, 0x1f44c, 0x1f684, 0x1f690, 0x1f6c0, 0x1f6c0, 0x1f6e0, 0x1f6e0, 0x1f700, 0x1f784, 0x1f7c0, 0x1f7c8, 0x1f808, 0x1f80c, 0x1f840, 0x1f84c, 0x1fa84, 0x1fa90, 0x1fac0, 0x1fac0, 0x1fae0, 0x1fae0, 0x1fb00, 0x1fb84, 0x1fbc0, 0x1fbc8, 0x1fc08, 0x1fc0c, 0x1fc40, 0x1fc4c, 0x1fe84, 0x1fe90, 0x1fec0, 0x1fec0, 0x1fee0, 0x1fee0, 0x1ff00, 0x1ff84, 0x1ffc0, 0x1ffc8, 0x30000, 0x30030, 0x30100, 0x30144, 0x30190, 0x301d0, 0x30200, 0x30318, 0x30400, 0x3052c, 0x30540, 0x3061c, 0x30800, 0x30834, 0x308c0, 0x30908, 0x30910, 0x309ac, 0x30a00, 0x30a04, 0x30a0c, 0x30a2c, 0x30a44, 0x30a50, 0x30a74, 0x30c24, 0x30d08, 0x30d14, 0x30d1c, 0x30d20, 0x30d3c, 0x30d50, 0x31200, 0x3120c, 0x31220, 0x31220, 0x31240, 0x31240, 0x31600, 0x31600, 0x31608, 0x3160c, 0x31a00, 0x31a1c, 0x31e04, 0x31e20, 0x31e38, 0x31e3c, 0x31e80, 0x31e80, 0x31e88, 0x31ea8, 0x31eb0, 0x31eb4, 0x31ec8, 0x31ed4, 0x31fb8, 0x32004, 0x32208, 0x3223c, 0x32600, 0x32630, 0x32a00, 0x32abc, 0x32b00, 0x32b70, 0x33000, 0x33048, 0x33060, 0x3309c, 0x330f0, 0x33148, 0x33160, 0x3319c, 0x331f0, 0x332e4, 0x332f8, 0x333e4, 0x333f8, 0x33448, 0x33460, 0x3349c, 0x334f0, 0x33548, 0x33560, 0x3359c, 0x335f0, 0x336e4, 0x336f8, 0x337e4, 0x337f8, 0x337fc, 0x33814, 0x33814, 0x3382c, 0x3382c, 0x33880, 0x3388c, 0x338e8, 0x338ec, 0x33900, 0x33948, 0x33960, 0x3399c, 0x339f0, 0x33ae4, 0x33af8, 0x33b10, 0x33b28, 0x33b28, 0x33b3c, 0x33b50, 0x33bf0, 0x33c10, 0x33c28, 0x33c28, 0x33c3c, 0x33c50, 0x33cf0, 0x33cfc, 0x34000, 0x34030, 0x34100, 0x34144, 0x34190, 0x341d0, 0x34200, 0x34318, 0x34400, 0x3452c, 0x34540, 0x3461c, 0x34800, 0x34834, 0x348c0, 0x34908, 0x34910, 0x349ac, 0x34a00, 0x34a04, 0x34a0c, 0x34a2c, 0x34a44, 0x34a50, 0x34a74, 0x34c24, 0x34d08, 0x34d14, 0x34d1c, 0x34d20, 0x34d3c, 0x34d50, 0x35200, 0x3520c, 0x35220, 0x35220, 0x35240, 0x35240, 0x35600, 0x35600, 0x35608, 0x3560c, 0x35a00, 0x35a1c, 0x35e04, 0x35e20, 0x35e38, 0x35e3c, 0x35e80, 0x35e80, 0x35e88, 0x35ea8, 0x35eb0, 0x35eb4, 0x35ec8, 0x35ed4, 0x35fb8, 0x36004, 0x36208, 0x3623c, 0x36600, 0x36630, 0x36a00, 0x36abc, 0x36b00, 0x36b70, 0x37000, 0x37048, 0x37060, 0x3709c, 0x370f0, 0x37148, 0x37160, 0x3719c, 0x371f0, 0x372e4, 0x372f8, 0x373e4, 0x373f8, 0x37448, 0x37460, 0x3749c, 0x374f0, 0x37548, 0x37560, 0x3759c, 0x375f0, 0x376e4, 0x376f8, 0x377e4, 0x377f8, 0x377fc, 0x37814, 0x37814, 0x3782c, 0x3782c, 0x37880, 0x3788c, 0x378e8, 0x378ec, 0x37900, 0x37948, 0x37960, 0x3799c, 0x379f0, 0x37ae4, 0x37af8, 0x37b10, 0x37b28, 0x37b28, 0x37b3c, 0x37b50, 0x37bf0, 0x37c10, 0x37c28, 0x37c28, 0x37c3c, 0x37c50, 0x37cf0, 0x37cfc, 0x38000, 0x38030, 0x38100, 0x38144, 0x38190, 0x381d0, 0x38200, 0x38318, 0x38400, 0x3852c, 0x38540, 0x3861c, 0x38800, 0x38834, 0x388c0, 0x38908, 0x38910, 0x389ac, 0x38a00, 0x38a04, 0x38a0c, 0x38a2c, 0x38a44, 0x38a50, 0x38a74, 0x38c24, 0x38d08, 0x38d14, 0x38d1c, 0x38d20, 0x38d3c, 0x38d50, 0x39200, 0x3920c, 0x39220, 0x39220, 0x39240, 0x39240, 0x39600, 0x39600, 0x39608, 0x3960c, 0x39a00, 0x39a1c, 0x39e04, 0x39e20, 0x39e38, 0x39e3c, 0x39e80, 0x39e80, 0x39e88, 0x39ea8, 0x39eb0, 0x39eb4, 0x39ec8, 0x39ed4, 0x39fb8, 0x3a004, 0x3a208, 0x3a23c, 0x3a600, 0x3a630, 0x3aa00, 0x3aabc, 0x3ab00, 0x3ab70, 0x3b000, 0x3b048, 0x3b060, 0x3b09c, 0x3b0f0, 0x3b148, 0x3b160, 0x3b19c, 0x3b1f0, 0x3b2e4, 0x3b2f8, 0x3b3e4, 0x3b3f8, 0x3b448, 0x3b460, 0x3b49c, 0x3b4f0, 0x3b548, 0x3b560, 0x3b59c, 0x3b5f0, 0x3b6e4, 0x3b6f8, 0x3b7e4, 0x3b7f8, 0x3b7fc, 0x3b814, 0x3b814, 0x3b82c, 0x3b82c, 0x3b880, 0x3b88c, 0x3b8e8, 0x3b8ec, 0x3b900, 0x3b948, 0x3b960, 0x3b99c, 0x3b9f0, 0x3bae4, 0x3baf8, 0x3bb10, 0x3bb28, 0x3bb28, 0x3bb3c, 0x3bb50, 0x3bbf0, 0x3bc10, 0x3bc28, 0x3bc28, 0x3bc3c, 0x3bc50, 0x3bcf0, 0x3bcfc, 0x3c000, 0x3c030, 0x3c100, 0x3c144, 0x3c190, 0x3c1d0, 0x3c200, 0x3c318, 0x3c400, 0x3c52c, 0x3c540, 0x3c61c, 0x3c800, 0x3c834, 0x3c8c0, 0x3c908, 0x3c910, 0x3c9ac, 0x3ca00, 0x3ca04, 0x3ca0c, 0x3ca2c, 0x3ca44, 0x3ca50, 0x3ca74, 0x3cc24, 0x3cd08, 0x3cd14, 0x3cd1c, 0x3cd20, 0x3cd3c, 0x3cd50, 0x3d200, 0x3d20c, 0x3d220, 0x3d220, 0x3d240, 0x3d240, 0x3d600, 0x3d600, 0x3d608, 0x3d60c, 0x3da00, 0x3da1c, 0x3de04, 0x3de20, 0x3de38, 0x3de3c, 0x3de80, 0x3de80, 0x3de88, 0x3dea8, 0x3deb0, 0x3deb4, 0x3dec8, 0x3ded4, 0x3dfb8, 0x3e004, 0x3e208, 0x3e23c, 0x3e600, 0x3e630, 0x3ea00, 0x3eabc, 0x3eb00, 0x3eb70, 0x3f000, 0x3f048, 0x3f060, 0x3f09c, 0x3f0f0, 0x3f148, 0x3f160, 0x3f19c, 0x3f1f0, 0x3f2e4, 0x3f2f8, 0x3f3e4, 0x3f3f8, 0x3f448, 0x3f460, 0x3f49c, 0x3f4f0, 0x3f548, 0x3f560, 0x3f59c, 0x3f5f0, 0x3f6e4, 0x3f6f8, 0x3f7e4, 0x3f7f8, 0x3f7fc, 0x3f814, 0x3f814, 0x3f82c, 0x3f82c, 0x3f880, 0x3f88c, 0x3f8e8, 0x3f8ec, 0x3f900, 0x3f948, 0x3f960, 0x3f99c, 0x3f9f0, 0x3fae4, 0x3faf8, 0x3fb10, 0x3fb28, 0x3fb28, 0x3fb3c, 0x3fb50, 0x3fbf0, 0x3fc10, 0x3fc28, 0x3fc28, 0x3fc3c, 0x3fc50, 0x3fcf0, 0x3fcfc, 0x40000, 0x4000c, 0x40040, 0x40068, 0x40080, 0x40144, 0x40180, 0x4018c, 0x40200, 0x40298, 0x402ac, 0x4033c, 0x403f8, 0x403fc, 0x41304, 0x413c4, 0x41400, 0x4141c, 0x41480, 0x414d0, 0x44000, 0x44078, 0x440c0, 0x44278, 0x442c0, 0x44478, 0x444c0, 0x44678, 0x446c0, 0x44878, 0x448c0, 0x449fc, 0x45000, 0x45068, 0x45080, 0x45084, 0x450a0, 0x450b0, 0x45200, 0x45268, 0x45280, 0x45284, 0x452a0, 0x452b0, 0x460c0, 0x460e4, 0x47000, 0x4708c, 0x47200, 0x47250, 0x47400, 0x47420, 0x47600, 0x47618, 0x47800, 0x47814, 0x48000, 0x4800c, 0x48040, 0x48068, 0x48080, 0x48144, 0x48180, 0x4818c, 0x48200, 0x48298, 0x482ac, 0x4833c, 0x483f8, 0x483fc, 0x49304, 0x493c4, 0x49400, 0x4941c, 0x49480, 0x494d0, 0x4c000, 0x4c078, 0x4c0c0, 0x4c278, 0x4c2c0, 0x4c478, 0x4c4c0, 0x4c678, 0x4c6c0, 0x4c878, 0x4c8c0, 0x4c9fc, 0x4d000, 0x4d068, 0x4d080, 0x4d084, 0x4d0a0, 0x4d0b0, 0x4d200, 0x4d268, 0x4d280, 0x4d284, 0x4d2a0, 0x4d2b0, 0x4e0c0, 0x4e0e4, 0x4f000, 0x4f08c, 0x4f200, 0x4f250, 0x4f400, 0x4f420, 0x4f600, 0x4f618, 0x4f800, 0x4f814, 0x50000, 0x500cc, 0x50400, 0x50400, 0x50800, 0x508cc, 0x50c00, 0x50c00, 0x51000, 0x5101c, 0x51300, 0x51308, }; int i; struct adapter *ap = netdev2adap(dev); static const unsigned int *reg_ranges; int arr_size = 0, buf_size = 0; if (is_t4(ap->params.chip)) { reg_ranges = &t4_reg_ranges[0]; arr_size = ARRAY_SIZE(t4_reg_ranges); buf_size = T4_REGMAP_SIZE; } else { reg_ranges = &t5_reg_ranges[0]; arr_size = ARRAY_SIZE(t5_reg_ranges); buf_size = T5_REGMAP_SIZE; } regs->version = mk_adap_vers(ap); memset(buf, 0, buf_size); for (i = 0; i < arr_size; i += 2) reg_block_dump(ap, buf, reg_ranges[i], reg_ranges[i + 1]); } static int restart_autoneg(struct net_device *dev) { struct port_info *p = netdev_priv(dev); if (!netif_running(dev)) return -EAGAIN; if (p->link_cfg.autoneg != AUTONEG_ENABLE) return -EINVAL; t4_restart_aneg(p->adapter, p->adapter->fn, p->tx_chan); return 0; } static int identify_port(struct net_device *dev, enum ethtool_phys_id_state state) { unsigned int val; struct adapter *adap = netdev2adap(dev); if (state == ETHTOOL_ID_ACTIVE) val = 0xffff; else if (state == ETHTOOL_ID_INACTIVE) val = 0; else return -EINVAL; return t4_identify_port(adap, adap->fn, netdev2pinfo(dev)->viid, val); } static unsigned int from_fw_linkcaps(unsigned int type, unsigned int caps) { unsigned int v = 0; if (type == FW_PORT_TYPE_BT_SGMII || type == FW_PORT_TYPE_BT_XFI || type == FW_PORT_TYPE_BT_XAUI) { v |= SUPPORTED_TP; if (caps & FW_PORT_CAP_SPEED_100M) v |= SUPPORTED_100baseT_Full; if (caps & FW_PORT_CAP_SPEED_1G) v |= SUPPORTED_1000baseT_Full; if (caps & FW_PORT_CAP_SPEED_10G) v |= SUPPORTED_10000baseT_Full; } else if (type == FW_PORT_TYPE_KX4 || type == FW_PORT_TYPE_KX) { v |= SUPPORTED_Backplane; if (caps & FW_PORT_CAP_SPEED_1G) v |= SUPPORTED_1000baseKX_Full; if (caps & FW_PORT_CAP_SPEED_10G) v |= SUPPORTED_10000baseKX4_Full; } else if (type == FW_PORT_TYPE_KR) v |= SUPPORTED_Backplane | SUPPORTED_10000baseKR_Full; else if (type == FW_PORT_TYPE_BP_AP) v |= SUPPORTED_Backplane | SUPPORTED_10000baseR_FEC | SUPPORTED_10000baseKR_Full | SUPPORTED_1000baseKX_Full; else if (type == FW_PORT_TYPE_BP4_AP) v |= SUPPORTED_Backplane | SUPPORTED_10000baseR_FEC | SUPPORTED_10000baseKR_Full | SUPPORTED_1000baseKX_Full | SUPPORTED_10000baseKX4_Full; else if (type == FW_PORT_TYPE_FIBER_XFI || type == FW_PORT_TYPE_FIBER_XAUI || type == FW_PORT_TYPE_SFP) v |= SUPPORTED_FIBRE; else if (type == FW_PORT_TYPE_BP40_BA) v |= SUPPORTED_40000baseSR4_Full; if (caps & FW_PORT_CAP_ANEG) v |= SUPPORTED_Autoneg; return v; } static unsigned int to_fw_linkcaps(unsigned int caps) { unsigned int v = 0; if (caps & ADVERTISED_100baseT_Full) v |= FW_PORT_CAP_SPEED_100M; if (caps & ADVERTISED_1000baseT_Full) v |= FW_PORT_CAP_SPEED_1G; if (caps & ADVERTISED_10000baseT_Full) v |= FW_PORT_CAP_SPEED_10G; if (caps & ADVERTISED_40000baseSR4_Full) v |= FW_PORT_CAP_SPEED_40G; return v; } static int get_settings(struct net_device *dev, struct ethtool_cmd *cmd) { const struct port_info *p = netdev_priv(dev); if (p->port_type == FW_PORT_TYPE_BT_SGMII || p->port_type == FW_PORT_TYPE_BT_XFI || p->port_type == FW_PORT_TYPE_BT_XAUI) cmd->port = PORT_TP; else if (p->port_type == FW_PORT_TYPE_FIBER_XFI || p->port_type == FW_PORT_TYPE_FIBER_XAUI) cmd->port = PORT_FIBRE; else if (p->port_type == FW_PORT_TYPE_SFP || p->port_type == FW_PORT_TYPE_QSFP_10G || p->port_type == FW_PORT_TYPE_QSFP) { if (p->mod_type == FW_PORT_MOD_TYPE_LR || p->mod_type == FW_PORT_MOD_TYPE_SR || p->mod_type == FW_PORT_MOD_TYPE_ER || p->mod_type == FW_PORT_MOD_TYPE_LRM) cmd->port = PORT_FIBRE; else if (p->mod_type == FW_PORT_MOD_TYPE_TWINAX_PASSIVE || p->mod_type == FW_PORT_MOD_TYPE_TWINAX_ACTIVE) cmd->port = PORT_DA; else cmd->port = PORT_OTHER; } else cmd->port = PORT_OTHER; if (p->mdio_addr >= 0) { cmd->phy_address = p->mdio_addr; cmd->transceiver = XCVR_EXTERNAL; cmd->mdio_support = p->port_type == FW_PORT_TYPE_BT_SGMII ? MDIO_SUPPORTS_C22 : MDIO_SUPPORTS_C45; } else { cmd->phy_address = 0; /* not really, but no better option */ cmd->transceiver = XCVR_INTERNAL; cmd->mdio_support = 0; } cmd->supported = from_fw_linkcaps(p->port_type, p->link_cfg.supported); cmd->advertising = from_fw_linkcaps(p->port_type, p->link_cfg.advertising); ethtool_cmd_speed_set(cmd, netif_carrier_ok(dev) ? p->link_cfg.speed : 0); cmd->duplex = DUPLEX_FULL; cmd->autoneg = p->link_cfg.autoneg; cmd->maxtxpkt = 0; cmd->maxrxpkt = 0; return 0; } static unsigned int speed_to_caps(int speed) { if (speed == 100) return FW_PORT_CAP_SPEED_100M; if (speed == 1000) return FW_PORT_CAP_SPEED_1G; if (speed == 10000) return FW_PORT_CAP_SPEED_10G; if (speed == 40000) return FW_PORT_CAP_SPEED_40G; return 0; } static int set_settings(struct net_device *dev, struct ethtool_cmd *cmd) { unsigned int cap; struct port_info *p = netdev_priv(dev); struct link_config *lc = &p->link_cfg; u32 speed = ethtool_cmd_speed(cmd); if (cmd->duplex != DUPLEX_FULL) /* only full-duplex supported */ return -EINVAL; if (!(lc->supported & FW_PORT_CAP_ANEG)) { /* * PHY offers a single speed. See if that's what's * being requested. */ if (cmd->autoneg == AUTONEG_DISABLE && (lc->supported & speed_to_caps(speed))) return 0; return -EINVAL; } if (cmd->autoneg == AUTONEG_DISABLE) { cap = speed_to_caps(speed); if (!(lc->supported & cap) || (speed == 1000) || (speed == 10000) || (speed == 40000)) return -EINVAL; lc->requested_speed = cap; lc->advertising = 0; } else { cap = to_fw_linkcaps(cmd->advertising); if (!(lc->supported & cap)) return -EINVAL; lc->requested_speed = 0; lc->advertising = cap | FW_PORT_CAP_ANEG; } lc->autoneg = cmd->autoneg; if (netif_running(dev)) return t4_link_start(p->adapter, p->adapter->fn, p->tx_chan, lc); return 0; } static void get_pauseparam(struct net_device *dev, struct ethtool_pauseparam *epause) { struct port_info *p = netdev_priv(dev); epause->autoneg = (p->link_cfg.requested_fc & PAUSE_AUTONEG) != 0; epause->rx_pause = (p->link_cfg.fc & PAUSE_RX) != 0; epause->tx_pause = (p->link_cfg.fc & PAUSE_TX) != 0; } static int set_pauseparam(struct net_device *dev, struct ethtool_pauseparam *epause) { struct port_info *p = netdev_priv(dev); struct link_config *lc = &p->link_cfg; if (epause->autoneg == AUTONEG_DISABLE) lc->requested_fc = 0; else if (lc->supported & FW_PORT_CAP_ANEG) lc->requested_fc = PAUSE_AUTONEG; else return -EINVAL; if (epause->rx_pause) lc->requested_fc |= PAUSE_RX; if (epause->tx_pause) lc->requested_fc |= PAUSE_TX; if (netif_running(dev)) return t4_link_start(p->adapter, p->adapter->fn, p->tx_chan, lc); return 0; } static void get_sge_param(struct net_device *dev, struct ethtool_ringparam *e) { const struct port_info *pi = netdev_priv(dev); const struct sge *s = &pi->adapter->sge; e->rx_max_pending = MAX_RX_BUFFERS; e->rx_mini_max_pending = MAX_RSPQ_ENTRIES; e->rx_jumbo_max_pending = 0; e->tx_max_pending = MAX_TXQ_ENTRIES; e->rx_pending = s->ethrxq[pi->first_qset].fl.size - 8; e->rx_mini_pending = s->ethrxq[pi->first_qset].rspq.size; e->rx_jumbo_pending = 0; e->tx_pending = s->ethtxq[pi->first_qset].q.size; } static int set_sge_param(struct net_device *dev, struct ethtool_ringparam *e) { int i; const struct port_info *pi = netdev_priv(dev); struct adapter *adapter = pi->adapter; struct sge *s = &adapter->sge; if (e->rx_pending > MAX_RX_BUFFERS || e->rx_jumbo_pending || e->tx_pending > MAX_TXQ_ENTRIES || e->rx_mini_pending > MAX_RSPQ_ENTRIES || e->rx_mini_pending < MIN_RSPQ_ENTRIES || e->rx_pending < MIN_FL_ENTRIES || e->tx_pending < MIN_TXQ_ENTRIES) return -EINVAL; if (adapter->flags & FULL_INIT_DONE) return -EBUSY; for (i = 0; i < pi->nqsets; ++i) { s->ethtxq[pi->first_qset + i].q.size = e->tx_pending; s->ethrxq[pi->first_qset + i].fl.size = e->rx_pending + 8; s->ethrxq[pi->first_qset + i].rspq.size = e->rx_mini_pending; } return 0; } static int closest_timer(const struct sge *s, int time) { int i, delta, match = 0, min_delta = INT_MAX; for (i = 0; i < ARRAY_SIZE(s->timer_val); i++) { delta = time - s->timer_val[i]; if (delta < 0) delta = -delta; if (delta < min_delta) { min_delta = delta; match = i; } } return match; } static int closest_thres(const struct sge *s, int thres) { int i, delta, match = 0, min_delta = INT_MAX; for (i = 0; i < ARRAY_SIZE(s->counter_val); i++) { delta = thres - s->counter_val[i]; if (delta < 0) delta = -delta; if (delta < min_delta) { min_delta = delta; match = i; } } return match; } /* * Return a queue's interrupt hold-off time in us. 0 means no timer. */ static unsigned int qtimer_val(const struct adapter *adap, const struct sge_rspq *q) { unsigned int idx = q->intr_params >> 1; return idx < SGE_NTIMERS ? adap->sge.timer_val[idx] : 0; } /** * set_rspq_intr_params - set a queue's interrupt holdoff parameters * @q: the Rx queue * @us: the hold-off time in us, or 0 to disable timer * @cnt: the hold-off packet count, or 0 to disable counter * * Sets an Rx queue's interrupt hold-off time and packet count. At least * one of the two needs to be enabled for the queue to generate interrupts. */ static int set_rspq_intr_params(struct sge_rspq *q, unsigned int us, unsigned int cnt) { struct adapter *adap = q->adap; if ((us | cnt) == 0) cnt = 1; if (cnt) { int err; u32 v, new_idx; new_idx = closest_thres(&adap->sge, cnt); if (q->desc && q->pktcnt_idx != new_idx) { /* the queue has already been created, update it */ v = FW_PARAMS_MNEM(FW_PARAMS_MNEM_DMAQ) | FW_PARAMS_PARAM_X(FW_PARAMS_PARAM_DMAQ_IQ_INTCNTTHRESH) | FW_PARAMS_PARAM_YZ(q->cntxt_id); err = t4_set_params(adap, adap->fn, adap->fn, 0, 1, &v, &new_idx); if (err) return err; } q->pktcnt_idx = new_idx; } us = us == 0 ? 6 : closest_timer(&adap->sge, us); q->intr_params = QINTR_TIMER_IDX(us) | (cnt > 0 ? QINTR_CNT_EN : 0); return 0; } /** * set_rx_intr_params - set a net devices's RX interrupt holdoff paramete! * @dev: the network device * @us: the hold-off time in us, or 0 to disable timer * @cnt: the hold-off packet count, or 0 to disable counter * * Set the RX interrupt hold-off parameters for a network device. */ static int set_rx_intr_params(struct net_device *dev, unsigned int us, unsigned int cnt) { int i, err; struct port_info *pi = netdev_priv(dev); struct adapter *adap = pi->adapter; struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset]; for (i = 0; i < pi->nqsets; i++, q++) { err = set_rspq_intr_params(&q->rspq, us, cnt); if (err) return err; } return 0; } static int set_coalesce(struct net_device *dev, struct ethtool_coalesce *c) { return set_rx_intr_params(dev, c->rx_coalesce_usecs, c->rx_max_coalesced_frames); } static int get_coalesce(struct net_device *dev, struct ethtool_coalesce *c) { const struct port_info *pi = netdev_priv(dev); const struct adapter *adap = pi->adapter; const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq; c->rx_coalesce_usecs = qtimer_val(adap, rq); c->rx_max_coalesced_frames = (rq->intr_params & QINTR_CNT_EN) ? adap->sge.counter_val[rq->pktcnt_idx] : 0; return 0; } /** * eeprom_ptov - translate a physical EEPROM address to virtual * @phys_addr: the physical EEPROM address * @fn: the PCI function number * @sz: size of function-specific area * * Translate a physical EEPROM address to virtual. The first 1K is * accessed through virtual addresses starting at 31K, the rest is * accessed through virtual addresses starting at 0. * * The mapping is as follows: * [0..1K) -> [31K..32K) * [1K..1K+A) -> [31K-A..31K) * [1K+A..ES) -> [0..ES-A-1K) * * where A = @fn * @sz, and ES = EEPROM size. */ static int eeprom_ptov(unsigned int phys_addr, unsigned int fn, unsigned int sz) { fn *= sz; if (phys_addr < 1024) return phys_addr + (31 << 10); if (phys_addr < 1024 + fn) return 31744 - fn + phys_addr - 1024; if (phys_addr < EEPROMSIZE) return phys_addr - 1024 - fn; return -EINVAL; } /* * The next two routines implement eeprom read/write from physical addresses. */ static int eeprom_rd_phys(struct adapter *adap, unsigned int phys_addr, u32 *v) { int vaddr = eeprom_ptov(phys_addr, adap->fn, EEPROMPFSIZE); if (vaddr >= 0) vaddr = pci_read_vpd(adap->pdev, vaddr, sizeof(u32), v); return vaddr < 0 ? vaddr : 0; } static int eeprom_wr_phys(struct adapter *adap, unsigned int phys_addr, u32 v) { int vaddr = eeprom_ptov(phys_addr, adap->fn, EEPROMPFSIZE); if (vaddr >= 0) vaddr = pci_write_vpd(adap->pdev, vaddr, sizeof(u32), &v); return vaddr < 0 ? vaddr : 0; } #define EEPROM_MAGIC 0x38E2F10C static int get_eeprom(struct net_device *dev, struct ethtool_eeprom *e, u8 *data) { int i, err = 0; struct adapter *adapter = netdev2adap(dev); u8 *buf = kmalloc(EEPROMSIZE, GFP_KERNEL); if (!buf) return -ENOMEM; e->magic = EEPROM_MAGIC; for (i = e->offset & ~3; !err && i < e->offset + e->len; i += 4) err = eeprom_rd_phys(adapter, i, (u32 *)&buf[i]); if (!err) memcpy(data, buf + e->offset, e->len); kfree(buf); return err; } static int set_eeprom(struct net_device *dev, struct ethtool_eeprom *eeprom, u8 *data) { u8 *buf; int err = 0; u32 aligned_offset, aligned_len, *p; struct adapter *adapter = netdev2adap(dev); if (eeprom->magic != EEPROM_MAGIC) return -EINVAL; aligned_offset = eeprom->offset & ~3; aligned_len = (eeprom->len + (eeprom->offset & 3) + 3) & ~3; if (adapter->fn > 0) { u32 start = 1024 + adapter->fn * EEPROMPFSIZE; if (aligned_offset < start || aligned_offset + aligned_len > start + EEPROMPFSIZE) return -EPERM; } if (aligned_offset != eeprom->offset || aligned_len != eeprom->len) { /* * RMW possibly needed for first or last words. */ buf = kmalloc(aligned_len, GFP_KERNEL); if (!buf) return -ENOMEM; err = eeprom_rd_phys(adapter, aligned_offset, (u32 *)buf); if (!err && aligned_len > 4) err = eeprom_rd_phys(adapter, aligned_offset + aligned_len - 4, (u32 *)&buf[aligned_len - 4]); if (err) goto out; memcpy(buf + (eeprom->offset & 3), data, eeprom->len); } else buf = data; err = t4_seeprom_wp(adapter, false); if (err) goto out; for (p = (u32 *)buf; !err && aligned_len; aligned_len -= 4, p++) { err = eeprom_wr_phys(adapter, aligned_offset, *p); aligned_offset += 4; } if (!err) err = t4_seeprom_wp(adapter, true); out: if (buf != data) kfree(buf); return err; } static int set_flash(struct net_device *netdev, struct ethtool_flash *ef) { int ret; const struct firmware *fw; struct adapter *adap = netdev2adap(netdev); ef->data[sizeof(ef->data) - 1] = '\0'; ret = request_firmware(&fw, ef->data, adap->pdev_dev); if (ret < 0) return ret; ret = t4_load_fw(adap, fw->data, fw->size); release_firmware(fw); if (!ret) dev_info(adap->pdev_dev, "loaded firmware %s\n", ef->data); return ret; } #define WOL_SUPPORTED (WAKE_BCAST | WAKE_MAGIC) #define BCAST_CRC 0xa0ccc1a6 static void get_wol(struct net_device *dev, struct ethtool_wolinfo *wol) { wol->supported = WAKE_BCAST | WAKE_MAGIC; wol->wolopts = netdev2adap(dev)->wol; memset(&wol->sopass, 0, sizeof(wol->sopass)); } static int set_wol(struct net_device *dev, struct ethtool_wolinfo *wol) { int err = 0; struct port_info *pi = netdev_priv(dev); if (wol->wolopts & ~WOL_SUPPORTED) return -EINVAL; t4_wol_magic_enable(pi->adapter, pi->tx_chan, (wol->wolopts & WAKE_MAGIC) ? dev->dev_addr : NULL); if (wol->wolopts & WAKE_BCAST) { err = t4_wol_pat_enable(pi->adapter, pi->tx_chan, 0xfe, ~0ULL, ~0ULL, 0, false); if (!err) err = t4_wol_pat_enable(pi->adapter, pi->tx_chan, 1, ~6ULL, ~0ULL, BCAST_CRC, true); } else t4_wol_pat_enable(pi->adapter, pi->tx_chan, 0, 0, 0, 0, false); return err; } static int cxgb_set_features(struct net_device *dev, netdev_features_t features) { const struct port_info *pi = netdev_priv(dev); netdev_features_t changed = dev->features ^ features; int err; if (!(changed & NETIF_F_HW_VLAN_CTAG_RX)) return 0; err = t4_set_rxmode(pi->adapter, pi->adapter->fn, pi->viid, -1, -1, -1, -1, !!(features & NETIF_F_HW_VLAN_CTAG_RX), true); if (unlikely(err)) dev->features = features ^ NETIF_F_HW_VLAN_CTAG_RX; return err; } static u32 get_rss_table_size(struct net_device *dev) { const struct port_info *pi = netdev_priv(dev); return pi->rss_size; } static int get_rss_table(struct net_device *dev, u32 *p, u8 *key) { const struct port_info *pi = netdev_priv(dev); unsigned int n = pi->rss_size; while (n--) p[n] = pi->rss[n]; return 0; } static int set_rss_table(struct net_device *dev, const u32 *p, const u8 *key) { unsigned int i; struct port_info *pi = netdev_priv(dev); for (i = 0; i < pi->rss_size; i++) pi->rss[i] = p[i]; if (pi->adapter->flags & FULL_INIT_DONE) return write_rss(pi, pi->rss); return 0; } static int get_rxnfc(struct net_device *dev, struct ethtool_rxnfc *info, u32 *rules) { const struct port_info *pi = netdev_priv(dev); switch (info->cmd) { case ETHTOOL_GRXFH: { unsigned int v = pi->rss_mode; info->data = 0; switch (info->flow_type) { case TCP_V4_FLOW: if (v & FW_RSS_VI_CONFIG_CMD_IP4FOURTUPEN) info->data = RXH_IP_SRC | RXH_IP_DST | RXH_L4_B_0_1 | RXH_L4_B_2_3; else if (v & FW_RSS_VI_CONFIG_CMD_IP4TWOTUPEN) info->data = RXH_IP_SRC | RXH_IP_DST; break; case UDP_V4_FLOW: if ((v & FW_RSS_VI_CONFIG_CMD_IP4FOURTUPEN) && (v & FW_RSS_VI_CONFIG_CMD_UDPEN)) info->data = RXH_IP_SRC | RXH_IP_DST | RXH_L4_B_0_1 | RXH_L4_B_2_3; else if (v & FW_RSS_VI_CONFIG_CMD_IP4TWOTUPEN) info->data = RXH_IP_SRC | RXH_IP_DST; break; case SCTP_V4_FLOW: case AH_ESP_V4_FLOW: case IPV4_FLOW: if (v & FW_RSS_VI_CONFIG_CMD_IP4TWOTUPEN) info->data = RXH_IP_SRC | RXH_IP_DST; break; case TCP_V6_FLOW: if (v & FW_RSS_VI_CONFIG_CMD_IP6FOURTUPEN) info->data = RXH_IP_SRC | RXH_IP_DST | RXH_L4_B_0_1 | RXH_L4_B_2_3; else if (v & FW_RSS_VI_CONFIG_CMD_IP6TWOTUPEN) info->data = RXH_IP_SRC | RXH_IP_DST; break; case UDP_V6_FLOW: if ((v & FW_RSS_VI_CONFIG_CMD_IP6FOURTUPEN) && (v & FW_RSS_VI_CONFIG_CMD_UDPEN)) info->data = RXH_IP_SRC | RXH_IP_DST | RXH_L4_B_0_1 | RXH_L4_B_2_3; else if (v & FW_RSS_VI_CONFIG_CMD_IP6TWOTUPEN) info->data = RXH_IP_SRC | RXH_IP_DST; break; case SCTP_V6_FLOW: case AH_ESP_V6_FLOW: case IPV6_FLOW: if (v & FW_RSS_VI_CONFIG_CMD_IP6TWOTUPEN) info->data = RXH_IP_SRC | RXH_IP_DST; break; } return 0; } case ETHTOOL_GRXRINGS: info->data = pi->nqsets; return 0; } return -EOPNOTSUPP; } static const struct ethtool_ops cxgb_ethtool_ops = { .get_settings = get_settings, .set_settings = set_settings, .get_drvinfo = get_drvinfo, .get_msglevel = get_msglevel, .set_msglevel = set_msglevel, .get_ringparam = get_sge_param, .set_ringparam = set_sge_param, .get_coalesce = get_coalesce, .set_coalesce = set_coalesce, .get_eeprom_len = get_eeprom_len, .get_eeprom = get_eeprom, .set_eeprom = set_eeprom, .get_pauseparam = get_pauseparam, .set_pauseparam = set_pauseparam, .get_link = ethtool_op_get_link, .get_strings = get_strings, .set_phys_id = identify_port, .nway_reset = restart_autoneg, .get_sset_count = get_sset_count, .get_ethtool_stats = get_stats, .get_regs_len = get_regs_len, .get_regs = get_regs, .get_wol = get_wol, .set_wol = set_wol, .get_rxnfc = get_rxnfc, .get_rxfh_indir_size = get_rss_table_size, .get_rxfh = get_rss_table, .set_rxfh = set_rss_table, .flash_device = set_flash, }; /* * debugfs support */ static ssize_t mem_read(struct file *file, char __user *buf, size_t count, loff_t *ppos) { loff_t pos = *ppos; loff_t avail = file_inode(file)->i_size; unsigned int mem = (uintptr_t)file->private_data & 3; struct adapter *adap = file->private_data - mem; __be32 *data; int ret; if (pos < 0) return -EINVAL; if (pos >= avail) return 0; if (count > avail - pos) count = avail - pos; data = t4_alloc_mem(count); if (!data) return -ENOMEM; spin_lock(&adap->win0_lock); ret = t4_memory_rw(adap, 0, mem, pos, count, data, T4_MEMORY_READ); spin_unlock(&adap->win0_lock); if (ret) { t4_free_mem(data); return ret; } ret = copy_to_user(buf, data, count); t4_free_mem(data); if (ret) return -EFAULT; *ppos = pos + count; return count; } static const struct file_operations mem_debugfs_fops = { .owner = THIS_MODULE, .open = simple_open, .read = mem_read, .llseek = default_llseek, }; static void add_debugfs_mem(struct adapter *adap, const char *name, unsigned int idx, unsigned int size_mb) { struct dentry *de; de = debugfs_create_file(name, S_IRUSR, adap->debugfs_root, (void *)adap + idx, &mem_debugfs_fops); if (de && de->d_inode) de->d_inode->i_size = size_mb << 20; } static int setup_debugfs(struct adapter *adap) { int i; u32 size; if (IS_ERR_OR_NULL(adap->debugfs_root)) return -1; i = t4_read_reg(adap, MA_TARGET_MEM_ENABLE); if (i & EDRAM0_ENABLE) { size = t4_read_reg(adap, MA_EDRAM0_BAR); add_debugfs_mem(adap, "edc0", MEM_EDC0, EDRAM_SIZE_GET(size)); } if (i & EDRAM1_ENABLE) { size = t4_read_reg(adap, MA_EDRAM1_BAR); add_debugfs_mem(adap, "edc1", MEM_EDC1, EDRAM_SIZE_GET(size)); } if (is_t4(adap->params.chip)) { size = t4_read_reg(adap, MA_EXT_MEMORY_BAR); if (i & EXT_MEM_ENABLE) add_debugfs_mem(adap, "mc", MEM_MC, EXT_MEM_SIZE_GET(size)); } else { if (i & EXT_MEM_ENABLE) { size = t4_read_reg(adap, MA_EXT_MEMORY_BAR); add_debugfs_mem(adap, "mc0", MEM_MC0, EXT_MEM_SIZE_GET(size)); } if (i & EXT_MEM1_ENABLE) { size = t4_read_reg(adap, MA_EXT_MEMORY1_BAR); add_debugfs_mem(adap, "mc1", MEM_MC1, EXT_MEM_SIZE_GET(size)); } } if (adap->l2t) debugfs_create_file("l2t", S_IRUSR, adap->debugfs_root, adap, &t4_l2t_fops); return 0; } /* * upper-layer driver support */ /* * Allocate an active-open TID and set it to the supplied value. */ int cxgb4_alloc_atid(struct tid_info *t, void *data) { int atid = -1; spin_lock_bh(&t->atid_lock); if (t->afree) { union aopen_entry *p = t->afree; atid = (p - t->atid_tab) + t->atid_base; t->afree = p->next; p->data = data; t->atids_in_use++; } spin_unlock_bh(&t->atid_lock); return atid; } EXPORT_SYMBOL(cxgb4_alloc_atid); /* * Release an active-open TID. */ void cxgb4_free_atid(struct tid_info *t, unsigned int atid) { union aopen_entry *p = &t->atid_tab[atid - t->atid_base]; spin_lock_bh(&t->atid_lock); p->next = t->afree; t->afree = p; t->atids_in_use--; spin_unlock_bh(&t->atid_lock); } EXPORT_SYMBOL(cxgb4_free_atid); /* * Allocate a server TID and set it to the supplied value. */ int cxgb4_alloc_stid(struct tid_info *t, int family, void *data) { int stid; spin_lock_bh(&t->stid_lock); if (family == PF_INET) { stid = find_first_zero_bit(t->stid_bmap, t->nstids); if (stid < t->nstids) __set_bit(stid, t->stid_bmap); else stid = -1; } else { stid = bitmap_find_free_region(t->stid_bmap, t->nstids, 2); if (stid < 0) stid = -1; } if (stid >= 0) { t->stid_tab[stid].data = data; stid += t->stid_base; /* IPv6 requires max of 520 bits or 16 cells in TCAM * This is equivalent to 4 TIDs. With CLIP enabled it * needs 2 TIDs. */ if (family == PF_INET) t->stids_in_use++; else t->stids_in_use += 4; } spin_unlock_bh(&t->stid_lock); return stid; } EXPORT_SYMBOL(cxgb4_alloc_stid); /* Allocate a server filter TID and set it to the supplied value. */ int cxgb4_alloc_sftid(struct tid_info *t, int family, void *data) { int stid; spin_lock_bh(&t->stid_lock); if (family == PF_INET) { stid = find_next_zero_bit(t->stid_bmap, t->nstids + t->nsftids, t->nstids); if (stid < (t->nstids + t->nsftids)) __set_bit(stid, t->stid_bmap); else stid = -1; } else { stid = -1; } if (stid >= 0) { t->stid_tab[stid].data = data; stid -= t->nstids; stid += t->sftid_base; t->stids_in_use++; } spin_unlock_bh(&t->stid_lock); return stid; } EXPORT_SYMBOL(cxgb4_alloc_sftid); /* Release a server TID. */ void cxgb4_free_stid(struct tid_info *t, unsigned int stid, int family) { /* Is it a server filter TID? */ if (t->nsftids && (stid >= t->sftid_base)) { stid -= t->sftid_base; stid += t->nstids; } else { stid -= t->stid_base; } spin_lock_bh(&t->stid_lock); if (family == PF_INET) __clear_bit(stid, t->stid_bmap); else bitmap_release_region(t->stid_bmap, stid, 2); t->stid_tab[stid].data = NULL; if (family == PF_INET) t->stids_in_use--; else t->stids_in_use -= 4; spin_unlock_bh(&t->stid_lock); } EXPORT_SYMBOL(cxgb4_free_stid); /* * Populate a TID_RELEASE WR. Caller must properly size the skb. */ static void mk_tid_release(struct sk_buff *skb, unsigned int chan, unsigned int tid) { struct cpl_tid_release *req; set_wr_txq(skb, CPL_PRIORITY_SETUP, chan); req = (struct cpl_tid_release *)__skb_put(skb, sizeof(*req)); INIT_TP_WR(req, tid); OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_TID_RELEASE, tid)); } /* * Queue a TID release request and if necessary schedule a work queue to * process it. */ static void cxgb4_queue_tid_release(struct tid_info *t, unsigned int chan, unsigned int tid) { void **p = &t->tid_tab[tid]; struct adapter *adap = container_of(t, struct adapter, tids); spin_lock_bh(&adap->tid_release_lock); *p = adap->tid_release_head; /* Low 2 bits encode the Tx channel number */ adap->tid_release_head = (void **)((uintptr_t)p | chan); if (!adap->tid_release_task_busy) { adap->tid_release_task_busy = true; queue_work(adap->workq, &adap->tid_release_task); } spin_unlock_bh(&adap->tid_release_lock); } /* * Process the list of pending TID release requests. */ static void process_tid_release_list(struct work_struct *work) { struct sk_buff *skb; struct adapter *adap; adap = container_of(work, struct adapter, tid_release_task); spin_lock_bh(&adap->tid_release_lock); while (adap->tid_release_head) { void **p = adap->tid_release_head; unsigned int chan = (uintptr_t)p & 3; p = (void *)p - chan; adap->tid_release_head = *p; *p = NULL; spin_unlock_bh(&adap->tid_release_lock); while (!(skb = alloc_skb(sizeof(struct cpl_tid_release), GFP_KERNEL))) schedule_timeout_uninterruptible(1); mk_tid_release(skb, chan, p - adap->tids.tid_tab); t4_ofld_send(adap, skb); spin_lock_bh(&adap->tid_release_lock); } adap->tid_release_task_busy = false; spin_unlock_bh(&adap->tid_release_lock); } /* * Release a TID and inform HW. If we are unable to allocate the release * message we defer to a work queue. */ void cxgb4_remove_tid(struct tid_info *t, unsigned int chan, unsigned int tid) { void *old; struct sk_buff *skb; struct adapter *adap = container_of(t, struct adapter, tids); old = t->tid_tab[tid]; skb = alloc_skb(sizeof(struct cpl_tid_release), GFP_ATOMIC); if (likely(skb)) { t->tid_tab[tid] = NULL; mk_tid_release(skb, chan, tid); t4_ofld_send(adap, skb); } else cxgb4_queue_tid_release(t, chan, tid); if (old) atomic_dec(&t->tids_in_use); } EXPORT_SYMBOL(cxgb4_remove_tid); /* * Allocate and initialize the TID tables. Returns 0 on success. */ static int tid_init(struct tid_info *t) { size_t size; unsigned int stid_bmap_size; unsigned int natids = t->natids; struct adapter *adap = container_of(t, struct adapter, tids); stid_bmap_size = BITS_TO_LONGS(t->nstids + t->nsftids); size = t->ntids * sizeof(*t->tid_tab) + natids * sizeof(*t->atid_tab) + t->nstids * sizeof(*t->stid_tab) + t->nsftids * sizeof(*t->stid_tab) + stid_bmap_size * sizeof(long) + t->nftids * sizeof(*t->ftid_tab) + t->nsftids * sizeof(*t->ftid_tab); t->tid_tab = t4_alloc_mem(size); if (!t->tid_tab) return -ENOMEM; t->atid_tab = (union aopen_entry *)&t->tid_tab[t->ntids]; t->stid_tab = (struct serv_entry *)&t->atid_tab[natids]; t->stid_bmap = (unsigned long *)&t->stid_tab[t->nstids + t->nsftids]; t->ftid_tab = (struct filter_entry *)&t->stid_bmap[stid_bmap_size]; spin_lock_init(&t->stid_lock); spin_lock_init(&t->atid_lock); t->stids_in_use = 0; t->afree = NULL; t->atids_in_use = 0; atomic_set(&t->tids_in_use, 0); /* Setup the free list for atid_tab and clear the stid bitmap. */ if (natids) { while (--natids) t->atid_tab[natids - 1].next = &t->atid_tab[natids]; t->afree = t->atid_tab; } bitmap_zero(t->stid_bmap, t->nstids + t->nsftids); /* Reserve stid 0 for T4/T5 adapters */ if (!t->stid_base && (is_t4(adap->params.chip) || is_t5(adap->params.chip))) __set_bit(0, t->stid_bmap); return 0; } int cxgb4_clip_get(const struct net_device *dev, const struct in6_addr *lip) { struct adapter *adap; struct fw_clip_cmd c; adap = netdev2adap(dev); memset(&c, 0, sizeof(c)); c.op_to_write = htonl(FW_CMD_OP(FW_CLIP_CMD) | FW_CMD_REQUEST | FW_CMD_WRITE); c.alloc_to_len16 = htonl(F_FW_CLIP_CMD_ALLOC | FW_LEN16(c)); c.ip_hi = *(__be64 *)(lip->s6_addr); c.ip_lo = *(__be64 *)(lip->s6_addr + 8); return t4_wr_mbox_meat(adap, adap->mbox, &c, sizeof(c), &c, false); } EXPORT_SYMBOL(cxgb4_clip_get); int cxgb4_clip_release(const struct net_device *dev, const struct in6_addr *lip) { struct adapter *adap; struct fw_clip_cmd c; adap = netdev2adap(dev); memset(&c, 0, sizeof(c)); c.op_to_write = htonl(FW_CMD_OP(FW_CLIP_CMD) | FW_CMD_REQUEST | FW_CMD_READ); c.alloc_to_len16 = htonl(F_FW_CLIP_CMD_FREE | FW_LEN16(c)); c.ip_hi = *(__be64 *)(lip->s6_addr); c.ip_lo = *(__be64 *)(lip->s6_addr + 8); return t4_wr_mbox_meat(adap, adap->mbox, &c, sizeof(c), &c, false); } EXPORT_SYMBOL(cxgb4_clip_release); /** * cxgb4_create_server - create an IP server * @dev: the device * @stid: the server TID * @sip: local IP address to bind server to * @sport: the server's TCP port * @queue: queue to direct messages from this server to * * Create an IP server for the given port and address. * Returns <0 on error and one of the %NET_XMIT_* values on success. */ int cxgb4_create_server(const struct net_device *dev, unsigned int stid, __be32 sip, __be16 sport, __be16 vlan, unsigned int queue) { unsigned int chan; struct sk_buff *skb; struct adapter *adap; struct cpl_pass_open_req *req; int ret; skb = alloc_skb(sizeof(*req), GFP_KERNEL); if (!skb) return -ENOMEM; adap = netdev2adap(dev); req = (struct cpl_pass_open_req *)__skb_put(skb, sizeof(*req)); INIT_TP_WR(req, 0); OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_PASS_OPEN_REQ, stid)); req->local_port = sport; req->peer_port = htons(0); req->local_ip = sip; req->peer_ip = htonl(0); chan = rxq_to_chan(&adap->sge, queue); req->opt0 = cpu_to_be64(TX_CHAN(chan)); req->opt1 = cpu_to_be64(CONN_POLICY_ASK | SYN_RSS_ENABLE | SYN_RSS_QUEUE(queue)); ret = t4_mgmt_tx(adap, skb); return net_xmit_eval(ret); } EXPORT_SYMBOL(cxgb4_create_server); /* cxgb4_create_server6 - create an IPv6 server * @dev: the device * @stid: the server TID * @sip: local IPv6 address to bind server to * @sport: the server's TCP port * @queue: queue to direct messages from this server to * * Create an IPv6 server for the given port and address. * Returns <0 on error and one of the %NET_XMIT_* values on success. */ int cxgb4_create_server6(const struct net_device *dev, unsigned int stid, const struct in6_addr *sip, __be16 sport, unsigned int queue) { unsigned int chan; struct sk_buff *skb; struct adapter *adap; struct cpl_pass_open_req6 *req; int ret; skb = alloc_skb(sizeof(*req), GFP_KERNEL); if (!skb) return -ENOMEM; adap = netdev2adap(dev); req = (struct cpl_pass_open_req6 *)__skb_put(skb, sizeof(*req)); INIT_TP_WR(req, 0); OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_PASS_OPEN_REQ6, stid)); req->local_port = sport; req->peer_port = htons(0); req->local_ip_hi = *(__be64 *)(sip->s6_addr); req->local_ip_lo = *(__be64 *)(sip->s6_addr + 8); req->peer_ip_hi = cpu_to_be64(0); req->peer_ip_lo = cpu_to_be64(0); chan = rxq_to_chan(&adap->sge, queue); req->opt0 = cpu_to_be64(TX_CHAN(chan)); req->opt1 = cpu_to_be64(CONN_POLICY_ASK | SYN_RSS_ENABLE | SYN_RSS_QUEUE(queue)); ret = t4_mgmt_tx(adap, skb); return net_xmit_eval(ret); } EXPORT_SYMBOL(cxgb4_create_server6); int cxgb4_remove_server(const struct net_device *dev, unsigned int stid, unsigned int queue, bool ipv6) { struct sk_buff *skb; struct adapter *adap; struct cpl_close_listsvr_req *req; int ret; adap = netdev2adap(dev); skb = alloc_skb(sizeof(*req), GFP_KERNEL); if (!skb) return -ENOMEM; req = (struct cpl_close_listsvr_req *)__skb_put(skb, sizeof(*req)); INIT_TP_WR(req, 0); OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_CLOSE_LISTSRV_REQ, stid)); req->reply_ctrl = htons(NO_REPLY(0) | (ipv6 ? LISTSVR_IPV6(1) : LISTSVR_IPV6(0)) | QUEUENO(queue)); ret = t4_mgmt_tx(adap, skb); return net_xmit_eval(ret); } EXPORT_SYMBOL(cxgb4_remove_server); /** * cxgb4_best_mtu - find the entry in the MTU table closest to an MTU * @mtus: the HW MTU table * @mtu: the target MTU * @idx: index of selected entry in the MTU table * * Returns the index and the value in the HW MTU table that is closest to * but does not exceed @mtu, unless @mtu is smaller than any value in the * table, in which case that smallest available value is selected. */ unsigned int cxgb4_best_mtu(const unsigned short *mtus, unsigned short mtu, unsigned int *idx) { unsigned int i = 0; while (i < NMTUS - 1 && mtus[i + 1] <= mtu) ++i; if (idx) *idx = i; return mtus[i]; } EXPORT_SYMBOL(cxgb4_best_mtu); /** * cxgb4_best_aligned_mtu - find best MTU, [hopefully] data size aligned * @mtus: the HW MTU table * @header_size: Header Size * @data_size_max: maximum Data Segment Size * @data_size_align: desired Data Segment Size Alignment (2^N) * @mtu_idxp: HW MTU Table Index return value pointer (possibly NULL) * * Similar to cxgb4_best_mtu() but instead of searching the Hardware * MTU Table based solely on a Maximum MTU parameter, we break that * parameter up into a Header Size and Maximum Data Segment Size, and * provide a desired Data Segment Size Alignment. If we find an MTU in * the Hardware MTU Table which will result in a Data Segment Size with * the requested alignment _and_ that MTU isn't "too far" from the * closest MTU, then we'll return that rather than the closest MTU. */ unsigned int cxgb4_best_aligned_mtu(const unsigned short *mtus, unsigned short header_size, unsigned short data_size_max, unsigned short data_size_align, unsigned int *mtu_idxp) { unsigned short max_mtu = header_size + data_size_max; unsigned short data_size_align_mask = data_size_align - 1; int mtu_idx, aligned_mtu_idx; /* Scan the MTU Table till we find an MTU which is larger than our * Maximum MTU or we reach the end of the table. Along the way, * record the last MTU found, if any, which will result in a Data * Segment Length matching the requested alignment. */ for (mtu_idx = 0, aligned_mtu_idx = -1; mtu_idx < NMTUS; mtu_idx++) { unsigned short data_size = mtus[mtu_idx] - header_size; /* If this MTU minus the Header Size would result in a * Data Segment Size of the desired alignment, remember it. */ if ((data_size & data_size_align_mask) == 0) aligned_mtu_idx = mtu_idx; /* If we're not at the end of the Hardware MTU Table and the * next element is larger than our Maximum MTU, drop out of * the loop. */ if (mtu_idx+1 < NMTUS && mtus[mtu_idx+1] > max_mtu) break; } /* If we fell out of the loop because we ran to the end of the table, * then we just have to use the last [largest] entry. */ if (mtu_idx == NMTUS) mtu_idx--; /* If we found an MTU which resulted in the requested Data Segment * Length alignment and that's "not far" from the largest MTU which is * less than or equal to the maximum MTU, then use that. */ if (aligned_mtu_idx >= 0 && mtu_idx - aligned_mtu_idx <= 1) mtu_idx = aligned_mtu_idx; /* If the caller has passed in an MTU Index pointer, pass the * MTU Index back. Return the MTU value. */ if (mtu_idxp) *mtu_idxp = mtu_idx; return mtus[mtu_idx]; } EXPORT_SYMBOL(cxgb4_best_aligned_mtu); /** * cxgb4_port_chan - get the HW channel of a port * @dev: the net device for the port * * Return the HW Tx channel of the given port. */ unsigned int cxgb4_port_chan(const struct net_device *dev) { return netdev2pinfo(dev)->tx_chan; } EXPORT_SYMBOL(cxgb4_port_chan); unsigned int cxgb4_dbfifo_count(const struct net_device *dev, int lpfifo) { struct adapter *adap = netdev2adap(dev); u32 v1, v2, lp_count, hp_count; v1 = t4_read_reg(adap, A_SGE_DBFIFO_STATUS); v2 = t4_read_reg(adap, SGE_DBFIFO_STATUS2); if (is_t4(adap->params.chip)) { lp_count = G_LP_COUNT(v1); hp_count = G_HP_COUNT(v1); } else { lp_count = G_LP_COUNT_T5(v1); hp_count = G_HP_COUNT_T5(v2); } return lpfifo ? lp_count : hp_count; } EXPORT_SYMBOL(cxgb4_dbfifo_count); /** * cxgb4_port_viid - get the VI id of a port * @dev: the net device for the port * * Return the VI id of the given port. */ unsigned int cxgb4_port_viid(const struct net_device *dev) { return netdev2pinfo(dev)->viid; } EXPORT_SYMBOL(cxgb4_port_viid); /** * cxgb4_port_idx - get the index of a port * @dev: the net device for the port * * Return the index of the given port. */ unsigned int cxgb4_port_idx(const struct net_device *dev) { return netdev2pinfo(dev)->port_id; } EXPORT_SYMBOL(cxgb4_port_idx); void cxgb4_get_tcp_stats(struct pci_dev *pdev, struct tp_tcp_stats *v4, struct tp_tcp_stats *v6) { struct adapter *adap = pci_get_drvdata(pdev); spin_lock(&adap->stats_lock); t4_tp_get_tcp_stats(adap, v4, v6); spin_unlock(&adap->stats_lock); } EXPORT_SYMBOL(cxgb4_get_tcp_stats); void cxgb4_iscsi_init(struct net_device *dev, unsigned int tag_mask, const unsigned int *pgsz_order) { struct adapter *adap = netdev2adap(dev); t4_write_reg(adap, ULP_RX_ISCSI_TAGMASK, tag_mask); t4_write_reg(adap, ULP_RX_ISCSI_PSZ, HPZ0(pgsz_order[0]) | HPZ1(pgsz_order[1]) | HPZ2(pgsz_order[2]) | HPZ3(pgsz_order[3])); } EXPORT_SYMBOL(cxgb4_iscsi_init); int cxgb4_flush_eq_cache(struct net_device *dev) { struct adapter *adap = netdev2adap(dev); int ret; ret = t4_fwaddrspace_write(adap, adap->mbox, 0xe1000000 + A_SGE_CTXT_CMD, 0x20000000); return ret; } EXPORT_SYMBOL(cxgb4_flush_eq_cache); static int read_eq_indices(struct adapter *adap, u16 qid, u16 *pidx, u16 *cidx) { u32 addr = t4_read_reg(adap, A_SGE_DBQ_CTXT_BADDR) + 24 * qid + 8; __be64 indices; int ret; spin_lock(&adap->win0_lock); ret = t4_memory_rw(adap, 0, MEM_EDC0, addr, sizeof(indices), (__be32 *)&indices, T4_MEMORY_READ); spin_unlock(&adap->win0_lock); if (!ret) { *cidx = (be64_to_cpu(indices) >> 25) & 0xffff; *pidx = (be64_to_cpu(indices) >> 9) & 0xffff; } return ret; } int cxgb4_sync_txq_pidx(struct net_device *dev, u16 qid, u16 pidx, u16 size) { struct adapter *adap = netdev2adap(dev); u16 hw_pidx, hw_cidx; int ret; ret = read_eq_indices(adap, qid, &hw_pidx, &hw_cidx); if (ret) goto out; if (pidx != hw_pidx) { u16 delta; if (pidx >= hw_pidx) delta = pidx - hw_pidx; else delta = size - hw_pidx + pidx; wmb(); t4_write_reg(adap, MYPF_REG(SGE_PF_KDOORBELL), QID(qid) | PIDX(delta)); } out: return ret; } EXPORT_SYMBOL(cxgb4_sync_txq_pidx); void cxgb4_disable_db_coalescing(struct net_device *dev) { struct adapter *adap; adap = netdev2adap(dev); t4_set_reg_field(adap, A_SGE_DOORBELL_CONTROL, F_NOCOALESCE, F_NOCOALESCE); } EXPORT_SYMBOL(cxgb4_disable_db_coalescing); void cxgb4_enable_db_coalescing(struct net_device *dev) { struct adapter *adap; adap = netdev2adap(dev); t4_set_reg_field(adap, A_SGE_DOORBELL_CONTROL, F_NOCOALESCE, 0); } EXPORT_SYMBOL(cxgb4_enable_db_coalescing); int cxgb4_read_tpte(struct net_device *dev, u32 stag, __be32 *tpte) { struct adapter *adap; u32 offset, memtype, memaddr; u32 edc0_size, edc1_size, mc0_size, mc1_size; u32 edc0_end, edc1_end, mc0_end, mc1_end; int ret; adap = netdev2adap(dev); offset = ((stag >> 8) * 32) + adap->vres.stag.start; /* Figure out where the offset lands in the Memory Type/Address scheme. * This code assumes that the memory is laid out starting at offset 0 * with no breaks as: EDC0, EDC1, MC0, MC1. All cards have both EDC0 * and EDC1. Some cards will have neither MC0 nor MC1, most cards have * MC0, and some have both MC0 and MC1. */ edc0_size = EDRAM_SIZE_GET(t4_read_reg(adap, MA_EDRAM0_BAR)) << 20; edc1_size = EDRAM_SIZE_GET(t4_read_reg(adap, MA_EDRAM1_BAR)) << 20; mc0_size = EXT_MEM_SIZE_GET(t4_read_reg(adap, MA_EXT_MEMORY_BAR)) << 20; edc0_end = edc0_size; edc1_end = edc0_end + edc1_size; mc0_end = edc1_end + mc0_size; if (offset < edc0_end) { memtype = MEM_EDC0; memaddr = offset; } else if (offset < edc1_end) { memtype = MEM_EDC1; memaddr = offset - edc0_end; } else { if (offset < mc0_end) { memtype = MEM_MC0; memaddr = offset - edc1_end; } else if (is_t4(adap->params.chip)) { /* T4 only has a single memory channel */ goto err; } else { mc1_size = EXT_MEM_SIZE_GET( t4_read_reg(adap, MA_EXT_MEMORY1_BAR)) << 20; mc1_end = mc0_end + mc1_size; if (offset < mc1_end) { memtype = MEM_MC1; memaddr = offset - mc0_end; } else { /* offset beyond the end of any memory */ goto err; } } } spin_lock(&adap->win0_lock); ret = t4_memory_rw(adap, 0, memtype, memaddr, 32, tpte, T4_MEMORY_READ); spin_unlock(&adap->win0_lock); return ret; err: dev_err(adap->pdev_dev, "stag %#x, offset %#x out of range\n", stag, offset); return -EINVAL; } EXPORT_SYMBOL(cxgb4_read_tpte); u64 cxgb4_read_sge_timestamp(struct net_device *dev) { u32 hi, lo; struct adapter *adap; adap = netdev2adap(dev); lo = t4_read_reg(adap, SGE_TIMESTAMP_LO); hi = GET_TSVAL(t4_read_reg(adap, SGE_TIMESTAMP_HI)); return ((u64)hi << 32) | (u64)lo; } EXPORT_SYMBOL(cxgb4_read_sge_timestamp); static struct pci_driver cxgb4_driver; static void check_neigh_update(struct neighbour *neigh) { const struct device *parent; const struct net_device *netdev = neigh->dev; if (netdev->priv_flags & IFF_802_1Q_VLAN) netdev = vlan_dev_real_dev(netdev); parent = netdev->dev.parent; if (parent && parent->driver == &cxgb4_driver.driver) t4_l2t_update(dev_get_drvdata(parent), neigh); } static int netevent_cb(struct notifier_block *nb, unsigned long event, void *data) { switch (event) { case NETEVENT_NEIGH_UPDATE: check_neigh_update(data); break; case NETEVENT_REDIRECT: default: break; } return 0; } static bool netevent_registered; static struct notifier_block cxgb4_netevent_nb = { .notifier_call = netevent_cb }; static void drain_db_fifo(struct adapter *adap, int usecs) { u32 v1, v2, lp_count, hp_count; do { v1 = t4_read_reg(adap, A_SGE_DBFIFO_STATUS); v2 = t4_read_reg(adap, SGE_DBFIFO_STATUS2); if (is_t4(adap->params.chip)) { lp_count = G_LP_COUNT(v1); hp_count = G_HP_COUNT(v1); } else { lp_count = G_LP_COUNT_T5(v1); hp_count = G_HP_COUNT_T5(v2); } if (lp_count == 0 && hp_count == 0) break; set_current_state(TASK_UNINTERRUPTIBLE); schedule_timeout(usecs_to_jiffies(usecs)); } while (1); } static void disable_txq_db(struct sge_txq *q) { unsigned long flags; spin_lock_irqsave(&q->db_lock, flags); q->db_disabled = 1; spin_unlock_irqrestore(&q->db_lock, flags); } static void enable_txq_db(struct adapter *adap, struct sge_txq *q) { spin_lock_irq(&q->db_lock); if (q->db_pidx_inc) { /* Make sure that all writes to the TX descriptors * are committed before we tell HW about them. */ wmb(); t4_write_reg(adap, MYPF_REG(SGE_PF_KDOORBELL), QID(q->cntxt_id) | PIDX(q->db_pidx_inc)); q->db_pidx_inc = 0; } q->db_disabled = 0; spin_unlock_irq(&q->db_lock); } static void disable_dbs(struct adapter *adap) { int i; for_each_ethrxq(&adap->sge, i) disable_txq_db(&adap->sge.ethtxq[i].q); for_each_ofldrxq(&adap->sge, i) disable_txq_db(&adap->sge.ofldtxq[i].q); for_each_port(adap, i) disable_txq_db(&adap->sge.ctrlq[i].q); } static void enable_dbs(struct adapter *adap) { int i; for_each_ethrxq(&adap->sge, i) enable_txq_db(adap, &adap->sge.ethtxq[i].q); for_each_ofldrxq(&adap->sge, i) enable_txq_db(adap, &adap->sge.ofldtxq[i].q); for_each_port(adap, i) enable_txq_db(adap, &adap->sge.ctrlq[i].q); } static void notify_rdma_uld(struct adapter *adap, enum cxgb4_control cmd) { if (adap->uld_handle[CXGB4_ULD_RDMA]) ulds[CXGB4_ULD_RDMA].control(adap->uld_handle[CXGB4_ULD_RDMA], cmd); } static void process_db_full(struct work_struct *work) { struct adapter *adap; adap = container_of(work, struct adapter, db_full_task); drain_db_fifo(adap, dbfifo_drain_delay); enable_dbs(adap); notify_rdma_uld(adap, CXGB4_CONTROL_DB_EMPTY); t4_set_reg_field(adap, SGE_INT_ENABLE3, DBFIFO_HP_INT | DBFIFO_LP_INT, DBFIFO_HP_INT | DBFIFO_LP_INT); } static void sync_txq_pidx(struct adapter *adap, struct sge_txq *q) { u16 hw_pidx, hw_cidx; int ret; spin_lock_irq(&q->db_lock); ret = read_eq_indices(adap, (u16)q->cntxt_id, &hw_pidx, &hw_cidx); if (ret) goto out; if (q->db_pidx != hw_pidx) { u16 delta; if (q->db_pidx >= hw_pidx) delta = q->db_pidx - hw_pidx; else delta = q->size - hw_pidx + q->db_pidx; wmb(); t4_write_reg(adap, MYPF_REG(SGE_PF_KDOORBELL), QID(q->cntxt_id) | PIDX(delta)); } out: q->db_disabled = 0; q->db_pidx_inc = 0; spin_unlock_irq(&q->db_lock); if (ret) CH_WARN(adap, "DB drop recovery failed.\n"); } static void recover_all_queues(struct adapter *adap) { int i; for_each_ethrxq(&adap->sge, i) sync_txq_pidx(adap, &adap->sge.ethtxq[i].q); for_each_ofldrxq(&adap->sge, i) sync_txq_pidx(adap, &adap->sge.ofldtxq[i].q); for_each_port(adap, i) sync_txq_pidx(adap, &adap->sge.ctrlq[i].q); } static void process_db_drop(struct work_struct *work) { struct adapter *adap; adap = container_of(work, struct adapter, db_drop_task); if (is_t4(adap->params.chip)) { drain_db_fifo(adap, dbfifo_drain_delay); notify_rdma_uld(adap, CXGB4_CONTROL_DB_DROP); drain_db_fifo(adap, dbfifo_drain_delay); recover_all_queues(adap); drain_db_fifo(adap, dbfifo_drain_delay); enable_dbs(adap); notify_rdma_uld(adap, CXGB4_CONTROL_DB_EMPTY); } else { u32 dropped_db = t4_read_reg(adap, 0x010ac); u16 qid = (dropped_db >> 15) & 0x1ffff; u16 pidx_inc = dropped_db & 0x1fff; unsigned int s_qpp; unsigned short udb_density; unsigned long qpshift; int page; u32 udb; dev_warn(adap->pdev_dev, "Dropped DB 0x%x qid %d bar2 %d coalesce %d pidx %d\n", dropped_db, qid, (dropped_db >> 14) & 1, (dropped_db >> 13) & 1, pidx_inc); drain_db_fifo(adap, 1); s_qpp = QUEUESPERPAGEPF1 * adap->fn; udb_density = 1 << QUEUESPERPAGEPF0_GET(t4_read_reg(adap, SGE_EGRESS_QUEUES_PER_PAGE_PF) >> s_qpp); qpshift = PAGE_SHIFT - ilog2(udb_density); udb = qid << qpshift; udb &= PAGE_MASK; page = udb / PAGE_SIZE; udb += (qid - (page * udb_density)) * 128; writel(PIDX(pidx_inc), adap->bar2 + udb + 8); /* Re-enable BAR2 WC */ t4_set_reg_field(adap, 0x10b0, 1<<15, 1<<15); } t4_set_reg_field(adap, A_SGE_DOORBELL_CONTROL, F_DROPPED_DB, 0); } void t4_db_full(struct adapter *adap) { if (is_t4(adap->params.chip)) { disable_dbs(adap); notify_rdma_uld(adap, CXGB4_CONTROL_DB_FULL); t4_set_reg_field(adap, SGE_INT_ENABLE3, DBFIFO_HP_INT | DBFIFO_LP_INT, 0); queue_work(adap->workq, &adap->db_full_task); } } void t4_db_dropped(struct adapter *adap) { if (is_t4(adap->params.chip)) { disable_dbs(adap); notify_rdma_uld(adap, CXGB4_CONTROL_DB_FULL); } queue_work(adap->workq, &adap->db_drop_task); } static void uld_attach(struct adapter *adap, unsigned int uld) { void *handle; struct cxgb4_lld_info lli; unsigned short i; lli.pdev = adap->pdev; lli.pf = adap->fn; lli.l2t = adap->l2t; lli.tids = &adap->tids; lli.ports = adap->port; lli.vr = &adap->vres; lli.mtus = adap->params.mtus; if (uld == CXGB4_ULD_RDMA) { lli.rxq_ids = adap->sge.rdma_rxq; lli.ciq_ids = adap->sge.rdma_ciq; lli.nrxq = adap->sge.rdmaqs; lli.nciq = adap->sge.rdmaciqs; } else if (uld == CXGB4_ULD_ISCSI) { lli.rxq_ids = adap->sge.ofld_rxq; lli.nrxq = adap->sge.ofldqsets; } lli.ntxq = adap->sge.ofldqsets; lli.nchan = adap->params.nports; lli.nports = adap->params.nports; lli.wr_cred = adap->params.ofldq_wr_cred; lli.adapter_type = adap->params.chip; lli.iscsi_iolen = MAXRXDATA_GET(t4_read_reg(adap, TP_PARA_REG2)); lli.cclk_ps = 1000000000 / adap->params.vpd.cclk; lli.udb_density = 1 << QUEUESPERPAGEPF0_GET( t4_read_reg(adap, SGE_EGRESS_QUEUES_PER_PAGE_PF) >> (adap->fn * 4)); lli.ucq_density = 1 << QUEUESPERPAGEPF0_GET( t4_read_reg(adap, SGE_INGRESS_QUEUES_PER_PAGE_PF) >> (adap->fn * 4)); lli.filt_mode = adap->params.tp.vlan_pri_map; /* MODQ_REQ_MAP sets queues 0-3 to chan 0-3 */ for (i = 0; i < NCHAN; i++) lli.tx_modq[i] = i; lli.gts_reg = adap->regs + MYPF_REG(SGE_PF_GTS); lli.db_reg = adap->regs + MYPF_REG(SGE_PF_KDOORBELL); lli.fw_vers = adap->params.fw_vers; lli.dbfifo_int_thresh = dbfifo_int_thresh; lli.sge_ingpadboundary = adap->sge.fl_align; lli.sge_egrstatuspagesize = adap->sge.stat_len; lli.sge_pktshift = adap->sge.pktshift; lli.enable_fw_ofld_conn = adap->flags & FW_OFLD_CONN; lli.max_ordird_qp = adap->params.max_ordird_qp; lli.max_ird_adapter = adap->params.max_ird_adapter; lli.ulptx_memwrite_dsgl = adap->params.ulptx_memwrite_dsgl; handle = ulds[uld].add(&lli); if (IS_ERR(handle)) { dev_warn(adap->pdev_dev, "could not attach to the %s driver, error %ld\n", uld_str[uld], PTR_ERR(handle)); return; } adap->uld_handle[uld] = handle; if (!netevent_registered) { register_netevent_notifier(&cxgb4_netevent_nb); netevent_registered = true; } if (adap->flags & FULL_INIT_DONE) ulds[uld].state_change(handle, CXGB4_STATE_UP); } static void attach_ulds(struct adapter *adap) { unsigned int i; spin_lock(&adap_rcu_lock); list_add_tail_rcu(&adap->rcu_node, &adap_rcu_list); spin_unlock(&adap_rcu_lock); mutex_lock(&uld_mutex); list_add_tail(&adap->list_node, &adapter_list); for (i = 0; i < CXGB4_ULD_MAX; i++) if (ulds[i].add) uld_attach(adap, i); mutex_unlock(&uld_mutex); } static void detach_ulds(struct adapter *adap) { unsigned int i; mutex_lock(&uld_mutex); list_del(&adap->list_node); for (i = 0; i < CXGB4_ULD_MAX; i++) if (adap->uld_handle[i]) { ulds[i].state_change(adap->uld_handle[i], CXGB4_STATE_DETACH); adap->uld_handle[i] = NULL; } if (netevent_registered && list_empty(&adapter_list)) { unregister_netevent_notifier(&cxgb4_netevent_nb); netevent_registered = false; } mutex_unlock(&uld_mutex); spin_lock(&adap_rcu_lock); list_del_rcu(&adap->rcu_node); spin_unlock(&adap_rcu_lock); } static void notify_ulds(struct adapter *adap, enum cxgb4_state new_state) { unsigned int i; mutex_lock(&uld_mutex); for (i = 0; i < CXGB4_ULD_MAX; i++) if (adap->uld_handle[i]) ulds[i].state_change(adap->uld_handle[i], new_state); mutex_unlock(&uld_mutex); } /** * cxgb4_register_uld - register an upper-layer driver * @type: the ULD type * @p: the ULD methods * * Registers an upper-layer driver with this driver and notifies the ULD * about any presently available devices that support its type. Returns * %-EBUSY if a ULD of the same type is already registered. */ int cxgb4_register_uld(enum cxgb4_uld type, const struct cxgb4_uld_info *p) { int ret = 0; struct adapter *adap; if (type >= CXGB4_ULD_MAX) return -EINVAL; mutex_lock(&uld_mutex); if (ulds[type].add) { ret = -EBUSY; goto out; } ulds[type] = *p; list_for_each_entry(adap, &adapter_list, list_node) uld_attach(adap, type); out: mutex_unlock(&uld_mutex); return ret; } EXPORT_SYMBOL(cxgb4_register_uld); /** * cxgb4_unregister_uld - unregister an upper-layer driver * @type: the ULD type * * Unregisters an existing upper-layer driver. */ int cxgb4_unregister_uld(enum cxgb4_uld type) { struct adapter *adap; if (type >= CXGB4_ULD_MAX) return -EINVAL; mutex_lock(&uld_mutex); list_for_each_entry(adap, &adapter_list, list_node) adap->uld_handle[type] = NULL; ulds[type].add = NULL; mutex_unlock(&uld_mutex); return 0; } EXPORT_SYMBOL(cxgb4_unregister_uld); /* Check if netdev on which event is occured belongs to us or not. Return * success (true) if it belongs otherwise failure (false). * Called with rcu_read_lock() held. */ static bool cxgb4_netdev(const struct net_device *netdev) { struct adapter *adap; int i; list_for_each_entry_rcu(adap, &adap_rcu_list, rcu_node) for (i = 0; i < MAX_NPORTS; i++) if (adap->port[i] == netdev) return true; return false; } static int clip_add(struct net_device *event_dev, struct inet6_ifaddr *ifa, unsigned long event) { int ret = NOTIFY_DONE; rcu_read_lock(); if (cxgb4_netdev(event_dev)) { switch (event) { case NETDEV_UP: ret = cxgb4_clip_get(event_dev, (const struct in6_addr *)ifa->addr.s6_addr); if (ret < 0) { rcu_read_unlock(); return ret; } ret = NOTIFY_OK; break; case NETDEV_DOWN: cxgb4_clip_release(event_dev, (const struct in6_addr *)ifa->addr.s6_addr); ret = NOTIFY_OK; break; default: break; } } rcu_read_unlock(); return ret; } static int cxgb4_inet6addr_handler(struct notifier_block *this, unsigned long event, void *data) { struct inet6_ifaddr *ifa = data; struct net_device *event_dev; int ret = NOTIFY_DONE; struct bonding *bond = netdev_priv(ifa->idev->dev); struct list_head *iter; struct slave *slave; struct pci_dev *first_pdev = NULL; if (ifa->idev->dev->priv_flags & IFF_802_1Q_VLAN) { event_dev = vlan_dev_real_dev(ifa->idev->dev); ret = clip_add(event_dev, ifa, event); } else if (ifa->idev->dev->flags & IFF_MASTER) { /* It is possible that two different adapters are bonded in one * bond. We need to find such different adapters and add clip * in all of them only once. */ bond_for_each_slave(bond, slave, iter) { if (!first_pdev) { ret = clip_add(slave->dev, ifa, event); /* If clip_add is success then only initialize * first_pdev since it means it is our device */ if (ret == NOTIFY_OK) first_pdev = to_pci_dev( slave->dev->dev.parent); } else if (first_pdev != to_pci_dev(slave->dev->dev.parent)) ret = clip_add(slave->dev, ifa, event); } } else ret = clip_add(ifa->idev->dev, ifa, event); return ret; } static struct notifier_block cxgb4_inet6addr_notifier = { .notifier_call = cxgb4_inet6addr_handler }; /* Retrieves IPv6 addresses from a root device (bond, vlan) associated with * a physical device. * The physical device reference is needed to send the actul CLIP command. */ static int update_dev_clip(struct net_device *root_dev, struct net_device *dev) { struct inet6_dev *idev = NULL; struct inet6_ifaddr *ifa; int ret = 0; idev = __in6_dev_get(root_dev); if (!idev) return ret; read_lock_bh(&idev->lock); list_for_each_entry(ifa, &idev->addr_list, if_list) { ret = cxgb4_clip_get(dev, (const struct in6_addr *)ifa->addr.s6_addr); if (ret < 0) break; } read_unlock_bh(&idev->lock); return ret; } static int update_root_dev_clip(struct net_device *dev) { struct net_device *root_dev = NULL; int i, ret = 0; /* First populate the real net device's IPv6 addresses */ ret = update_dev_clip(dev, dev); if (ret) return ret; /* Parse all bond and vlan devices layered on top of the physical dev */ for (i = 0; i < VLAN_N_VID; i++) { root_dev = __vlan_find_dev_deep_rcu(dev, htons(ETH_P_8021Q), i); if (!root_dev) continue; ret = update_dev_clip(root_dev, dev); if (ret) break; } return ret; } static void update_clip(const struct adapter *adap) { int i; struct net_device *dev; int ret; rcu_read_lock(); for (i = 0; i < MAX_NPORTS; i++) { dev = adap->port[i]; ret = 0; if (dev) ret = update_root_dev_clip(dev); if (ret < 0) break; } rcu_read_unlock(); } /** * cxgb_up - enable the adapter * @adap: adapter being enabled * * Called when the first port is enabled, this function performs the * actions necessary to make an adapter operational, such as completing * the initialization of HW modules, and enabling interrupts. * * Must be called with the rtnl lock held. */ static int cxgb_up(struct adapter *adap) { int err; err = setup_sge_queues(adap); if (err) goto out; err = setup_rss(adap); if (err) goto freeq; if (adap->flags & USING_MSIX) { name_msix_vecs(adap); err = request_irq(adap->msix_info[0].vec, t4_nondata_intr, 0, adap->msix_info[0].desc, adap); if (err) goto irq_err; err = request_msix_queue_irqs(adap); if (err) { free_irq(adap->msix_info[0].vec, adap); goto irq_err; } } else { err = request_irq(adap->pdev->irq, t4_intr_handler(adap), (adap->flags & USING_MSI) ? 0 : IRQF_SHARED, adap->port[0]->name, adap); if (err) goto irq_err; } enable_rx(adap); t4_sge_start(adap); t4_intr_enable(adap); adap->flags |= FULL_INIT_DONE; notify_ulds(adap, CXGB4_STATE_UP); update_clip(adap); out: return err; irq_err: dev_err(adap->pdev_dev, "request_irq failed, err %d\n", err); freeq: t4_free_sge_resources(adap); goto out; } static void cxgb_down(struct adapter *adapter) { t4_intr_disable(adapter); cancel_work_sync(&adapter->tid_release_task); cancel_work_sync(&adapter->db_full_task); cancel_work_sync(&adapter->db_drop_task); adapter->tid_release_task_busy = false; adapter->tid_release_head = NULL; if (adapter->flags & USING_MSIX) { free_msix_queue_irqs(adapter); free_irq(adapter->msix_info[0].vec, adapter); } else free_irq(adapter->pdev->irq, adapter); quiesce_rx(adapter); t4_sge_stop(adapter); t4_free_sge_resources(adapter); adapter->flags &= ~FULL_INIT_DONE; } /* * net_device operations */ static int cxgb_open(struct net_device *dev) { int err; struct port_info *pi = netdev_priv(dev); struct adapter *adapter = pi->adapter; netif_carrier_off(dev); if (!(adapter->flags & FULL_INIT_DONE)) { err = cxgb_up(adapter); if (err < 0) return err; } err = link_start(dev); if (!err) netif_tx_start_all_queues(dev); return err; } static int cxgb_close(struct net_device *dev) { struct port_info *pi = netdev_priv(dev); struct adapter *adapter = pi->adapter; netif_tx_stop_all_queues(dev); netif_carrier_off(dev); return t4_enable_vi(adapter, adapter->fn, pi->viid, false, false); } /* Return an error number if the indicated filter isn't writable ... */ static int writable_filter(struct filter_entry *f) { if (f->locked) return -EPERM; if (f->pending) return -EBUSY; return 0; } /* Delete the filter at the specified index (if valid). The checks for all * the common problems with doing this like the filter being locked, currently * pending in another operation, etc. */ static int delete_filter(struct adapter *adapter, unsigned int fidx) { struct filter_entry *f; int ret; if (fidx >= adapter->tids.nftids + adapter->tids.nsftids) return -EINVAL; f = &adapter->tids.ftid_tab[fidx]; ret = writable_filter(f); if (ret) return ret; if (f->valid) return del_filter_wr(adapter, fidx); return 0; } int cxgb4_create_server_filter(const struct net_device *dev, unsigned int stid, __be32 sip, __be16 sport, __be16 vlan, unsigned int queue, unsigned char port, unsigned char mask) { int ret; struct filter_entry *f; struct adapter *adap; int i; u8 *val; adap = netdev2adap(dev); /* Adjust stid to correct filter index */ stid -= adap->tids.sftid_base; stid += adap->tids.nftids; /* Check to make sure the filter requested is writable ... */ f = &adap->tids.ftid_tab[stid]; ret = writable_filter(f); if (ret) return ret; /* Clear out any old resources being used by the filter before * we start constructing the new filter. */ if (f->valid) clear_filter(adap, f); /* Clear out filter specifications */ memset(&f->fs, 0, sizeof(struct ch_filter_specification)); f->fs.val.lport = cpu_to_be16(sport); f->fs.mask.lport = ~0; val = (u8 *)&sip; if ((val[0] | val[1] | val[2] | val[3]) != 0) { for (i = 0; i < 4; i++) { f->fs.val.lip[i] = val[i]; f->fs.mask.lip[i] = ~0; } if (adap->params.tp.vlan_pri_map & F_PORT) { f->fs.val.iport = port; f->fs.mask.iport = mask; } } if (adap->params.tp.vlan_pri_map & F_PROTOCOL) { f->fs.val.proto = IPPROTO_TCP; f->fs.mask.proto = ~0; } f->fs.dirsteer = 1; f->fs.iq = queue; /* Mark filter as locked */ f->locked = 1; f->fs.rpttid = 1; ret = set_filter_wr(adap, stid); if (ret) { clear_filter(adap, f); return ret; } return 0; } EXPORT_SYMBOL(cxgb4_create_server_filter); int cxgb4_remove_server_filter(const struct net_device *dev, unsigned int stid, unsigned int queue, bool ipv6) { int ret; struct filter_entry *f; struct adapter *adap; adap = netdev2adap(dev); /* Adjust stid to correct filter index */ stid -= adap->tids.sftid_base; stid += adap->tids.nftids; f = &adap->tids.ftid_tab[stid]; /* Unlock the filter */ f->locked = 0; ret = delete_filter(adap, stid); if (ret) return ret; return 0; } EXPORT_SYMBOL(cxgb4_remove_server_filter); static struct rtnl_link_stats64 *cxgb_get_stats(struct net_device *dev, struct rtnl_link_stats64 *ns) { struct port_stats stats; struct port_info *p = netdev_priv(dev); struct adapter *adapter = p->adapter; /* Block retrieving statistics during EEH error * recovery. Otherwise, the recovery might fail * and the PCI device will be removed permanently */ spin_lock(&adapter->stats_lock); if (!netif_device_present(dev)) { spin_unlock(&adapter->stats_lock); return ns; } t4_get_port_stats(adapter, p->tx_chan, &stats); spin_unlock(&adapter->stats_lock); ns->tx_bytes = stats.tx_octets; ns->tx_packets = stats.tx_frames; ns->rx_bytes = stats.rx_octets; ns->rx_packets = stats.rx_frames; ns->multicast = stats.rx_mcast_frames; /* detailed rx_errors */ ns->rx_length_errors = stats.rx_jabber + stats.rx_too_long + stats.rx_runt; ns->rx_over_errors = 0; ns->rx_crc_errors = stats.rx_fcs_err; ns->rx_frame_errors = stats.rx_symbol_err; ns->rx_fifo_errors = stats.rx_ovflow0 + stats.rx_ovflow1 + stats.rx_ovflow2 + stats.rx_ovflow3 + stats.rx_trunc0 + stats.rx_trunc1 + stats.rx_trunc2 + stats.rx_trunc3; ns->rx_missed_errors = 0; /* detailed tx_errors */ ns->tx_aborted_errors = 0; ns->tx_carrier_errors = 0; ns->tx_fifo_errors = 0; ns->tx_heartbeat_errors = 0; ns->tx_window_errors = 0; ns->tx_errors = stats.tx_error_frames; ns->rx_errors = stats.rx_symbol_err + stats.rx_fcs_err + ns->rx_length_errors + stats.rx_len_err + ns->rx_fifo_errors; return ns; } static int cxgb_ioctl(struct net_device *dev, struct ifreq *req, int cmd) { unsigned int mbox; int ret = 0, prtad, devad; struct port_info *pi = netdev_priv(dev); struct mii_ioctl_data *data = (struct mii_ioctl_data *)&req->ifr_data; switch (cmd) { case SIOCGMIIPHY: if (pi->mdio_addr < 0) return -EOPNOTSUPP; data->phy_id = pi->mdio_addr; break; case SIOCGMIIREG: case SIOCSMIIREG: if (mdio_phy_id_is_c45(data->phy_id)) { prtad = mdio_phy_id_prtad(data->phy_id); devad = mdio_phy_id_devad(data->phy_id); } else if (data->phy_id < 32) { prtad = data->phy_id; devad = 0; data->reg_num &= 0x1f; } else return -EINVAL; mbox = pi->adapter->fn; if (cmd == SIOCGMIIREG) ret = t4_mdio_rd(pi->adapter, mbox, prtad, devad, data->reg_num, &data->val_out); else ret = t4_mdio_wr(pi->adapter, mbox, prtad, devad, data->reg_num, data->val_in); break; default: return -EOPNOTSUPP; } return ret; } static void cxgb_set_rxmode(struct net_device *dev) { /* unfortunately we can't return errors to the stack */ set_rxmode(dev, -1, false); } static int cxgb_change_mtu(struct net_device *dev, int new_mtu) { int ret; struct port_info *pi = netdev_priv(dev); if (new_mtu < 81 || new_mtu > MAX_MTU) /* accommodate SACK */ return -EINVAL; ret = t4_set_rxmode(pi->adapter, pi->adapter->fn, pi->viid, new_mtu, -1, -1, -1, -1, true); if (!ret) dev->mtu = new_mtu; return ret; } static int cxgb_set_mac_addr(struct net_device *dev, void *p) { int ret; struct sockaddr *addr = p; struct port_info *pi = netdev_priv(dev); if (!is_valid_ether_addr(addr->sa_data)) return -EADDRNOTAVAIL; ret = t4_change_mac(pi->adapter, pi->adapter->fn, pi->viid, pi->xact_addr_filt, addr->sa_data, true, true); if (ret < 0) return ret; memcpy(dev->dev_addr, addr->sa_data, dev->addr_len); pi->xact_addr_filt = ret; return 0; } #ifdef CONFIG_NET_POLL_CONTROLLER static void cxgb_netpoll(struct net_device *dev) { struct port_info *pi = netdev_priv(dev); struct adapter *adap = pi->adapter; if (adap->flags & USING_MSIX) { int i; struct sge_eth_rxq *rx = &adap->sge.ethrxq[pi->first_qset]; for (i = pi->nqsets; i; i--, rx++) t4_sge_intr_msix(0, &rx->rspq); } else t4_intr_handler(adap)(0, adap); } #endif static const struct net_device_ops cxgb4_netdev_ops = { .ndo_open = cxgb_open, .ndo_stop = cxgb_close, .ndo_start_xmit = t4_eth_xmit, .ndo_select_queue = cxgb_select_queue, .ndo_get_stats64 = cxgb_get_stats, .ndo_set_rx_mode = cxgb_set_rxmode, .ndo_set_mac_address = cxgb_set_mac_addr, .ndo_set_features = cxgb_set_features, .ndo_validate_addr = eth_validate_addr, .ndo_do_ioctl = cxgb_ioctl, .ndo_change_mtu = cxgb_change_mtu, #ifdef CONFIG_NET_POLL_CONTROLLER .ndo_poll_controller = cxgb_netpoll, #endif }; void t4_fatal_err(struct adapter *adap) { t4_set_reg_field(adap, SGE_CONTROL, GLOBALENABLE, 0); t4_intr_disable(adap); dev_alert(adap->pdev_dev, "encountered fatal error, adapter stopped\n"); } /* Return the specified PCI-E Configuration Space register from our Physical * Function. We try first via a Firmware LDST Command since we prefer to let * the firmware own all of these registers, but if that fails we go for it * directly ourselves. */ static u32 t4_read_pcie_cfg4(struct adapter *adap, int reg) { struct fw_ldst_cmd ldst_cmd; u32 val; int ret; /* Construct and send the Firmware LDST Command to retrieve the * specified PCI-E Configuration Space register. */ memset(&ldst_cmd, 0, sizeof(ldst_cmd)); ldst_cmd.op_to_addrspace = htonl(FW_CMD_OP(FW_LDST_CMD) | FW_CMD_REQUEST | FW_CMD_READ | FW_LDST_CMD_ADDRSPACE(FW_LDST_ADDRSPC_FUNC_PCIE)); ldst_cmd.cycles_to_len16 = htonl(FW_LEN16(ldst_cmd)); ldst_cmd.u.pcie.select_naccess = FW_LDST_CMD_NACCESS(1); ldst_cmd.u.pcie.ctrl_to_fn = (FW_LDST_CMD_LC | FW_LDST_CMD_FN(adap->fn)); ldst_cmd.u.pcie.r = reg; ret = t4_wr_mbox(adap, adap->mbox, &ldst_cmd, sizeof(ldst_cmd), &ldst_cmd); /* If the LDST Command suucceeded, exctract the returned register * value. Otherwise read it directly ourself. */ if (ret == 0) val = ntohl(ldst_cmd.u.pcie.data[0]); else t4_hw_pci_read_cfg4(adap, reg, &val); return val; } static void setup_memwin(struct adapter *adap) { u32 mem_win0_base, mem_win1_base, mem_win2_base, mem_win2_aperture; if (is_t4(adap->params.chip)) { u32 bar0; /* Truncation intentional: we only read the bottom 32-bits of * the 64-bit BAR0/BAR1 ... We use the hardware backdoor * mechanism to read BAR0 instead of using * pci_resource_start() because we could be operating from * within a Virtual Machine which is trapping our accesses to * our Configuration Space and we need to set up the PCI-E * Memory Window decoders with the actual addresses which will * be coming across the PCI-E link. */ bar0 = t4_read_pcie_cfg4(adap, PCI_BASE_ADDRESS_0); bar0 &= PCI_BASE_ADDRESS_MEM_MASK; adap->t4_bar0 = bar0; mem_win0_base = bar0 + MEMWIN0_BASE; mem_win1_base = bar0 + MEMWIN1_BASE; mem_win2_base = bar0 + MEMWIN2_BASE; mem_win2_aperture = MEMWIN2_APERTURE; } else { /* For T5, only relative offset inside the PCIe BAR is passed */ mem_win0_base = MEMWIN0_BASE; mem_win1_base = MEMWIN1_BASE; mem_win2_base = MEMWIN2_BASE_T5; mem_win2_aperture = MEMWIN2_APERTURE_T5; } t4_write_reg(adap, PCIE_MEM_ACCESS_REG(PCIE_MEM_ACCESS_BASE_WIN, 0), mem_win0_base | BIR(0) | WINDOW(ilog2(MEMWIN0_APERTURE) - 10)); t4_write_reg(adap, PCIE_MEM_ACCESS_REG(PCIE_MEM_ACCESS_BASE_WIN, 1), mem_win1_base | BIR(0) | WINDOW(ilog2(MEMWIN1_APERTURE) - 10)); t4_write_reg(adap, PCIE_MEM_ACCESS_REG(PCIE_MEM_ACCESS_BASE_WIN, 2), mem_win2_base | BIR(0) | WINDOW(ilog2(mem_win2_aperture) - 10)); t4_read_reg(adap, PCIE_MEM_ACCESS_REG(PCIE_MEM_ACCESS_BASE_WIN, 2)); } static void setup_memwin_rdma(struct adapter *adap) { if (adap->vres.ocq.size) { u32 start; unsigned int sz_kb; start = t4_read_pcie_cfg4(adap, PCI_BASE_ADDRESS_2); start &= PCI_BASE_ADDRESS_MEM_MASK; start += OCQ_WIN_OFFSET(adap->pdev, &adap->vres); sz_kb = roundup_pow_of_two(adap->vres.ocq.size) >> 10; t4_write_reg(adap, PCIE_MEM_ACCESS_REG(PCIE_MEM_ACCESS_BASE_WIN, 3), start | BIR(1) | WINDOW(ilog2(sz_kb))); t4_write_reg(adap, PCIE_MEM_ACCESS_REG(PCIE_MEM_ACCESS_OFFSET, 3), adap->vres.ocq.start); t4_read_reg(adap, PCIE_MEM_ACCESS_REG(PCIE_MEM_ACCESS_OFFSET, 3)); } } static int adap_init1(struct adapter *adap, struct fw_caps_config_cmd *c) { u32 v; int ret; /* get device capabilities */ memset(c, 0, sizeof(*c)); c->op_to_write = htonl(FW_CMD_OP(FW_CAPS_CONFIG_CMD) | FW_CMD_REQUEST | FW_CMD_READ); c->cfvalid_to_len16 = htonl(FW_LEN16(*c)); ret = t4_wr_mbox(adap, adap->fn, c, sizeof(*c), c); if (ret < 0) return ret; /* select capabilities we'll be using */ if (c->niccaps & htons(FW_CAPS_CONFIG_NIC_VM)) { if (!vf_acls) c->niccaps ^= htons(FW_CAPS_CONFIG_NIC_VM); else c->niccaps = htons(FW_CAPS_CONFIG_NIC_VM); } else if (vf_acls) { dev_err(adap->pdev_dev, "virtualization ACLs not supported"); return ret; } c->op_to_write = htonl(FW_CMD_OP(FW_CAPS_CONFIG_CMD) | FW_CMD_REQUEST | FW_CMD_WRITE); ret = t4_wr_mbox(adap, adap->fn, c, sizeof(*c), NULL); if (ret < 0) return ret; ret = t4_config_glbl_rss(adap, adap->fn, FW_RSS_GLB_CONFIG_CMD_MODE_BASICVIRTUAL, FW_RSS_GLB_CONFIG_CMD_TNLMAPEN | FW_RSS_GLB_CONFIG_CMD_TNLALLLKP); if (ret < 0) return ret; ret = t4_cfg_pfvf(adap, adap->fn, adap->fn, 0, MAX_EGRQ, 64, MAX_INGQ, 0, 0, 4, 0xf, 0xf, 16, FW_CMD_CAP_PF, FW_CMD_CAP_PF); if (ret < 0) return ret; t4_sge_init(adap); /* tweak some settings */ t4_write_reg(adap, TP_SHIFT_CNT, 0x64f8849); t4_write_reg(adap, ULP_RX_TDDP_PSZ, HPZ0(PAGE_SHIFT - 12)); t4_write_reg(adap, TP_PIO_ADDR, TP_INGRESS_CONFIG); v = t4_read_reg(adap, TP_PIO_DATA); t4_write_reg(adap, TP_PIO_DATA, v & ~CSUM_HAS_PSEUDO_HDR); /* first 4 Tx modulation queues point to consecutive Tx channels */ adap->params.tp.tx_modq_map = 0xE4; t4_write_reg(adap, A_TP_TX_MOD_QUEUE_REQ_MAP, V_TX_MOD_QUEUE_REQ_MAP(adap->params.tp.tx_modq_map)); /* associate each Tx modulation queue with consecutive Tx channels */ v = 0x84218421; t4_write_indirect(adap, TP_PIO_ADDR, TP_PIO_DATA, &v, 1, A_TP_TX_SCHED_HDR); t4_write_indirect(adap, TP_PIO_ADDR, TP_PIO_DATA, &v, 1, A_TP_TX_SCHED_FIFO); t4_write_indirect(adap, TP_PIO_ADDR, TP_PIO_DATA, &v, 1, A_TP_TX_SCHED_PCMD); #define T4_TX_MODQ_10G_WEIGHT_DEFAULT 16 /* in KB units */ if (is_offload(adap)) { t4_write_reg(adap, A_TP_TX_MOD_QUEUE_WEIGHT0, V_TX_MODQ_WEIGHT0(T4_TX_MODQ_10G_WEIGHT_DEFAULT) | V_TX_MODQ_WEIGHT1(T4_TX_MODQ_10G_WEIGHT_DEFAULT) | V_TX_MODQ_WEIGHT2(T4_TX_MODQ_10G_WEIGHT_DEFAULT) | V_TX_MODQ_WEIGHT3(T4_TX_MODQ_10G_WEIGHT_DEFAULT)); t4_write_reg(adap, A_TP_TX_MOD_CHANNEL_WEIGHT, V_TX_MODQ_WEIGHT0(T4_TX_MODQ_10G_WEIGHT_DEFAULT) | V_TX_MODQ_WEIGHT1(T4_TX_MODQ_10G_WEIGHT_DEFAULT) | V_TX_MODQ_WEIGHT2(T4_TX_MODQ_10G_WEIGHT_DEFAULT) | V_TX_MODQ_WEIGHT3(T4_TX_MODQ_10G_WEIGHT_DEFAULT)); } /* get basic stuff going */ return t4_early_init(adap, adap->fn); } /* * Max # of ATIDs. The absolute HW max is 16K but we keep it lower. */ #define MAX_ATIDS 8192U /* * Phase 0 of initialization: contact FW, obtain config, perform basic init. * * If the firmware we're dealing with has Configuration File support, then * we use that to perform all configuration */ /* * Tweak configuration based on module parameters, etc. Most of these have * defaults assigned to them by Firmware Configuration Files (if we're using * them) but need to be explicitly set if we're using hard-coded * initialization. But even in the case of using Firmware Configuration * Files, we'd like to expose the ability to change these via module * parameters so these are essentially common tweaks/settings for * Configuration Files and hard-coded initialization ... */ static int adap_init0_tweaks(struct adapter *adapter) { /* * Fix up various Host-Dependent Parameters like Page Size, Cache * Line Size, etc. The firmware default is for a 4KB Page Size and * 64B Cache Line Size ... */ t4_fixup_host_params(adapter, PAGE_SIZE, L1_CACHE_BYTES); /* * Process module parameters which affect early initialization. */ if (rx_dma_offset != 2 && rx_dma_offset != 0) { dev_err(&adapter->pdev->dev, "Ignoring illegal rx_dma_offset=%d, using 2\n", rx_dma_offset); rx_dma_offset = 2; } t4_set_reg_field(adapter, SGE_CONTROL, PKTSHIFT_MASK, PKTSHIFT(rx_dma_offset)); /* * Don't include the "IP Pseudo Header" in CPL_RX_PKT checksums: Linux * adds the pseudo header itself. */ t4_tp_wr_bits_indirect(adapter, TP_INGRESS_CONFIG, CSUM_HAS_PSEUDO_HDR, 0); return 0; } /* * Attempt to initialize the adapter via a Firmware Configuration File. */ static int adap_init0_config(struct adapter *adapter, int reset) { struct fw_caps_config_cmd caps_cmd; const struct firmware *cf; unsigned long mtype = 0, maddr = 0; u32 finiver, finicsum, cfcsum; int ret; int config_issued = 0; char *fw_config_file, fw_config_file_path[256]; char *config_name = NULL; /* * Reset device if necessary. */ if (reset) { ret = t4_fw_reset(adapter, adapter->mbox, PIORSTMODE | PIORST); if (ret < 0) goto bye; } /* * If we have a T4 configuration file under /lib/firmware/cxgb4/, * then use that. Otherwise, use the configuration file stored * in the adapter flash ... */ switch (CHELSIO_CHIP_VERSION(adapter->params.chip)) { case CHELSIO_T4: fw_config_file = FW4_CFNAME; break; case CHELSIO_T5: fw_config_file = FW5_CFNAME; break; default: dev_err(adapter->pdev_dev, "Device %d is not supported\n", adapter->pdev->device); ret = -EINVAL; goto bye; } ret = request_firmware(&cf, fw_config_file, adapter->pdev_dev); if (ret < 0) { config_name = "On FLASH"; mtype = FW_MEMTYPE_CF_FLASH; maddr = t4_flash_cfg_addr(adapter); } else { u32 params[7], val[7]; sprintf(fw_config_file_path, "/lib/firmware/%s", fw_config_file); config_name = fw_config_file_path; if (cf->size >= FLASH_CFG_MAX_SIZE) ret = -ENOMEM; else { params[0] = (FW_PARAMS_MNEM(FW_PARAMS_MNEM_DEV) | FW_PARAMS_PARAM_X(FW_PARAMS_PARAM_DEV_CF)); ret = t4_query_params(adapter, adapter->mbox, adapter->fn, 0, 1, params, val); if (ret == 0) { /* * For t4_memory_rw() below addresses and * sizes have to be in terms of multiples of 4 * bytes. So, if the Configuration File isn't * a multiple of 4 bytes in length we'll have * to write that out separately since we can't * guarantee that the bytes following the * residual byte in the buffer returned by * request_firmware() are zeroed out ... */ size_t resid = cf->size & 0x3; size_t size = cf->size & ~0x3; __be32 *data = (__be32 *)cf->data; mtype = FW_PARAMS_PARAM_Y_GET(val[0]); maddr = FW_PARAMS_PARAM_Z_GET(val[0]) << 16; spin_lock(&adapter->win0_lock); ret = t4_memory_rw(adapter, 0, mtype, maddr, size, data, T4_MEMORY_WRITE); if (ret == 0 && resid != 0) { union { __be32 word; char buf[4]; } last; int i; last.word = data[size >> 2]; for (i = resid; i < 4; i++) last.buf[i] = 0; ret = t4_memory_rw(adapter, 0, mtype, maddr + size, 4, &last.word, T4_MEMORY_WRITE); } spin_unlock(&adapter->win0_lock); } } release_firmware(cf); if (ret) goto bye; } /* * Issue a Capability Configuration command to the firmware to get it * to parse the Configuration File. We don't use t4_fw_config_file() * because we want the ability to modify various features after we've * processed the configuration file ... */ memset(&caps_cmd, 0, sizeof(caps_cmd)); caps_cmd.op_to_write = htonl(FW_CMD_OP(FW_CAPS_CONFIG_CMD) | FW_CMD_REQUEST | FW_CMD_READ); caps_cmd.cfvalid_to_len16 = htonl(FW_CAPS_CONFIG_CMD_CFVALID | FW_CAPS_CONFIG_CMD_MEMTYPE_CF(mtype) | FW_CAPS_CONFIG_CMD_MEMADDR64K_CF(maddr >> 16) | FW_LEN16(caps_cmd)); ret = t4_wr_mbox(adapter, adapter->mbox, &caps_cmd, sizeof(caps_cmd), &caps_cmd); /* If the CAPS_CONFIG failed with an ENOENT (for a Firmware * Configuration File in FLASH), our last gasp effort is to use the * Firmware Configuration File which is embedded in the firmware. A * very few early versions of the firmware didn't have one embedded * but we can ignore those. */ if (ret == -ENOENT) { memset(&caps_cmd, 0, sizeof(caps_cmd)); caps_cmd.op_to_write = htonl(FW_CMD_OP(FW_CAPS_CONFIG_CMD) | FW_CMD_REQUEST | FW_CMD_READ); caps_cmd.cfvalid_to_len16 = htonl(FW_LEN16(caps_cmd)); ret = t4_wr_mbox(adapter, adapter->mbox, &caps_cmd, sizeof(caps_cmd), &caps_cmd); config_name = "Firmware Default"; } config_issued = 1; if (ret < 0) goto bye; finiver = ntohl(caps_cmd.finiver); finicsum = ntohl(caps_cmd.finicsum); cfcsum = ntohl(caps_cmd.cfcsum); if (finicsum != cfcsum) dev_warn(adapter->pdev_dev, "Configuration File checksum "\ "mismatch: [fini] csum=%#x, computed csum=%#x\n", finicsum, cfcsum); /* * And now tell the firmware to use the configuration we just loaded. */ caps_cmd.op_to_write = htonl(FW_CMD_OP(FW_CAPS_CONFIG_CMD) | FW_CMD_REQUEST | FW_CMD_WRITE); caps_cmd.cfvalid_to_len16 = htonl(FW_LEN16(caps_cmd)); ret = t4_wr_mbox(adapter, adapter->mbox, &caps_cmd, sizeof(caps_cmd), NULL); if (ret < 0) goto bye; /* * Tweak configuration based on system architecture, module * parameters, etc. */ ret = adap_init0_tweaks(adapter); if (ret < 0) goto bye; /* * And finally tell the firmware to initialize itself using the * parameters from the Configuration File. */ ret = t4_fw_initialize(adapter, adapter->mbox); if (ret < 0) goto bye; /* * Return successfully and note that we're operating with parameters * not supplied by the driver, rather than from hard-wired * initialization constants burried in the driver. */ adapter->flags |= USING_SOFT_PARAMS; dev_info(adapter->pdev_dev, "Successfully configured using Firmware "\ "Configuration File \"%s\", version %#x, computed checksum %#x\n", config_name, finiver, cfcsum); return 0; /* * Something bad happened. Return the error ... (If the "error" * is that there's no Configuration File on the adapter we don't * want to issue a warning since this is fairly common.) */ bye: if (config_issued && ret != -ENOENT) dev_warn(adapter->pdev_dev, "\"%s\" configuration file error %d\n", config_name, -ret); return ret; } /* * Attempt to initialize the adapter via hard-coded, driver supplied * parameters ... */ static int adap_init0_no_config(struct adapter *adapter, int reset) { struct sge *s = &adapter->sge; struct fw_caps_config_cmd caps_cmd; u32 v; int i, ret; /* * Reset device if necessary */ if (reset) { ret = t4_fw_reset(adapter, adapter->mbox, PIORSTMODE | PIORST); if (ret < 0) goto bye; } /* * Get device capabilities and select which we'll be using. */ memset(&caps_cmd, 0, sizeof(caps_cmd)); caps_cmd.op_to_write = htonl(FW_CMD_OP(FW_CAPS_CONFIG_CMD) | FW_CMD_REQUEST | FW_CMD_READ); caps_cmd.cfvalid_to_len16 = htonl(FW_LEN16(caps_cmd)); ret = t4_wr_mbox(adapter, adapter->mbox, &caps_cmd, sizeof(caps_cmd), &caps_cmd); if (ret < 0) goto bye; if (caps_cmd.niccaps & htons(FW_CAPS_CONFIG_NIC_VM)) { if (!vf_acls) caps_cmd.niccaps ^= htons(FW_CAPS_CONFIG_NIC_VM); else caps_cmd.niccaps = htons(FW_CAPS_CONFIG_NIC_VM); } else if (vf_acls) { dev_err(adapter->pdev_dev, "virtualization ACLs not supported"); goto bye; } caps_cmd.op_to_write = htonl(FW_CMD_OP(FW_CAPS_CONFIG_CMD) | FW_CMD_REQUEST | FW_CMD_WRITE); ret = t4_wr_mbox(adapter, adapter->mbox, &caps_cmd, sizeof(caps_cmd), NULL); if (ret < 0) goto bye; /* * Tweak configuration based on system architecture, module * parameters, etc. */ ret = adap_init0_tweaks(adapter); if (ret < 0) goto bye; /* * Select RSS Global Mode we want to use. We use "Basic Virtual" * mode which maps each Virtual Interface to its own section of * the RSS Table and we turn on all map and hash enables ... */ adapter->flags |= RSS_TNLALLLOOKUP; ret = t4_config_glbl_rss(adapter, adapter->mbox, FW_RSS_GLB_CONFIG_CMD_MODE_BASICVIRTUAL, FW_RSS_GLB_CONFIG_CMD_TNLMAPEN | FW_RSS_GLB_CONFIG_CMD_HASHTOEPLITZ | ((adapter->flags & RSS_TNLALLLOOKUP) ? FW_RSS_GLB_CONFIG_CMD_TNLALLLKP : 0)); if (ret < 0) goto bye; /* * Set up our own fundamental resource provisioning ... */ ret = t4_cfg_pfvf(adapter, adapter->mbox, adapter->fn, 0, PFRES_NEQ, PFRES_NETHCTRL, PFRES_NIQFLINT, PFRES_NIQ, PFRES_TC, PFRES_NVI, FW_PFVF_CMD_CMASK_MASK, pfvfres_pmask(adapter, adapter->fn, 0), PFRES_NEXACTF, PFRES_R_CAPS, PFRES_WX_CAPS); if (ret < 0) goto bye; /* * Perform low level SGE initialization. We need to do this before we * send the firmware the INITIALIZE command because that will cause * any other PF Drivers which are waiting for the Master * Initialization to proceed forward. */ for (i = 0; i < SGE_NTIMERS - 1; i++) s->timer_val[i] = min(intr_holdoff[i], MAX_SGE_TIMERVAL); s->timer_val[SGE_NTIMERS - 1] = MAX_SGE_TIMERVAL; s->counter_val[0] = 1; for (i = 1; i < SGE_NCOUNTERS; i++) s->counter_val[i] = min(intr_cnt[i - 1], THRESHOLD_0_GET(THRESHOLD_0_MASK)); t4_sge_init(adapter); #ifdef CONFIG_PCI_IOV /* * Provision resource limits for Virtual Functions. We currently * grant them all the same static resource limits except for the Port * Access Rights Mask which we're assigning based on the PF. All of * the static provisioning stuff for both the PF and VF really needs * to be managed in a persistent manner for each device which the * firmware controls. */ { int pf, vf; for (pf = 0; pf < ARRAY_SIZE(num_vf); pf++) { if (num_vf[pf] <= 0) continue; /* VF numbering starts at 1! */ for (vf = 1; vf <= num_vf[pf]; vf++) { ret = t4_cfg_pfvf(adapter, adapter->mbox, pf, vf, VFRES_NEQ, VFRES_NETHCTRL, VFRES_NIQFLINT, VFRES_NIQ, VFRES_TC, VFRES_NVI, FW_PFVF_CMD_CMASK_MASK, pfvfres_pmask( adapter, pf, vf), VFRES_NEXACTF, VFRES_R_CAPS, VFRES_WX_CAPS); if (ret < 0) dev_warn(adapter->pdev_dev, "failed to "\ "provision pf/vf=%d/%d; " "err=%d\n", pf, vf, ret); } } } #endif /* * Set up the default filter mode. Later we'll want to implement this * via a firmware command, etc. ... This needs to be done before the * firmare initialization command ... If the selected set of fields * isn't equal to the default value, we'll need to make sure that the * field selections will fit in the 36-bit budget. */ if (tp_vlan_pri_map != TP_VLAN_PRI_MAP_DEFAULT) { int j, bits = 0; for (j = TP_VLAN_PRI_MAP_FIRST; j <= TP_VLAN_PRI_MAP_LAST; j++) switch (tp_vlan_pri_map & (1 << j)) { case 0: /* compressed filter field not enabled */ break; case FCOE_MASK: bits += 1; break; case PORT_MASK: bits += 3; break; case VNIC_ID_MASK: bits += 17; break; case VLAN_MASK: bits += 17; break; case TOS_MASK: bits += 8; break; case PROTOCOL_MASK: bits += 8; break; case ETHERTYPE_MASK: bits += 16; break; case MACMATCH_MASK: bits += 9; break; case MPSHITTYPE_MASK: bits += 3; break; case FRAGMENTATION_MASK: bits += 1; break; } if (bits > 36) { dev_err(adapter->pdev_dev, "tp_vlan_pri_map=%#x needs %d bits > 36;"\ " using %#x\n", tp_vlan_pri_map, bits, TP_VLAN_PRI_MAP_DEFAULT); tp_vlan_pri_map = TP_VLAN_PRI_MAP_DEFAULT; } } v = tp_vlan_pri_map; t4_write_indirect(adapter, TP_PIO_ADDR, TP_PIO_DATA, &v, 1, TP_VLAN_PRI_MAP); /* * We need Five Tuple Lookup mode to be set in TP_GLOBAL_CONFIG order * to support any of the compressed filter fields above. Newer * versions of the firmware do this automatically but it doesn't hurt * to set it here. Meanwhile, we do _not_ need to set Lookup Every * Packet in TP_INGRESS_CONFIG to support matching non-TCP packets * since the firmware automatically turns this on and off when we have * a non-zero number of filters active (since it does have a * performance impact). */ if (tp_vlan_pri_map) t4_set_reg_field(adapter, TP_GLOBAL_CONFIG, FIVETUPLELOOKUP_MASK, FIVETUPLELOOKUP_MASK); /* * Tweak some settings. */ t4_write_reg(adapter, TP_SHIFT_CNT, SYNSHIFTMAX(6) | RXTSHIFTMAXR1(4) | RXTSHIFTMAXR2(15) | PERSHIFTBACKOFFMAX(8) | PERSHIFTMAX(8) | KEEPALIVEMAXR1(4) | KEEPALIVEMAXR2(9)); /* * Get basic stuff going by issuing the Firmware Initialize command. * Note that this _must_ be after all PFVF commands ... */ ret = t4_fw_initialize(adapter, adapter->mbox); if (ret < 0) goto bye; /* * Return successfully! */ dev_info(adapter->pdev_dev, "Successfully configured using built-in "\ "driver parameters\n"); return 0; /* * Something bad happened. Return the error ... */ bye: return ret; } static struct fw_info fw_info_array[] = { { .chip = CHELSIO_T4, .fs_name = FW4_CFNAME, .fw_mod_name = FW4_FNAME, .fw_hdr = { .chip = FW_HDR_CHIP_T4, .fw_ver = __cpu_to_be32(FW_VERSION(T4)), .intfver_nic = FW_INTFVER(T4, NIC), .intfver_vnic = FW_INTFVER(T4, VNIC), .intfver_ri = FW_INTFVER(T4, RI), .intfver_iscsi = FW_INTFVER(T4, ISCSI), .intfver_fcoe = FW_INTFVER(T4, FCOE), }, }, { .chip = CHELSIO_T5, .fs_name = FW5_CFNAME, .fw_mod_name = FW5_FNAME, .fw_hdr = { .chip = FW_HDR_CHIP_T5, .fw_ver = __cpu_to_be32(FW_VERSION(T5)), .intfver_nic = FW_INTFVER(T5, NIC), .intfver_vnic = FW_INTFVER(T5, VNIC), .intfver_ri = FW_INTFVER(T5, RI), .intfver_iscsi = FW_INTFVER(T5, ISCSI), .intfver_fcoe = FW_INTFVER(T5, FCOE), }, } }; static struct fw_info *find_fw_info(int chip) { int i; for (i = 0; i < ARRAY_SIZE(fw_info_array); i++) { if (fw_info_array[i].chip == chip) return &fw_info_array[i]; } return NULL; } /* * Phase 0 of initialization: contact FW, obtain config, perform basic init. */ static int adap_init0(struct adapter *adap) { int ret; u32 v, port_vec; enum dev_state state; u32 params[7], val[7]; struct fw_caps_config_cmd caps_cmd; int reset = 1; /* * Contact FW, advertising Master capability (and potentially forcing * ourselves as the Master PF if our module parameter force_init is * set). */ ret = t4_fw_hello(adap, adap->mbox, adap->fn, force_init ? MASTER_MUST : MASTER_MAY, &state); if (ret < 0) { dev_err(adap->pdev_dev, "could not connect to FW, error %d\n", ret); return ret; } if (ret == adap->mbox) adap->flags |= MASTER_PF; if (force_init && state == DEV_STATE_INIT) state = DEV_STATE_UNINIT; /* * If we're the Master PF Driver and the device is uninitialized, * then let's consider upgrading the firmware ... (We always want * to check the firmware version number in order to A. get it for * later reporting and B. to warn if the currently loaded firmware * is excessively mismatched relative to the driver.) */ t4_get_fw_version(adap, &adap->params.fw_vers); t4_get_tp_version(adap, &adap->params.tp_vers); if ((adap->flags & MASTER_PF) && state != DEV_STATE_INIT) { struct fw_info *fw_info; struct fw_hdr *card_fw; const struct firmware *fw; const u8 *fw_data = NULL; unsigned int fw_size = 0; /* This is the firmware whose headers the driver was compiled * against */ fw_info = find_fw_info(CHELSIO_CHIP_VERSION(adap->params.chip)); if (fw_info == NULL) { dev_err(adap->pdev_dev, "unable to get firmware info for chip %d.\n", CHELSIO_CHIP_VERSION(adap->params.chip)); return -EINVAL; } /* allocate memory to read the header of the firmware on the * card */ card_fw = t4_alloc_mem(sizeof(*card_fw)); /* Get FW from from /lib/firmware/ */ ret = request_firmware(&fw, fw_info->fw_mod_name, adap->pdev_dev); if (ret < 0) { dev_err(adap->pdev_dev, "unable to load firmware image %s, error %d\n", fw_info->fw_mod_name, ret); } else { fw_data = fw->data; fw_size = fw->size; } /* upgrade FW logic */ ret = t4_prep_fw(adap, fw_info, fw_data, fw_size, card_fw, state, &reset); /* Cleaning up */ if (fw != NULL) release_firmware(fw); t4_free_mem(card_fw); if (ret < 0) goto bye; } /* * Grab VPD parameters. This should be done after we establish a * connection to the firmware since some of the VPD parameters * (notably the Core Clock frequency) are retrieved via requests to * the firmware. On the other hand, we need these fairly early on * so we do this right after getting ahold of the firmware. */ ret = get_vpd_params(adap, &adap->params.vpd); if (ret < 0) goto bye; /* * Find out what ports are available to us. Note that we need to do * this before calling adap_init0_no_config() since it needs nports * and portvec ... */ v = FW_PARAMS_MNEM(FW_PARAMS_MNEM_DEV) | FW_PARAMS_PARAM_X(FW_PARAMS_PARAM_DEV_PORTVEC); ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 1, &v, &port_vec); if (ret < 0) goto bye; adap->params.nports = hweight32(port_vec); adap->params.portvec = port_vec; /* * If the firmware is initialized already (and we're not forcing a * master initialization), note that we're living with existing * adapter parameters. Otherwise, it's time to try initializing the * adapter ... */ if (state == DEV_STATE_INIT) { dev_info(adap->pdev_dev, "Coming up as %s: "\ "Adapter already initialized\n", adap->flags & MASTER_PF ? "MASTER" : "SLAVE"); adap->flags |= USING_SOFT_PARAMS; } else { dev_info(adap->pdev_dev, "Coming up as MASTER: "\ "Initializing adapter\n"); /* * If the firmware doesn't support Configuration * Files warn user and exit, */ if (ret < 0) dev_warn(adap->pdev_dev, "Firmware doesn't support " "configuration file.\n"); if (force_old_init) ret = adap_init0_no_config(adap, reset); else { /* * Find out whether we're dealing with a version of * the firmware which has configuration file support. */ params[0] = (FW_PARAMS_MNEM(FW_PARAMS_MNEM_DEV) | FW_PARAMS_PARAM_X(FW_PARAMS_PARAM_DEV_CF)); ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 1, params, val); /* * If the firmware doesn't support Configuration * Files, use the old Driver-based, hard-wired * initialization. Otherwise, try using the * Configuration File support and fall back to the * Driver-based initialization if there's no * Configuration File found. */ if (ret < 0) ret = adap_init0_no_config(adap, reset); else { /* * The firmware provides us with a memory * buffer where we can load a Configuration * File from the host if we want to override * the Configuration File in flash. */ ret = adap_init0_config(adap, reset); if (ret == -ENOENT) { dev_info(adap->pdev_dev, "No Configuration File present " "on adapter. Using hard-wired " "configuration parameters.\n"); ret = adap_init0_no_config(adap, reset); } } } if (ret < 0) { dev_err(adap->pdev_dev, "could not initialize adapter, error %d\n", -ret); goto bye; } } /* * If we're living with non-hard-coded parameters (either from a * Firmware Configuration File or values programmed by a different PF * Driver), give the SGE code a chance to pull in anything that it * needs ... Note that this must be called after we retrieve our VPD * parameters in order to know how to convert core ticks to seconds. */ if (adap->flags & USING_SOFT_PARAMS) { ret = t4_sge_init(adap); if (ret < 0) goto bye; } if (is_bypass_device(adap->pdev->device)) adap->params.bypass = 1; /* * Grab some of our basic fundamental operating parameters. */ #define FW_PARAM_DEV(param) \ (FW_PARAMS_MNEM(FW_PARAMS_MNEM_DEV) | \ FW_PARAMS_PARAM_X(FW_PARAMS_PARAM_DEV_##param)) #define FW_PARAM_PFVF(param) \ FW_PARAMS_MNEM(FW_PARAMS_MNEM_PFVF) | \ FW_PARAMS_PARAM_X(FW_PARAMS_PARAM_PFVF_##param)| \ FW_PARAMS_PARAM_Y(0) | \ FW_PARAMS_PARAM_Z(0) params[0] = FW_PARAM_PFVF(EQ_START); params[1] = FW_PARAM_PFVF(L2T_START); params[2] = FW_PARAM_PFVF(L2T_END); params[3] = FW_PARAM_PFVF(FILTER_START); params[4] = FW_PARAM_PFVF(FILTER_END); params[5] = FW_PARAM_PFVF(IQFLINT_START); ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 6, params, val); if (ret < 0) goto bye; adap->sge.egr_start = val[0]; adap->l2t_start = val[1]; adap->l2t_end = val[2]; adap->tids.ftid_base = val[3]; adap->tids.nftids = val[4] - val[3] + 1; adap->sge.ingr_start = val[5]; /* query params related to active filter region */ params[0] = FW_PARAM_PFVF(ACTIVE_FILTER_START); params[1] = FW_PARAM_PFVF(ACTIVE_FILTER_END); ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 2, params, val); /* If Active filter size is set we enable establishing * offload connection through firmware work request */ if ((val[0] != val[1]) && (ret >= 0)) { adap->flags |= FW_OFLD_CONN; adap->tids.aftid_base = val[0]; adap->tids.aftid_end = val[1]; } /* If we're running on newer firmware, let it know that we're * prepared to deal with encapsulated CPL messages. Older * firmware won't understand this and we'll just get * unencapsulated messages ... */ params[0] = FW_PARAM_PFVF(CPLFW4MSG_ENCAP); val[0] = 1; (void) t4_set_params(adap, adap->mbox, adap->fn, 0, 1, params, val); /* * Find out whether we're allowed to use the T5+ ULPTX MEMWRITE DSGL * capability. Earlier versions of the firmware didn't have the * ULPTX_MEMWRITE_DSGL so we'll interpret a query failure as no * permission to use ULPTX MEMWRITE DSGL. */ if (is_t4(adap->params.chip)) { adap->params.ulptx_memwrite_dsgl = false; } else { params[0] = FW_PARAM_DEV(ULPTX_MEMWRITE_DSGL); ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 1, params, val); adap->params.ulptx_memwrite_dsgl = (ret == 0 && val[0] != 0); } /* * Get device capabilities so we can determine what resources we need * to manage. */ memset(&caps_cmd, 0, sizeof(caps_cmd)); caps_cmd.op_to_write = htonl(FW_CMD_OP(FW_CAPS_CONFIG_CMD) | FW_CMD_REQUEST | FW_CMD_READ); caps_cmd.cfvalid_to_len16 = htonl(FW_LEN16(caps_cmd)); ret = t4_wr_mbox(adap, adap->mbox, &caps_cmd, sizeof(caps_cmd), &caps_cmd); if (ret < 0) goto bye; if (caps_cmd.ofldcaps) { /* query offload-related parameters */ params[0] = FW_PARAM_DEV(NTID); params[1] = FW_PARAM_PFVF(SERVER_START); params[2] = FW_PARAM_PFVF(SERVER_END); params[3] = FW_PARAM_PFVF(TDDP_START); params[4] = FW_PARAM_PFVF(TDDP_END); params[5] = FW_PARAM_DEV(FLOWC_BUFFIFO_SZ); ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 6, params, val); if (ret < 0) goto bye; adap->tids.ntids = val[0]; adap->tids.natids = min(adap->tids.ntids / 2, MAX_ATIDS); adap->tids.stid_base = val[1]; adap->tids.nstids = val[2] - val[1] + 1; /* * Setup server filter region. Divide the availble filter * region into two parts. Regular filters get 1/3rd and server * filters get 2/3rd part. This is only enabled if workarond * path is enabled. * 1. For regular filters. * 2. Server filter: This are special filters which are used * to redirect SYN packets to offload queue. */ if (adap->flags & FW_OFLD_CONN && !is_bypass(adap)) { adap->tids.sftid_base = adap->tids.ftid_base + DIV_ROUND_UP(adap->tids.nftids, 3); adap->tids.nsftids = adap->tids.nftids - DIV_ROUND_UP(adap->tids.nftids, 3); adap->tids.nftids = adap->tids.sftid_base - adap->tids.ftid_base; } adap->vres.ddp.start = val[3]; adap->vres.ddp.size = val[4] - val[3] + 1; adap->params.ofldq_wr_cred = val[5]; adap->params.offload = 1; } if (caps_cmd.rdmacaps) { params[0] = FW_PARAM_PFVF(STAG_START); params[1] = FW_PARAM_PFVF(STAG_END); params[2] = FW_PARAM_PFVF(RQ_START); params[3] = FW_PARAM_PFVF(RQ_END); params[4] = FW_PARAM_PFVF(PBL_START); params[5] = FW_PARAM_PFVF(PBL_END); ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 6, params, val); if (ret < 0) goto bye; adap->vres.stag.start = val[0]; adap->vres.stag.size = val[1] - val[0] + 1; adap->vres.rq.start = val[2]; adap->vres.rq.size = val[3] - val[2] + 1; adap->vres.pbl.start = val[4]; adap->vres.pbl.size = val[5] - val[4] + 1; params[0] = FW_PARAM_PFVF(SQRQ_START); params[1] = FW_PARAM_PFVF(SQRQ_END); params[2] = FW_PARAM_PFVF(CQ_START); params[3] = FW_PARAM_PFVF(CQ_END); params[4] = FW_PARAM_PFVF(OCQ_START); params[5] = FW_PARAM_PFVF(OCQ_END); ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 6, params, val); if (ret < 0) goto bye; adap->vres.qp.start = val[0]; adap->vres.qp.size = val[1] - val[0] + 1; adap->vres.cq.start = val[2]; adap->vres.cq.size = val[3] - val[2] + 1; adap->vres.ocq.start = val[4]; adap->vres.ocq.size = val[5] - val[4] + 1; params[0] = FW_PARAM_DEV(MAXORDIRD_QP); params[1] = FW_PARAM_DEV(MAXIRD_ADAPTER); ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 2, params, val); if (ret < 0) { adap->params.max_ordird_qp = 8; adap->params.max_ird_adapter = 32 * adap->tids.ntids; ret = 0; } else { adap->params.max_ordird_qp = val[0]; adap->params.max_ird_adapter = val[1]; } dev_info(adap->pdev_dev, "max_ordird_qp %d max_ird_adapter %d\n", adap->params.max_ordird_qp, adap->params.max_ird_adapter); } if (caps_cmd.iscsicaps) { params[0] = FW_PARAM_PFVF(ISCSI_START); params[1] = FW_PARAM_PFVF(ISCSI_END); ret = t4_query_params(adap, adap->mbox, adap->fn, 0, 2, params, val); if (ret < 0) goto bye; adap->vres.iscsi.start = val[0]; adap->vres.iscsi.size = val[1] - val[0] + 1; } #undef FW_PARAM_PFVF #undef FW_PARAM_DEV /* The MTU/MSS Table is initialized by now, so load their values. If * we're initializing the adapter, then we'll make any modifications * we want to the MTU/MSS Table and also initialize the congestion * parameters. */ t4_read_mtu_tbl(adap, adap->params.mtus, NULL); if (state != DEV_STATE_INIT) { int i; /* The default MTU Table contains values 1492 and 1500. * However, for TCP, it's better to have two values which are * a multiple of 8 +/- 4 bytes apart near this popular MTU. * This allows us to have a TCP Data Payload which is a * multiple of 8 regardless of what combination of TCP Options * are in use (always a multiple of 4 bytes) which is * important for performance reasons. For instance, if no * options are in use, then we have a 20-byte IP header and a * 20-byte TCP header. In this case, a 1500-byte MSS would * result in a TCP Data Payload of 1500 - 40 == 1460 bytes * which is not a multiple of 8. So using an MSS of 1488 in * this case results in a TCP Data Payload of 1448 bytes which * is a multiple of 8. On the other hand, if 12-byte TCP Time * Stamps have been negotiated, then an MTU of 1500 bytes * results in a TCP Data Payload of 1448 bytes which, as * above, is a multiple of 8 bytes ... */ for (i = 0; i < NMTUS; i++) if (adap->params.mtus[i] == 1492) { adap->params.mtus[i] = 1488; break; } t4_load_mtus(adap, adap->params.mtus, adap->params.a_wnd, adap->params.b_wnd); } t4_init_tp_params(adap); adap->flags |= FW_OK; return 0; /* * Something bad happened. If a command timed out or failed with EIO * FW does not operate within its spec or something catastrophic * happened to HW/FW, stop issuing commands. */ bye: if (ret != -ETIMEDOUT && ret != -EIO) t4_fw_bye(adap, adap->mbox); return ret; } /* EEH callbacks */ static pci_ers_result_t eeh_err_detected(struct pci_dev *pdev, pci_channel_state_t state) { int i; struct adapter *adap = pci_get_drvdata(pdev); if (!adap) goto out; rtnl_lock(); adap->flags &= ~FW_OK; notify_ulds(adap, CXGB4_STATE_START_RECOVERY); spin_lock(&adap->stats_lock); for_each_port(adap, i) { struct net_device *dev = adap->port[i]; netif_device_detach(dev); netif_carrier_off(dev); } spin_unlock(&adap->stats_lock); if (adap->flags & FULL_INIT_DONE) cxgb_down(adap); rtnl_unlock(); if ((adap->flags & DEV_ENABLED)) { pci_disable_device(pdev); adap->flags &= ~DEV_ENABLED; } out: return state == pci_channel_io_perm_failure ? PCI_ERS_RESULT_DISCONNECT : PCI_ERS_RESULT_NEED_RESET; } static pci_ers_result_t eeh_slot_reset(struct pci_dev *pdev) { int i, ret; struct fw_caps_config_cmd c; struct adapter *adap = pci_get_drvdata(pdev); if (!adap) { pci_restore_state(pdev); pci_save_state(pdev); return PCI_ERS_RESULT_RECOVERED; } if (!(adap->flags & DEV_ENABLED)) { if (pci_enable_device(pdev)) { dev_err(&pdev->dev, "Cannot reenable PCI " "device after reset\n"); return PCI_ERS_RESULT_DISCONNECT; } adap->flags |= DEV_ENABLED; } pci_set_master(pdev); pci_restore_state(pdev); pci_save_state(pdev); pci_cleanup_aer_uncorrect_error_status(pdev); if (t4_wait_dev_ready(adap) < 0) return PCI_ERS_RESULT_DISCONNECT; if (t4_fw_hello(adap, adap->fn, adap->fn, MASTER_MUST, NULL) < 0) return PCI_ERS_RESULT_DISCONNECT; adap->flags |= FW_OK; if (adap_init1(adap, &c)) return PCI_ERS_RESULT_DISCONNECT; for_each_port(adap, i) { struct port_info *p = adap2pinfo(adap, i); ret = t4_alloc_vi(adap, adap->fn, p->tx_chan, adap->fn, 0, 1, NULL, NULL); if (ret < 0) return PCI_ERS_RESULT_DISCONNECT; p->viid = ret; p->xact_addr_filt = -1; } t4_load_mtus(adap, adap->params.mtus, adap->params.a_wnd, adap->params.b_wnd); setup_memwin(adap); if (cxgb_up(adap)) return PCI_ERS_RESULT_DISCONNECT; return PCI_ERS_RESULT_RECOVERED; } static void eeh_resume(struct pci_dev *pdev) { int i; struct adapter *adap = pci_get_drvdata(pdev); if (!adap) return; rtnl_lock(); for_each_port(adap, i) { struct net_device *dev = adap->port[i]; if (netif_running(dev)) { link_start(dev); cxgb_set_rxmode(dev); } netif_device_attach(dev); } rtnl_unlock(); } static const struct pci_error_handlers cxgb4_eeh = { .error_detected = eeh_err_detected, .slot_reset = eeh_slot_reset, .resume = eeh_resume, }; static inline bool is_x_10g_port(const struct link_config *lc) { return (lc->supported & FW_PORT_CAP_SPEED_10G) != 0 || (lc->supported & FW_PORT_CAP_SPEED_40G) != 0; } static inline void init_rspq(struct adapter *adap, struct sge_rspq *q, unsigned int us, unsigned int cnt, unsigned int size, unsigned int iqe_size) { q->adap = adap; set_rspq_intr_params(q, us, cnt); q->iqe_len = iqe_size; q->size = size; } /* * Perform default configuration of DMA queues depending on the number and type * of ports we found and the number of available CPUs. Most settings can be * modified by the admin prior to actual use. */ static void cfg_queues(struct adapter *adap) { struct sge *s = &adap->sge; int i, n10g = 0, qidx = 0; #ifndef CONFIG_CHELSIO_T4_DCB int q10g = 0; #endif int ciq_size; for_each_port(adap, i) n10g += is_x_10g_port(&adap2pinfo(adap, i)->link_cfg); #ifdef CONFIG_CHELSIO_T4_DCB /* For Data Center Bridging support we need to be able to support up * to 8 Traffic Priorities; each of which will be assigned to its * own TX Queue in order to prevent Head-Of-Line Blocking. */ if (adap->params.nports * 8 > MAX_ETH_QSETS) { dev_err(adap->pdev_dev, "MAX_ETH_QSETS=%d < %d!\n", MAX_ETH_QSETS, adap->params.nports * 8); BUG_ON(1); } for_each_port(adap, i) { struct port_info *pi = adap2pinfo(adap, i); pi->first_qset = qidx; pi->nqsets = 8; qidx += pi->nqsets; } #else /* !CONFIG_CHELSIO_T4_DCB */ /* * We default to 1 queue per non-10G port and up to # of cores queues * per 10G port. */ if (n10g) q10g = (MAX_ETH_QSETS - (adap->params.nports - n10g)) / n10g; if (q10g > netif_get_num_default_rss_queues()) q10g = netif_get_num_default_rss_queues(); for_each_port(adap, i) { struct port_info *pi = adap2pinfo(adap, i); pi->first_qset = qidx; pi->nqsets = is_x_10g_port(&pi->link_cfg) ? q10g : 1; qidx += pi->nqsets; } #endif /* !CONFIG_CHELSIO_T4_DCB */ s->ethqsets = qidx; s->max_ethqsets = qidx; /* MSI-X may lower it later */ if (is_offload(adap)) { /* * For offload we use 1 queue/channel if all ports are up to 1G, * otherwise we divide all available queues amongst the channels * capped by the number of available cores. */ if (n10g) { i = min_t(int, ARRAY_SIZE(s->ofldrxq), num_online_cpus()); s->ofldqsets = roundup(i, adap->params.nports); } else s->ofldqsets = adap->params.nports; /* For RDMA one Rx queue per channel suffices */ s->rdmaqs = adap->params.nports; s->rdmaciqs = adap->params.nports; } for (i = 0; i < ARRAY_SIZE(s->ethrxq); i++) { struct sge_eth_rxq *r = &s->ethrxq[i]; init_rspq(adap, &r->rspq, 5, 10, 1024, 64); r->fl.size = 72; } for (i = 0; i < ARRAY_SIZE(s->ethtxq); i++) s->ethtxq[i].q.size = 1024; for (i = 0; i < ARRAY_SIZE(s->ctrlq); i++) s->ctrlq[i].q.size = 512; for (i = 0; i < ARRAY_SIZE(s->ofldtxq); i++) s->ofldtxq[i].q.size = 1024; for (i = 0; i < ARRAY_SIZE(s->ofldrxq); i++) { struct sge_ofld_rxq *r = &s->ofldrxq[i]; init_rspq(adap, &r->rspq, 5, 1, 1024, 64); r->rspq.uld = CXGB4_ULD_ISCSI; r->fl.size = 72; } for (i = 0; i < ARRAY_SIZE(s->rdmarxq); i++) { struct sge_ofld_rxq *r = &s->rdmarxq[i]; init_rspq(adap, &r->rspq, 5, 1, 511, 64); r->rspq.uld = CXGB4_ULD_RDMA; r->fl.size = 72; } ciq_size = 64 + adap->vres.cq.size + adap->tids.nftids; if (ciq_size > SGE_MAX_IQ_SIZE) { CH_WARN(adap, "CIQ size too small for available IQs\n"); ciq_size = SGE_MAX_IQ_SIZE; } for (i = 0; i < ARRAY_SIZE(s->rdmaciq); i++) { struct sge_ofld_rxq *r = &s->rdmaciq[i]; init_rspq(adap, &r->rspq, 5, 1, ciq_size, 64); r->rspq.uld = CXGB4_ULD_RDMA; } init_rspq(adap, &s->fw_evtq, 0, 1, 1024, 64); init_rspq(adap, &s->intrq, 0, 1, 2 * MAX_INGQ, 64); } /* * Reduce the number of Ethernet queues across all ports to at most n. * n provides at least one queue per port. */ static void reduce_ethqs(struct adapter *adap, int n) { int i; struct port_info *pi; while (n < adap->sge.ethqsets) for_each_port(adap, i) { pi = adap2pinfo(adap, i); if (pi->nqsets > 1) { pi->nqsets--; adap->sge.ethqsets--; if (adap->sge.ethqsets <= n) break; } } n = 0; for_each_port(adap, i) { pi = adap2pinfo(adap, i); pi->first_qset = n; n += pi->nqsets; } } /* 2 MSI-X vectors needed for the FW queue and non-data interrupts */ #define EXTRA_VECS 2 static int enable_msix(struct adapter *adap) { int ofld_need = 0; int i, want, need; struct sge *s = &adap->sge; unsigned int nchan = adap->params.nports; struct msix_entry entries[MAX_INGQ + 1]; for (i = 0; i < ARRAY_SIZE(entries); ++i) entries[i].entry = i; want = s->max_ethqsets + EXTRA_VECS; if (is_offload(adap)) { want += s->rdmaqs + s->rdmaciqs + s->ofldqsets; /* need nchan for each possible ULD */ ofld_need = 3 * nchan; } #ifdef CONFIG_CHELSIO_T4_DCB /* For Data Center Bridging we need 8 Ethernet TX Priority Queues for * each port. */ need = 8 * adap->params.nports + EXTRA_VECS + ofld_need; #else need = adap->params.nports + EXTRA_VECS + ofld_need; #endif want = pci_enable_msix_range(adap->pdev, entries, need, want); if (want < 0) return want; /* * Distribute available vectors to the various queue groups. * Every group gets its minimum requirement and NIC gets top * priority for leftovers. */ i = want - EXTRA_VECS - ofld_need; if (i < s->max_ethqsets) { s->max_ethqsets = i; if (i < s->ethqsets) reduce_ethqs(adap, i); } if (is_offload(adap)) { i = want - EXTRA_VECS - s->max_ethqsets; i -= ofld_need - nchan; s->ofldqsets = (i / nchan) * nchan; /* round down */ } for (i = 0; i < want; ++i) adap->msix_info[i].vec = entries[i].vector; return 0; } #undef EXTRA_VECS static int init_rss(struct adapter *adap) { unsigned int i, j; for_each_port(adap, i) { struct port_info *pi = adap2pinfo(adap, i); pi->rss = kcalloc(pi->rss_size, sizeof(u16), GFP_KERNEL); if (!pi->rss) return -ENOMEM; for (j = 0; j < pi->rss_size; j++) pi->rss[j] = ethtool_rxfh_indir_default(j, pi->nqsets); } return 0; } static void print_port_info(const struct net_device *dev) { char buf[80]; char *bufp = buf; const char *spd = ""; const struct port_info *pi = netdev_priv(dev); const struct adapter *adap = pi->adapter; if (adap->params.pci.speed == PCI_EXP_LNKSTA_CLS_2_5GB) spd = " 2.5 GT/s"; else if (adap->params.pci.speed == PCI_EXP_LNKSTA_CLS_5_0GB) spd = " 5 GT/s"; else if (adap->params.pci.speed == PCI_EXP_LNKSTA_CLS_8_0GB) spd = " 8 GT/s"; if (pi->link_cfg.supported & FW_PORT_CAP_SPEED_100M) bufp += sprintf(bufp, "100/"); if (pi->link_cfg.supported & FW_PORT_CAP_SPEED_1G) bufp += sprintf(bufp, "1000/"); if (pi->link_cfg.supported & FW_PORT_CAP_SPEED_10G) bufp += sprintf(bufp, "10G/"); if (pi->link_cfg.supported & FW_PORT_CAP_SPEED_40G) bufp += sprintf(bufp, "40G/"); if (bufp != buf) --bufp; sprintf(bufp, "BASE-%s", t4_get_port_type_description(pi->port_type)); netdev_info(dev, "Chelsio %s rev %d %s %sNIC PCIe x%d%s%s\n", adap->params.vpd.id, CHELSIO_CHIP_RELEASE(adap->params.chip), buf, is_offload(adap) ? "R" : "", adap->params.pci.width, spd, (adap->flags & USING_MSIX) ? " MSI-X" : (adap->flags & USING_MSI) ? " MSI" : ""); netdev_info(dev, "S/N: %s, P/N: %s\n", adap->params.vpd.sn, adap->params.vpd.pn); } static void enable_pcie_relaxed_ordering(struct pci_dev *dev) { pcie_capability_set_word(dev, PCI_EXP_DEVCTL, PCI_EXP_DEVCTL_RELAX_EN); } /* * Free the following resources: * - memory used for tables * - MSI/MSI-X * - net devices * - resources FW is holding for us */ static void free_some_resources(struct adapter *adapter) { unsigned int i; t4_free_mem(adapter->l2t); t4_free_mem(adapter->tids.tid_tab); disable_msi(adapter); for_each_port(adapter, i) if (adapter->port[i]) { kfree(adap2pinfo(adapter, i)->rss); free_netdev(adapter->port[i]); } if (adapter->flags & FW_OK) t4_fw_bye(adapter, adapter->fn); } #define TSO_FLAGS (NETIF_F_TSO | NETIF_F_TSO6 | NETIF_F_TSO_ECN) #define VLAN_FEAT (NETIF_F_SG | NETIF_F_IP_CSUM | TSO_FLAGS | \ NETIF_F_IPV6_CSUM | NETIF_F_HIGHDMA) #define SEGMENT_SIZE 128 static int init_one(struct pci_dev *pdev, const struct pci_device_id *ent) { int func, i, err, s_qpp, qpp, num_seg; struct port_info *pi; bool highdma = false; struct adapter *adapter = NULL; void __iomem *regs; printk_once(KERN_INFO "%s - version %s\n", DRV_DESC, DRV_VERSION); err = pci_request_regions(pdev, KBUILD_MODNAME); if (err) { /* Just info, some other driver may have claimed the device. */ dev_info(&pdev->dev, "cannot obtain PCI resources\n"); return err; } err = pci_enable_device(pdev); if (err) { dev_err(&pdev->dev, "cannot enable PCI device\n"); goto out_release_regions; } regs = pci_ioremap_bar(pdev, 0); if (!regs) { dev_err(&pdev->dev, "cannot map device registers\n"); err = -ENOMEM; goto out_disable_device; } /* We control everything through one PF */ func = SOURCEPF_GET(readl(regs + PL_WHOAMI)); if (func != ent->driver_data) { iounmap(regs); pci_disable_device(pdev); pci_save_state(pdev); /* to restore SR-IOV later */ goto sriov; } if (!pci_set_dma_mask(pdev, DMA_BIT_MASK(64))) { highdma = true; err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64)); if (err) { dev_err(&pdev->dev, "unable to obtain 64-bit DMA for " "coherent allocations\n"); goto out_unmap_bar0; } } else { err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32)); if (err) { dev_err(&pdev->dev, "no usable DMA configuration\n"); goto out_unmap_bar0; } } pci_enable_pcie_error_reporting(pdev); enable_pcie_relaxed_ordering(pdev); pci_set_master(pdev); pci_save_state(pdev); adapter = kzalloc(sizeof(*adapter), GFP_KERNEL); if (!adapter) { err = -ENOMEM; goto out_unmap_bar0; } adapter->workq = create_singlethread_workqueue("cxgb4"); if (!adapter->workq) { err = -ENOMEM; goto out_free_adapter; } /* PCI device has been enabled */ adapter->flags |= DEV_ENABLED; adapter->regs = regs; adapter->pdev = pdev; adapter->pdev_dev = &pdev->dev; adapter->mbox = func; adapter->fn = func; adapter->msg_enable = dflt_msg_enable; memset(adapter->chan_map, 0xff, sizeof(adapter->chan_map)); spin_lock_init(&adapter->stats_lock); spin_lock_init(&adapter->tid_release_lock); INIT_WORK(&adapter->tid_release_task, process_tid_release_list); INIT_WORK(&adapter->db_full_task, process_db_full); INIT_WORK(&adapter->db_drop_task, process_db_drop); err = t4_prep_adapter(adapter); if (err) goto out_free_adapter; if (!is_t4(adapter->params.chip)) { s_qpp = QUEUESPERPAGEPF1 * adapter->fn; qpp = 1 << QUEUESPERPAGEPF0_GET(t4_read_reg(adapter, SGE_EGRESS_QUEUES_PER_PAGE_PF) >> s_qpp); num_seg = PAGE_SIZE / SEGMENT_SIZE; /* Each segment size is 128B. Write coalescing is enabled only * when SGE_EGRESS_QUEUES_PER_PAGE_PF reg value for the * queue is less no of segments that can be accommodated in * a page size. */ if (qpp > num_seg) { dev_err(&pdev->dev, "Incorrect number of egress queues per page\n"); err = -EINVAL; goto out_free_adapter; } adapter->bar2 = ioremap_wc(pci_resource_start(pdev, 2), pci_resource_len(pdev, 2)); if (!adapter->bar2) { dev_err(&pdev->dev, "cannot map device bar2 region\n"); err = -ENOMEM; goto out_free_adapter; } } setup_memwin(adapter); err = adap_init0(adapter); setup_memwin_rdma(adapter); if (err) goto out_unmap_bar; for_each_port(adapter, i) { struct net_device *netdev; netdev = alloc_etherdev_mq(sizeof(struct port_info), MAX_ETH_QSETS); if (!netdev) { err = -ENOMEM; goto out_free_dev; } SET_NETDEV_DEV(netdev, &pdev->dev); adapter->port[i] = netdev; pi = netdev_priv(netdev); pi->adapter = adapter; pi->xact_addr_filt = -1; pi->port_id = i; netdev->irq = pdev->irq; netdev->hw_features = NETIF_F_SG | TSO_FLAGS | NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM | NETIF_F_RXCSUM | NETIF_F_RXHASH | NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX; if (highdma) netdev->hw_features |= NETIF_F_HIGHDMA; netdev->features |= netdev->hw_features; netdev->vlan_features = netdev->features & VLAN_FEAT; netdev->priv_flags |= IFF_UNICAST_FLT; netdev->netdev_ops = &cxgb4_netdev_ops; #ifdef CONFIG_CHELSIO_T4_DCB netdev->dcbnl_ops = &cxgb4_dcb_ops; cxgb4_dcb_state_init(netdev); #endif netdev->ethtool_ops = &cxgb_ethtool_ops; } pci_set_drvdata(pdev, adapter); if (adapter->flags & FW_OK) { err = t4_port_init(adapter, func, func, 0); if (err) goto out_free_dev; } /* * Configure queues and allocate tables now, they can be needed as * soon as the first register_netdev completes. */ cfg_queues(adapter); adapter->l2t = t4_init_l2t(); if (!adapter->l2t) { /* We tolerate a lack of L2T, giving up some functionality */ dev_warn(&pdev->dev, "could not allocate L2T, continuing\n"); adapter->params.offload = 0; } if (is_offload(adapter) && tid_init(&adapter->tids) < 0) { dev_warn(&pdev->dev, "could not allocate TID table, " "continuing\n"); adapter->params.offload = 0; } /* See what interrupts we'll be using */ if (msi > 1 && enable_msix(adapter) == 0) adapter->flags |= USING_MSIX; else if (msi > 0 && pci_enable_msi(pdev) == 0) adapter->flags |= USING_MSI; err = init_rss(adapter); if (err) goto out_free_dev; /* * The card is now ready to go. If any errors occur during device * registration we do not fail the whole card but rather proceed only * with the ports we manage to register successfully. However we must * register at least one net device. */ for_each_port(adapter, i) { pi = adap2pinfo(adapter, i); netif_set_real_num_tx_queues(adapter->port[i], pi->nqsets); netif_set_real_num_rx_queues(adapter->port[i], pi->nqsets); err = register_netdev(adapter->port[i]); if (err) break; adapter->chan_map[pi->tx_chan] = i; print_port_info(adapter->port[i]); } if (i == 0) { dev_err(&pdev->dev, "could not register any net devices\n"); goto out_free_dev; } if (err) { dev_warn(&pdev->dev, "only %d net devices registered\n", i); err = 0; } if (cxgb4_debugfs_root) { adapter->debugfs_root = debugfs_create_dir(pci_name(pdev), cxgb4_debugfs_root); setup_debugfs(adapter); } /* PCIe EEH recovery on powerpc platforms needs fundamental reset */ pdev->needs_freset = 1; if (is_offload(adapter)) attach_ulds(adapter); sriov: #ifdef CONFIG_PCI_IOV if (func < ARRAY_SIZE(num_vf) && num_vf[func] > 0) if (pci_enable_sriov(pdev, num_vf[func]) == 0) dev_info(&pdev->dev, "instantiated %u virtual functions\n", num_vf[func]); #endif return 0; out_free_dev: free_some_resources(adapter); out_unmap_bar: if (!is_t4(adapter->params.chip)) iounmap(adapter->bar2); out_free_adapter: if (adapter->workq) destroy_workqueue(adapter->workq); kfree(adapter); out_unmap_bar0: iounmap(regs); out_disable_device: pci_disable_pcie_error_reporting(pdev); pci_disable_device(pdev); out_release_regions: pci_release_regions(pdev); return err; } static void remove_one(struct pci_dev *pdev) { struct adapter *adapter = pci_get_drvdata(pdev); #ifdef CONFIG_PCI_IOV pci_disable_sriov(pdev); #endif if (adapter) { int i; /* Tear down per-adapter Work Queue first since it can contain * references to our adapter data structure. */ destroy_workqueue(adapter->workq); if (is_offload(adapter)) detach_ulds(adapter); for_each_port(adapter, i) if (adapter->port[i]->reg_state == NETREG_REGISTERED) unregister_netdev(adapter->port[i]); debugfs_remove_recursive(adapter->debugfs_root); /* If we allocated filters, free up state associated with any * valid filters ... */ if (adapter->tids.ftid_tab) { struct filter_entry *f = &adapter->tids.ftid_tab[0]; for (i = 0; i < (adapter->tids.nftids + adapter->tids.nsftids); i++, f++) if (f->valid) clear_filter(adapter, f); } if (adapter->flags & FULL_INIT_DONE) cxgb_down(adapter); free_some_resources(adapter); iounmap(adapter->regs); if (!is_t4(adapter->params.chip)) iounmap(adapter->bar2); pci_disable_pcie_error_reporting(pdev); if ((adapter->flags & DEV_ENABLED)) { pci_disable_device(pdev); adapter->flags &= ~DEV_ENABLED; } pci_release_regions(pdev); synchronize_rcu(); kfree(adapter); } else pci_release_regions(pdev); } static struct pci_driver cxgb4_driver = { .name = KBUILD_MODNAME, .id_table = cxgb4_pci_tbl, .probe = init_one, .remove = remove_one, .shutdown = remove_one, .err_handler = &cxgb4_eeh, }; static int __init cxgb4_init_module(void) { int ret; /* Debugfs support is optional, just warn if this fails */ cxgb4_debugfs_root = debugfs_create_dir(KBUILD_MODNAME, NULL); if (!cxgb4_debugfs_root) pr_warn("could not create debugfs entry, continuing\n"); ret = pci_register_driver(&cxgb4_driver); if (ret < 0) debugfs_remove(cxgb4_debugfs_root); register_inet6addr_notifier(&cxgb4_inet6addr_notifier); return ret; } static void __exit cxgb4_cleanup_module(void) { unregister_inet6addr_notifier(&cxgb4_inet6addr_notifier); pci_unregister_driver(&cxgb4_driver); debugfs_remove(cxgb4_debugfs_root); /* NULL ok */ } module_init(cxgb4_init_module); module_exit(cxgb4_cleanup_module);