]> bbs.cooldavid.org Git - net-next-2.6.git/blobdiff - drivers/net/ixgbe/ixgbe_main.c
net: use netdev_mc_count and netdev_mc_empty when appropriate
[net-next-2.6.git] / drivers / net / ixgbe / ixgbe_main.c
index 9c9202f40b10655c74899d4f80401172728204a0..0792f151de99f299a4f8a5eaac85c0c43dffd42c 100644 (file)
 #include "ixgbe.h"
 #include "ixgbe_common.h"
 #include "ixgbe_dcb_82599.h"
+#include "ixgbe_sriov.h"
 
 char ixgbe_driver_name[] = "ixgbe";
 static const char ixgbe_driver_string[] =
                               "Intel(R) 10 Gigabit PCI Express Network Driver";
 
-#define DRV_VERSION "2.0.44-k2"
+#define DRV_VERSION "2.0.62-k2"
 const char ixgbe_driver_version[] = DRV_VERSION;
 static char ixgbe_copyright[] = "Copyright (c) 1999-2010 Intel Corporation.";
 
@@ -67,7 +68,7 @@ static const struct ixgbe_info *ixgbe_info_tbl[] = {
  * { Vendor ID, Device ID, SubVendor ID, SubDevice ID,
  *   Class, Class Mask, private data (not used) }
  */
-static struct pci_device_id ixgbe_pci_tbl[] = {
+static DEFINE_PCI_DEVICE_TABLE(ixgbe_pci_tbl) = {
        {PCI_VDEVICE(INTEL, IXGBE_DEV_ID_82598),
         board_82598 },
        {PCI_VDEVICE(INTEL, IXGBE_DEV_ID_82598AF_DUAL_PORT),
@@ -124,6 +125,13 @@ static struct notifier_block dca_notifier = {
 };
 #endif
 
+#ifdef CONFIG_PCI_IOV
+static unsigned int max_vfs;
+module_param(max_vfs, uint, 0);
+MODULE_PARM_DESC(max_vfs, "Maximum number of virtual functions to allocate "
+                 "per physical function");
+#endif /* CONFIG_PCI_IOV */
+
 MODULE_AUTHOR("Intel Corporation, <linux.nics@intel.com>");
 MODULE_DESCRIPTION("Intel(R) 10 Gigabit PCI Express Network Driver");
 MODULE_LICENSE("GPL");
@@ -131,6 +139,41 @@ MODULE_VERSION(DRV_VERSION);
 
 #define DEFAULT_DEBUG_LEVEL_SHIFT 3
 
+static inline void ixgbe_disable_sriov(struct ixgbe_adapter *adapter)
+{
+       struct ixgbe_hw *hw = &adapter->hw;
+       u32 gcr;
+       u32 gpie;
+       u32 vmdctl;
+
+#ifdef CONFIG_PCI_IOV
+       /* disable iov and allow time for transactions to clear */
+       pci_disable_sriov(adapter->pdev);
+#endif
+
+       /* turn off device IOV mode */
+       gcr = IXGBE_READ_REG(hw, IXGBE_GCR_EXT);
+       gcr &= ~(IXGBE_GCR_EXT_SRIOV);
+       IXGBE_WRITE_REG(hw, IXGBE_GCR_EXT, gcr);
+       gpie = IXGBE_READ_REG(hw, IXGBE_GPIE);
+       gpie &= ~IXGBE_GPIE_VTMODE_MASK;
+       IXGBE_WRITE_REG(hw, IXGBE_GPIE, gpie);
+
+       /* set default pool back to 0 */
+       vmdctl = IXGBE_READ_REG(hw, IXGBE_VT_CTL);
+       vmdctl &= ~IXGBE_VT_CTL_POOL_MASK;
+       IXGBE_WRITE_REG(hw, IXGBE_VT_CTL, vmdctl);
+
+       /* take a breather then clean up driver data */
+       msleep(100);
+       if (adapter->vfinfo)
+               kfree(adapter->vfinfo);
+       adapter->vfinfo = NULL;
+
+       adapter->num_vfs = 0;
+       adapter->flags &= ~IXGBE_FLAG_SRIOV_ENABLED;
+}
+
 static void ixgbe_release_hw_control(struct ixgbe_adapter *adapter)
 {
        u32 ctrl_ext;
@@ -451,7 +494,7 @@ static void ixgbe_update_rx_dca(struct ixgbe_adapter *adapter,
 {
        u32 rxctrl;
        int cpu = get_cpu();
-       int q = rx_ring - adapter->rx_ring;
+       int q = rx_ring->reg_idx;
 
        if (rx_ring->cpu != cpu) {
                rxctrl = IXGBE_READ_REG(&adapter->hw, IXGBE_DCA_RXCTRL(q));
@@ -479,7 +522,7 @@ static void ixgbe_update_tx_dca(struct ixgbe_adapter *adapter,
 {
        u32 txctrl;
        int cpu = get_cpu();
-       int q = tx_ring - adapter->tx_ring;
+       int q = tx_ring->reg_idx;
        struct ixgbe_hw *hw = &adapter->hw;
 
        if (tx_ring->cpu != cpu) {
@@ -513,12 +556,12 @@ static void ixgbe_setup_dca(struct ixgbe_adapter *adapter)
        IXGBE_WRITE_REG(&adapter->hw, IXGBE_DCA_CTRL, 2);
 
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               adapter->tx_ring[i].cpu = -1;
-               ixgbe_update_tx_dca(adapter, &adapter->tx_ring[i]);
+               adapter->tx_ring[i]->cpu = -1;
+               ixgbe_update_tx_dca(adapter, adapter->tx_ring[i]);
        }
        for (i = 0; i < adapter->num_rx_queues; i++) {
-               adapter->rx_ring[i].cpu = -1;
-               ixgbe_update_rx_dca(adapter, &adapter->rx_ring[i]);
+               adapter->rx_ring[i]->cpu = -1;
+               ixgbe_update_rx_dca(adapter, adapter->rx_ring[i]);
        }
 }
 
@@ -989,7 +1032,7 @@ static void ixgbe_configure_msix(struct ixgbe_adapter *adapter)
                                       adapter->num_rx_queues);
 
                for (i = 0; i < q_vector->rxr_count; i++) {
-                       j = adapter->rx_ring[r_idx].reg_idx;
+                       j = adapter->rx_ring[r_idx]->reg_idx;
                        ixgbe_set_ivar(adapter, 0, j, v_idx);
                        r_idx = find_next_bit(q_vector->rxr_idx,
                                              adapter->num_rx_queues,
@@ -999,7 +1042,7 @@ static void ixgbe_configure_msix(struct ixgbe_adapter *adapter)
                                       adapter->num_tx_queues);
 
                for (i = 0; i < q_vector->txr_count; i++) {
-                       j = adapter->tx_ring[r_idx].reg_idx;
+                       j = adapter->tx_ring[r_idx]->reg_idx;
                        ixgbe_set_ivar(adapter, 1, j, v_idx);
                        r_idx = find_next_bit(q_vector->txr_idx,
                                              adapter->num_tx_queues,
@@ -1025,7 +1068,12 @@ static void ixgbe_configure_msix(struct ixgbe_adapter *adapter)
 
        /* set up to autoclear timer, and the vectors */
        mask = IXGBE_EIMS_ENABLE_MASK;
-       mask &= ~(IXGBE_EIMS_OTHER | IXGBE_EIMS_LSC);
+       if (adapter->num_vfs)
+               mask &= ~(IXGBE_EIMS_OTHER |
+                         IXGBE_EIMS_MAILBOX |
+                         IXGBE_EIMS_LSC);
+       else
+               mask &= ~(IXGBE_EIMS_OTHER | IXGBE_EIMS_LSC);
        IXGBE_WRITE_REG(&adapter->hw, IXGBE_EIAC, mask);
 }
 
@@ -1134,7 +1182,7 @@ static void ixgbe_set_itr_msix(struct ixgbe_q_vector *q_vector)
 
        r_idx = find_first_bit(q_vector->txr_idx, adapter->num_tx_queues);
        for (i = 0; i < q_vector->txr_count; i++) {
-               tx_ring = &(adapter->tx_ring[r_idx]);
+               tx_ring = adapter->tx_ring[r_idx];
                ret_itr = ixgbe_update_itr(adapter, q_vector->eitr,
                                           q_vector->tx_itr,
                                           tx_ring->total_packets,
@@ -1149,7 +1197,7 @@ static void ixgbe_set_itr_msix(struct ixgbe_q_vector *q_vector)
 
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
        for (i = 0; i < q_vector->rxr_count; i++) {
-               rx_ring = &(adapter->rx_ring[r_idx]);
+               rx_ring = adapter->rx_ring[r_idx];
                ret_itr = ixgbe_update_itr(adapter, q_vector->eitr,
                                           q_vector->rx_itr,
                                           rx_ring->total_packets,
@@ -1254,6 +1302,9 @@ static irqreturn_t ixgbe_msix_lsc(int irq, void *data)
        if (eicr & IXGBE_EICR_LSC)
                ixgbe_check_lsc(adapter);
 
+       if (eicr & IXGBE_EICR_MAILBOX)
+               ixgbe_msg_task(adapter);
+
        if (hw->mac.type == ixgbe_mac_82598EB)
                ixgbe_check_fan_failure(adapter, eicr);
 
@@ -1268,7 +1319,7 @@ static irqreturn_t ixgbe_msix_lsc(int irq, void *data)
                        netif_tx_stop_all_queues(netdev);
                        for (i = 0; i < adapter->num_tx_queues; i++) {
                                struct ixgbe_ring *tx_ring =
-                                                          &adapter->tx_ring[i];
+                                                           adapter->tx_ring[i];
                                if (test_and_clear_bit(__IXGBE_FDIR_INIT_DONE,
                                                       &tx_ring->reinit_state))
                                        schedule_work(&adapter->fdir_reinit_task);
@@ -1327,7 +1378,7 @@ static irqreturn_t ixgbe_msix_clean_tx(int irq, void *data)
 
        r_idx = find_first_bit(q_vector->txr_idx, adapter->num_tx_queues);
        for (i = 0; i < q_vector->txr_count; i++) {
-               tx_ring = &(adapter->tx_ring[r_idx]);
+               tx_ring = adapter->tx_ring[r_idx];
                tx_ring->total_bytes = 0;
                tx_ring->total_packets = 0;
                r_idx = find_next_bit(q_vector->txr_idx, adapter->num_tx_queues,
@@ -1355,7 +1406,7 @@ static irqreturn_t ixgbe_msix_clean_rx(int irq, void *data)
 
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
        for (i = 0;  i < q_vector->rxr_count; i++) {
-               rx_ring = &(adapter->rx_ring[r_idx]);
+               rx_ring = adapter->rx_ring[r_idx];
                rx_ring->total_bytes = 0;
                rx_ring->total_packets = 0;
                r_idx = find_next_bit(q_vector->rxr_idx, adapter->num_rx_queues,
@@ -1385,7 +1436,7 @@ static irqreturn_t ixgbe_msix_clean_many(int irq, void *data)
 
        r_idx = find_first_bit(q_vector->txr_idx, adapter->num_tx_queues);
        for (i = 0; i < q_vector->txr_count; i++) {
-               ring = &(adapter->tx_ring[r_idx]);
+               ring = adapter->tx_ring[r_idx];
                ring->total_bytes = 0;
                ring->total_packets = 0;
                r_idx = find_next_bit(q_vector->txr_idx, adapter->num_tx_queues,
@@ -1394,7 +1445,7 @@ static irqreturn_t ixgbe_msix_clean_many(int irq, void *data)
 
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
        for (i = 0; i < q_vector->rxr_count; i++) {
-               ring = &(adapter->rx_ring[r_idx]);
+               ring = adapter->rx_ring[r_idx];
                ring->total_bytes = 0;
                ring->total_packets = 0;
                r_idx = find_next_bit(q_vector->rxr_idx, adapter->num_rx_queues,
@@ -1425,7 +1476,7 @@ static int ixgbe_clean_rxonly(struct napi_struct *napi, int budget)
        long r_idx;
 
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
-       rx_ring = &(adapter->rx_ring[r_idx]);
+       rx_ring = adapter->rx_ring[r_idx];
 #ifdef CONFIG_IXGBE_DCA
        if (adapter->flags & IXGBE_FLAG_DCA_ENABLED)
                ixgbe_update_rx_dca(adapter, rx_ring);
@@ -1466,7 +1517,7 @@ static int ixgbe_clean_rxtx_many(struct napi_struct *napi, int budget)
 
        r_idx = find_first_bit(q_vector->txr_idx, adapter->num_tx_queues);
        for (i = 0; i < q_vector->txr_count; i++) {
-               ring = &(adapter->tx_ring[r_idx]);
+               ring = adapter->tx_ring[r_idx];
 #ifdef CONFIG_IXGBE_DCA
                if (adapter->flags & IXGBE_FLAG_DCA_ENABLED)
                        ixgbe_update_tx_dca(adapter, ring);
@@ -1482,7 +1533,7 @@ static int ixgbe_clean_rxtx_many(struct napi_struct *napi, int budget)
        budget = max(budget, 1);
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
        for (i = 0; i < q_vector->rxr_count; i++) {
-               ring = &(adapter->rx_ring[r_idx]);
+               ring = adapter->rx_ring[r_idx];
 #ifdef CONFIG_IXGBE_DCA
                if (adapter->flags & IXGBE_FLAG_DCA_ENABLED)
                        ixgbe_update_rx_dca(adapter, ring);
@@ -1493,7 +1544,7 @@ static int ixgbe_clean_rxtx_many(struct napi_struct *napi, int budget)
        }
 
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
-       ring = &(adapter->rx_ring[r_idx]);
+       ring = adapter->rx_ring[r_idx];
        /* If all Rx work done, exit the polling mode */
        if (work_done < budget) {
                napi_complete(napi);
@@ -1526,7 +1577,7 @@ static int ixgbe_clean_txonly(struct napi_struct *napi, int budget)
        long r_idx;
 
        r_idx = find_first_bit(q_vector->txr_idx, adapter->num_tx_queues);
-       tx_ring = &(adapter->tx_ring[r_idx]);
+       tx_ring = adapter->tx_ring[r_idx];
 #ifdef CONFIG_IXGBE_DCA
        if (adapter->flags & IXGBE_FLAG_DCA_ENABLED)
                ixgbe_update_tx_dca(adapter, tx_ring);
@@ -1711,8 +1762,8 @@ static void ixgbe_set_itr(struct ixgbe_adapter *adapter)
        struct ixgbe_q_vector *q_vector = adapter->q_vector[0];
        u8 current_itr;
        u32 new_itr = q_vector->eitr;
-       struct ixgbe_ring *rx_ring = &adapter->rx_ring[0];
-       struct ixgbe_ring *tx_ring = &adapter->tx_ring[0];
+       struct ixgbe_ring *rx_ring = adapter->rx_ring[0];
+       struct ixgbe_ring *tx_ring = adapter->tx_ring[0];
 
        q_vector->tx_itr = ixgbe_update_itr(adapter, new_itr,
                                            q_vector->tx_itr,
@@ -1768,6 +1819,8 @@ static inline void ixgbe_irq_enable(struct ixgbe_adapter *adapter)
                mask |= IXGBE_EIMS_ECC;
                mask |= IXGBE_EIMS_GPI_SDP1;
                mask |= IXGBE_EIMS_GPI_SDP2;
+               if (adapter->num_vfs)
+                       mask |= IXGBE_EIMS_MAILBOX;
        }
        if (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE ||
            adapter->flags & IXGBE_FLAG_FDIR_PERFECT_CAPABLE)
@@ -1776,6 +1829,11 @@ static inline void ixgbe_irq_enable(struct ixgbe_adapter *adapter)
        IXGBE_WRITE_REG(&adapter->hw, IXGBE_EIMS, mask);
        ixgbe_irq_enable_queues(adapter, ~0);
        IXGBE_WRITE_FLUSH(&adapter->hw);
+
+       if (adapter->num_vfs > 32) {
+               u32 eitrsel = (1 << (adapter->num_vfs - 32)) - 1;
+               IXGBE_WRITE_REG(&adapter->hw, IXGBE_EITRSEL, eitrsel);
+       }
 }
 
 /**
@@ -1817,10 +1875,10 @@ static irqreturn_t ixgbe_intr(int irq, void *data)
        ixgbe_check_fan_failure(adapter, eicr);
 
        if (napi_schedule_prep(&(q_vector->napi))) {
-               adapter->tx_ring[0].total_packets = 0;
-               adapter->tx_ring[0].total_bytes = 0;
-               adapter->rx_ring[0].total_packets = 0;
-               adapter->rx_ring[0].total_bytes = 0;
+               adapter->tx_ring[0]->total_packets = 0;
+               adapter->tx_ring[0]->total_bytes = 0;
+               adapter->rx_ring[0]->total_packets = 0;
+               adapter->rx_ring[0]->total_bytes = 0;
                /* would disable interrupts here but EIAM disabled it */
                __napi_schedule(&(q_vector->napi));
        }
@@ -1905,6 +1963,8 @@ static inline void ixgbe_irq_disable(struct ixgbe_adapter *adapter)
                IXGBE_WRITE_REG(&adapter->hw, IXGBE_EIMC, 0xFFFF0000);
                IXGBE_WRITE_REG(&adapter->hw, IXGBE_EIMC_EX(0), ~0);
                IXGBE_WRITE_REG(&adapter->hw, IXGBE_EIMC_EX(1), ~0);
+               if (adapter->num_vfs > 32)
+                       IXGBE_WRITE_REG(&adapter->hw, IXGBE_EITRSEL, 0);
        }
        IXGBE_WRITE_FLUSH(&adapter->hw);
        if (adapter->flags & IXGBE_FLAG_MSIX_ENABLED) {
@@ -1950,7 +2010,7 @@ static void ixgbe_configure_tx(struct ixgbe_adapter *adapter)
 
        /* Setup the HW Tx Head and Tail descriptor pointers */
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               struct ixgbe_ring *ring = &adapter->tx_ring[i];
+               struct ixgbe_ring *ring = adapter->tx_ring[i];
                j = ring->reg_idx;
                tdba = ring->dma;
                tdlen = ring->count * sizeof(union ixgbe_adv_tx_desc);
@@ -1960,8 +2020,8 @@ static void ixgbe_configure_tx(struct ixgbe_adapter *adapter)
                IXGBE_WRITE_REG(hw, IXGBE_TDLEN(j), tdlen);
                IXGBE_WRITE_REG(hw, IXGBE_TDH(j), 0);
                IXGBE_WRITE_REG(hw, IXGBE_TDT(j), 0);
-               adapter->tx_ring[i].head = IXGBE_TDH(j);
-               adapter->tx_ring[i].tail = IXGBE_TDT(j);
+               adapter->tx_ring[i]->head = IXGBE_TDH(j);
+               adapter->tx_ring[i]->tail = IXGBE_TDT(j);
                /*
                 * Disable Tx Head Writeback RO bit, since this hoses
                 * bookkeeping if things aren't delivered in order.
@@ -1989,18 +2049,32 @@ static void ixgbe_configure_tx(struct ixgbe_adapter *adapter)
 
        if (hw->mac.type == ixgbe_mac_82599EB) {
                u32 rttdcs;
+               u32 mask;
 
                /* disable the arbiter while setting MTQC */
                rttdcs = IXGBE_READ_REG(hw, IXGBE_RTTDCS);
                rttdcs |= IXGBE_RTTDCS_ARBDIS;
                IXGBE_WRITE_REG(hw, IXGBE_RTTDCS, rttdcs);
 
-               /* We enable 8 traffic classes, DCB only */
-               if (adapter->flags & IXGBE_FLAG_DCB_ENABLED)
-                       IXGBE_WRITE_REG(hw, IXGBE_MTQC, (IXGBE_MTQC_RT_ENA |
-                                       IXGBE_MTQC_8TC_8TQ));
-               else
+               /* set transmit pool layout */
+               mask = (IXGBE_FLAG_SRIOV_ENABLED | IXGBE_FLAG_DCB_ENABLED);
+               switch (adapter->flags & mask) {
+
+               case (IXGBE_FLAG_SRIOV_ENABLED):
+                       IXGBE_WRITE_REG(hw, IXGBE_MTQC,
+                                       (IXGBE_MTQC_VT_ENA | IXGBE_MTQC_64VF));
+                       break;
+
+               case (IXGBE_FLAG_DCB_ENABLED):
+                       /* We enable 8 traffic classes, DCB only */
+                       IXGBE_WRITE_REG(hw, IXGBE_MTQC,
+                                     (IXGBE_MTQC_RT_ENA | IXGBE_MTQC_8TC_8TQ));
+                       break;
+
+               default:
                        IXGBE_WRITE_REG(hw, IXGBE_MTQC, IXGBE_MTQC_64Q_1PB);
+                       break;
+               }
 
                /* re-eable the arbiter */
                rttdcs &= ~IXGBE_RTTDCS_ARBDIS;
@@ -2059,12 +2133,16 @@ static u32 ixgbe_setup_mrqc(struct ixgbe_adapter *adapter)
 #ifdef CONFIG_IXGBE_DCB
                                 | IXGBE_FLAG_DCB_ENABLED
 #endif
+                                | IXGBE_FLAG_SRIOV_ENABLED
                                );
 
        switch (mask) {
        case (IXGBE_FLAG_RSS_ENABLED):
                mrqc = IXGBE_MRQC_RSSEN;
                break;
+       case (IXGBE_FLAG_SRIOV_ENABLED):
+               mrqc = IXGBE_MRQC_VMDQEN;
+               break;
 #ifdef CONFIG_IXGBE_DCB
        case (IXGBE_FLAG_DCB_ENABLED):
                mrqc = IXGBE_MRQC_RT8TCEN;
@@ -2090,7 +2168,7 @@ static void ixgbe_configure_rscctl(struct ixgbe_adapter *adapter, int index)
        u32 rscctrl;
        int rx_buf_len;
 
-       rx_ring = &adapter->rx_ring[index];
+       rx_ring = adapter->rx_ring[index];
        j = rx_ring->reg_idx;
        rx_buf_len = rx_ring->rx_buf_len;
        rscctrl = IXGBE_READ_REG(hw, IXGBE_RSCCTL(j));
@@ -2145,7 +2223,9 @@ static void ixgbe_configure_rx(struct ixgbe_adapter *adapter)
        int rx_buf_len;
 
        /* Decide whether to use packet split mode or not */
-       adapter->flags |= IXGBE_FLAG_RX_PS_ENABLED;
+       /* Do not use packet split if we're in SR-IOV Mode */
+       if (!adapter->num_vfs)
+               adapter->flags |= IXGBE_FLAG_RX_PS_ENABLED;
 
        /* Set the RX buffer length according to the mode */
        if (adapter->flags & IXGBE_FLAG_RX_PS_ENABLED) {
@@ -2157,7 +2237,9 @@ static void ixgbe_configure_rx(struct ixgbe_adapter *adapter)
                                      IXGBE_PSRTYPE_IPV4HDR |
                                      IXGBE_PSRTYPE_IPV6HDR |
                                      IXGBE_PSRTYPE_L2HDR;
-                       IXGBE_WRITE_REG(hw, IXGBE_PSRTYPE(0), psrtype);
+                       IXGBE_WRITE_REG(hw,
+                                       IXGBE_PSRTYPE(adapter->num_vfs),
+                                       psrtype);
                }
        } else {
                if (!(adapter->flags2 & IXGBE_FLAG2_RSC_ENABLED) &&
@@ -2184,7 +2266,7 @@ static void ixgbe_configure_rx(struct ixgbe_adapter *adapter)
 #endif
        IXGBE_WRITE_REG(hw, IXGBE_HLREG0, hlreg0);
 
-       rdlen = adapter->rx_ring[0].count * sizeof(union ixgbe_adv_rx_desc);
+       rdlen = adapter->rx_ring[0]->count * sizeof(union ixgbe_adv_rx_desc);
        /* disable receives while setting up the descriptors */
        rxctrl = IXGBE_READ_REG(hw, IXGBE_RXCTRL);
        IXGBE_WRITE_REG(hw, IXGBE_RXCTRL, rxctrl & ~IXGBE_RXCTRL_RXEN);
@@ -2194,7 +2276,7 @@ static void ixgbe_configure_rx(struct ixgbe_adapter *adapter)
         * the Base and Length of the Rx Descriptor Ring
         */
        for (i = 0; i < adapter->num_rx_queues; i++) {
-               rx_ring = &adapter->rx_ring[i];
+               rx_ring = adapter->rx_ring[i];
                rdba = rx_ring->dma;
                j = rx_ring->reg_idx;
                IXGBE_WRITE_REG(hw, IXGBE_RDBAL(j), (rdba & DMA_BIT_MASK(32)));
@@ -2243,6 +2325,30 @@ static void ixgbe_configure_rx(struct ixgbe_adapter *adapter)
                IXGBE_WRITE_REG(hw, IXGBE_RDRXCTL, rdrxctl);
        }
 
+       if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED) {
+               u32 vt_reg_bits;
+               u32 reg_offset, vf_shift;
+               u32 vmdctl = IXGBE_READ_REG(hw, IXGBE_VT_CTL);
+               vt_reg_bits = IXGBE_VMD_CTL_VMDQ_EN
+                       | IXGBE_VT_CTL_REPLEN;
+               vt_reg_bits |= (adapter->num_vfs <<
+                               IXGBE_VT_CTL_POOL_SHIFT);
+               IXGBE_WRITE_REG(hw, IXGBE_VT_CTL, vmdctl | vt_reg_bits);
+               IXGBE_WRITE_REG(hw, IXGBE_MRQC, 0);
+
+               vf_shift = adapter->num_vfs % 32;
+               reg_offset = adapter->num_vfs / 32;
+               IXGBE_WRITE_REG(hw, IXGBE_VFRE(0), 0);
+               IXGBE_WRITE_REG(hw, IXGBE_VFRE(1), 0);
+               IXGBE_WRITE_REG(hw, IXGBE_VFTE(0), 0);
+               IXGBE_WRITE_REG(hw, IXGBE_VFTE(1), 0);
+               /* Enable only the PF's pool for Tx/Rx */
+               IXGBE_WRITE_REG(hw, IXGBE_VFRE(reg_offset), (1 << vf_shift));
+               IXGBE_WRITE_REG(hw, IXGBE_VFTE(reg_offset), (1 << vf_shift));
+               IXGBE_WRITE_REG(hw, IXGBE_PFDTXGSWC, IXGBE_PFDTXGSWC_VT_LBEN);
+               ixgbe_set_vmolr(hw, adapter->num_vfs);
+       }
+
        /* Program MRQC for the distribution of queues */
        mrqc = ixgbe_setup_mrqc(adapter);
 
@@ -2274,6 +2380,20 @@ static void ixgbe_configure_rx(struct ixgbe_adapter *adapter)
        }
        IXGBE_WRITE_REG(hw, IXGBE_MRQC, mrqc);
 
+       if (adapter->num_vfs) {
+               u32 reg;
+
+               /* Map PF MAC address in RAR Entry 0 to first pool
+                * following VFs */
+               hw->mac.ops.set_vmdq(hw, 0, adapter->num_vfs);
+
+               /* Set up VF register offsets for selected VT Mode, i.e.
+                * 64 VFs for SR-IOV */
+               reg = IXGBE_READ_REG(hw, IXGBE_GCR_EXT);
+               reg |= IXGBE_GCR_EXT_SRIOV;
+               IXGBE_WRITE_REG(hw, IXGBE_GCR_EXT, reg);
+       }
+
        rxcsum = IXGBE_READ_REG(hw, IXGBE_RXCSUM);
 
        if (adapter->flags & IXGBE_FLAG_RSS_ENABLED ||
@@ -2312,15 +2432,17 @@ static void ixgbe_vlan_rx_add_vid(struct net_device *netdev, u16 vid)
 {
        struct ixgbe_adapter *adapter = netdev_priv(netdev);
        struct ixgbe_hw *hw = &adapter->hw;
+       int pool_ndx = adapter->num_vfs;
 
        /* add VID to filter table */
-       hw->mac.ops.set_vfta(&adapter->hw, vid, 0, true);
+       hw->mac.ops.set_vfta(&adapter->hw, vid, pool_ndx, true);
 }
 
 static void ixgbe_vlan_rx_kill_vid(struct net_device *netdev, u16 vid)
 {
        struct ixgbe_adapter *adapter = netdev_priv(netdev);
        struct ixgbe_hw *hw = &adapter->hw;
+       int pool_ndx = adapter->num_vfs;
 
        if (!test_bit(__IXGBE_DOWN, &adapter->state))
                ixgbe_irq_disable(adapter);
@@ -2331,7 +2453,7 @@ static void ixgbe_vlan_rx_kill_vid(struct net_device *netdev, u16 vid)
                ixgbe_irq_enable(adapter);
 
        /* remove VID from filter table */
-       hw->mac.ops.set_vfta(&adapter->hw, vid, 0, false);
+       hw->mac.ops.set_vfta(&adapter->hw, vid, pool_ndx, false);
 }
 
 static void ixgbe_vlan_rx_register(struct net_device *netdev,
@@ -2361,7 +2483,7 @@ static void ixgbe_vlan_rx_register(struct net_device *netdev,
        } else if (adapter->hw.mac.type == ixgbe_mac_82599EB) {
                for (i = 0; i < adapter->num_rx_queues; i++) {
                        u32 ctrl;
-                       j = adapter->rx_ring[i].reg_idx;
+                       j = adapter->rx_ring[i]->reg_idx;
                        ctrl = IXGBE_READ_REG(&adapter->hw, IXGBE_RXDCTL(j));
                        ctrl |= IXGBE_RXDCTL_VME;
                        IXGBE_WRITE_REG(&adapter->hw, IXGBE_RXDCTL(j), ctrl);
@@ -2414,7 +2536,7 @@ static u8 *ixgbe_addr_list_itr(struct ixgbe_hw *hw, u8 **mc_addr_ptr, u32 *vmdq)
  * responsible for configuring the hardware for proper unicast, multicast and
  * promiscuous mode.
  **/
-static void ixgbe_set_rx_mode(struct net_device *netdev)
+void ixgbe_set_rx_mode(struct net_device *netdev)
 {
        struct ixgbe_adapter *adapter = netdev_priv(netdev);
        struct ixgbe_hw *hw = &adapter->hw;
@@ -2446,14 +2568,16 @@ static void ixgbe_set_rx_mode(struct net_device *netdev)
        IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
 
        /* reprogram secondary unicast list */
-       hw->mac.ops.update_uc_addr_list(hw, &netdev->uc.list);
+       hw->mac.ops.update_uc_addr_list(hw, netdev);
 
        /* reprogram multicast list */
-       addr_count = netdev->mc_count;
+       addr_count = netdev_mc_count(netdev);
        if (addr_count)
                addr_list = netdev->mc_list->dmi_addr;
        hw->mac.ops.update_mc_addr_list(hw, addr_list, addr_count,
                                        ixgbe_addr_list_itr);
+       if (adapter->num_vfs)
+               ixgbe_restore_vf_multicasts(adapter);
 }
 
 static void ixgbe_napi_enable_all(struct ixgbe_adapter *adapter)
@@ -2522,7 +2646,7 @@ static void ixgbe_configure_dcb(struct ixgbe_adapter *adapter)
        ixgbe_dcb_hw_config(&adapter->hw, &adapter->dcb_cfg);
 
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               j = adapter->tx_ring[i].reg_idx;
+               j = adapter->tx_ring[i]->reg_idx;
                txdctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(j));
                /* PThresh workaround for Tx hang with DFP enabled. */
                txdctl |= 32;
@@ -2539,7 +2663,7 @@ static void ixgbe_configure_dcb(struct ixgbe_adapter *adapter)
                vlnctrl &= ~IXGBE_VLNCTRL_CFIEN;
                IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
                for (i = 0; i < adapter->num_rx_queues; i++) {
-                       j = adapter->rx_ring[i].reg_idx;
+                       j = adapter->rx_ring[i]->reg_idx;
                        vlnctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(j));
                        vlnctrl |= IXGBE_RXDCTL_VME;
                        IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(j), vlnctrl);
@@ -2579,7 +2703,7 @@ static void ixgbe_configure(struct ixgbe_adapter *adapter)
 #endif /* IXGBE_FCOE */
        if (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE) {
                for (i = 0; i < adapter->num_tx_queues; i++)
-                       adapter->tx_ring[i].atr_sample_rate =
+                       adapter->tx_ring[i]->atr_sample_rate =
                                                       adapter->atr_sample_rate;
                ixgbe_init_fdir_signature_82599(hw, adapter->fdir_pballoc);
        } else if (adapter->flags & IXGBE_FLAG_FDIR_PERFECT_CAPABLE) {
@@ -2589,8 +2713,8 @@ static void ixgbe_configure(struct ixgbe_adapter *adapter)
        ixgbe_configure_tx(adapter);
        ixgbe_configure_rx(adapter);
        for (i = 0; i < adapter->num_rx_queues; i++)
-               ixgbe_alloc_rx_buffers(adapter, &adapter->rx_ring[i],
-                                      (adapter->rx_ring[i].count - 1));
+               ixgbe_alloc_rx_buffers(adapter, adapter->rx_ring[i],
+                                      (adapter->rx_ring[i]->count - 1));
 }
 
 static inline bool ixgbe_is_sfp(struct ixgbe_hw *hw)
@@ -2673,7 +2797,7 @@ link_cfg_out:
 static inline void ixgbe_rx_desc_queue_enable(struct ixgbe_adapter *adapter,
                                              int rxr)
 {
-       int j = adapter->rx_ring[rxr].reg_idx;
+       int j = adapter->rx_ring[rxr]->reg_idx;
        int k;
 
        for (k = 0; k < IXGBE_MAX_RX_DESC_POLL; k++) {
@@ -2687,8 +2811,8 @@ static inline void ixgbe_rx_desc_queue_enable(struct ixgbe_adapter *adapter,
                DPRINTK(DRV, ERR, "RXDCTL.ENABLE on Rx queue %d "
                        "not set within the polling period\n", rxr);
        }
-       ixgbe_release_rx_desc(&adapter->hw, &adapter->rx_ring[rxr],
-                             (adapter->rx_ring[rxr].count - 1));
+       ixgbe_release_rx_desc(&adapter->hw, adapter->rx_ring[rxr],
+                             (adapter->rx_ring[rxr]->count - 1));
 }
 
 static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
@@ -2702,6 +2826,7 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
        u32 txdctl, rxdctl, mhadd;
        u32 dmatxctl;
        u32 gpie;
+       u32 ctrl_ext;
 
        ixgbe_get_hw_control(adapter);
 
@@ -2714,6 +2839,10 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
                        /* MSI only */
                        gpie = 0;
                }
+               if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED) {
+                       gpie &= ~IXGBE_GPIE_VTMODE_MASK;
+                       gpie |= IXGBE_GPIE_VTMODE_64;
+               }
                /* XXX: to interrupt immediately for EICS writes, enable this */
                /* gpie |= IXGBE_GPIE_EIMEN; */
                IXGBE_WRITE_REG(hw, IXGBE_GPIE, gpie);
@@ -2770,7 +2899,7 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
        }
 
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               j = adapter->tx_ring[i].reg_idx;
+               j = adapter->tx_ring[i]->reg_idx;
                txdctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(j));
                /* enable WTHRESH=8 descriptors, to encourage burst writeback */
                txdctl |= (8 << 16);
@@ -2784,14 +2913,26 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
                IXGBE_WRITE_REG(hw, IXGBE_DMATXCTL, dmatxctl);
        }
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               j = adapter->tx_ring[i].reg_idx;
+               j = adapter->tx_ring[i]->reg_idx;
                txdctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(j));
                txdctl |= IXGBE_TXDCTL_ENABLE;
                IXGBE_WRITE_REG(hw, IXGBE_TXDCTL(j), txdctl);
+               if (hw->mac.type == ixgbe_mac_82599EB) {
+                       int wait_loop = 10;
+                       /* poll for Tx Enable ready */
+                       do {
+                               msleep(1);
+                               txdctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(j));
+                       } while (--wait_loop &&
+                                !(txdctl & IXGBE_TXDCTL_ENABLE));
+                       if (!wait_loop)
+                               DPRINTK(DRV, ERR, "Could not enable "
+                                       "Tx Queue %d\n", j);
+               }
        }
 
        for (i = 0; i < num_rx_rings; i++) {
-               j = adapter->rx_ring[i].reg_idx;
+               j = adapter->rx_ring[i]->reg_idx;
                rxdctl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(j));
                /* enable PTHRESH=32 descriptors (half the internal cache)
                 * and HTHRESH=0 descriptors (to minimize latency on fetch),
@@ -2865,7 +3006,7 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
 
        for (i = 0; i < adapter->num_tx_queues; i++)
                set_bit(__IXGBE_FDIR_INIT_DONE,
-                       &(adapter->tx_ring[i].reinit_state));
+                       &(adapter->tx_ring[i]->reinit_state));
 
        /* enable transmits */
        netif_tx_start_all_queues(netdev);
@@ -2875,6 +3016,12 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
        adapter->flags |= IXGBE_FLAG_NEED_LINK_UPDATE;
        adapter->link_check_timeout = jiffies;
        mod_timer(&adapter->watchdog_timer, jiffies);
+
+       /* Set PF Reset Done bit so PF/VF Mail Ops can work */
+       ctrl_ext = IXGBE_READ_REG(hw, IXGBE_CTRL_EXT);
+       ctrl_ext |= IXGBE_CTRL_EXT_PFRSTD;
+       IXGBE_WRITE_REG(hw, IXGBE_CTRL_EXT, ctrl_ext);
+
        return 0;
 }
 
@@ -2923,7 +3070,8 @@ void ixgbe_reset(struct ixgbe_adapter *adapter)
        }
 
        /* reprogram the RAR[0] in case user changed it. */
-       hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, IXGBE_RAH_AV);
+       hw->mac.ops.set_rar(hw, 0, hw->mac.addr, adapter->num_vfs,
+                           IXGBE_RAH_AV);
 }
 
 /**
@@ -3029,7 +3177,7 @@ static void ixgbe_clean_all_rx_rings(struct ixgbe_adapter *adapter)
        int i;
 
        for (i = 0; i < adapter->num_rx_queues; i++)
-               ixgbe_clean_rx_ring(adapter, &adapter->rx_ring[i]);
+               ixgbe_clean_rx_ring(adapter, adapter->rx_ring[i]);
 }
 
 /**
@@ -3041,7 +3189,7 @@ static void ixgbe_clean_all_tx_rings(struct ixgbe_adapter *adapter)
        int i;
 
        for (i = 0; i < adapter->num_tx_queues; i++)
-               ixgbe_clean_tx_ring(adapter, &adapter->tx_ring[i]);
+               ixgbe_clean_tx_ring(adapter, adapter->tx_ring[i]);
 }
 
 void ixgbe_down(struct ixgbe_adapter *adapter)
@@ -3055,6 +3203,17 @@ void ixgbe_down(struct ixgbe_adapter *adapter)
        /* signal that we are down to the interrupt handler */
        set_bit(__IXGBE_DOWN, &adapter->state);
 
+       /* disable receive for all VFs and wait one second */
+       if (adapter->num_vfs) {
+               for (i = 0 ; i < adapter->num_vfs; i++)
+                       adapter->vfinfo[i].clear_to_send = 0;
+
+               /* ping all the active vfs to let them know we are going down */
+               ixgbe_ping_all_vfs(adapter);
+               /* Disable all VFTE/VFRE TX/RX */
+               ixgbe_disable_tx_rx(adapter);
+       }
+
        /* disable receives */
        rxctrl = IXGBE_READ_REG(hw, IXGBE_RXCTRL);
        IXGBE_WRITE_REG(hw, IXGBE_RXCTRL, rxctrl & ~IXGBE_RXCTRL_RXEN);
@@ -3081,7 +3240,7 @@ void ixgbe_down(struct ixgbe_adapter *adapter)
 
        /* disable transmits in the hardware now that interrupts are off */
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               j = adapter->tx_ring[i].reg_idx;
+               j = adapter->tx_ring[i]->reg_idx;
                txdctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(j));
                IXGBE_WRITE_REG(hw, IXGBE_TXDCTL(j),
                                (txdctl & ~IXGBE_TXDCTL_ENABLE));
@@ -3094,6 +3253,9 @@ void ixgbe_down(struct ixgbe_adapter *adapter)
 
        netif_carrier_off(netdev);
 
+       /* clear n-tuple filters that are cached */
+       ethtool_ntuple_flush(netdev);
+
        if (!pci_channel_offline(adapter->pdev))
                ixgbe_reset(adapter);
        ixgbe_clean_all_tx_rings(adapter);
@@ -3121,13 +3283,13 @@ static int ixgbe_poll(struct napi_struct *napi, int budget)
 
 #ifdef CONFIG_IXGBE_DCA
        if (adapter->flags & IXGBE_FLAG_DCA_ENABLED) {
-               ixgbe_update_tx_dca(adapter, adapter->tx_ring);
-               ixgbe_update_rx_dca(adapter, adapter->rx_ring);
+               ixgbe_update_tx_dca(adapter, adapter->tx_ring[0]);
+               ixgbe_update_rx_dca(adapter, adapter->rx_ring[0]);
        }
 #endif
 
-       tx_clean_complete = ixgbe_clean_tx_irq(q_vector, adapter->tx_ring);
-       ixgbe_clean_rx_irq(q_vector, adapter->rx_ring, &work_done, budget);
+       tx_clean_complete = ixgbe_clean_tx_irq(q_vector, adapter->tx_ring[0]);
+       ixgbe_clean_rx_irq(q_vector, adapter->rx_ring[0], &work_done, budget);
 
        if (!tx_clean_complete)
                work_done = budget;
@@ -3291,6 +3453,19 @@ static inline bool ixgbe_set_fcoe_queues(struct ixgbe_adapter *adapter)
 }
 
 #endif /* IXGBE_FCOE */
+/**
+ * ixgbe_set_sriov_queues: Allocate queues for IOV use
+ * @adapter: board private structure to initialize
+ *
+ * IOV doesn't actually use anything, so just NAK the
+ * request for now and let the other queue routines
+ * figure out what to do.
+ */
+static inline bool ixgbe_set_sriov_queues(struct ixgbe_adapter *adapter)
+{
+       return false;
+}
+
 /*
  * ixgbe_set_num_queues: Allocate queues for device, feature dependant
  * @adapter: board private structure to initialize
@@ -3304,6 +3479,15 @@ static inline bool ixgbe_set_fcoe_queues(struct ixgbe_adapter *adapter)
  **/
 static void ixgbe_set_num_queues(struct ixgbe_adapter *adapter)
 {
+       /* Start with base case */
+       adapter->num_rx_queues = 1;
+       adapter->num_tx_queues = 1;
+       adapter->num_rx_pools = adapter->num_rx_queues;
+       adapter->num_rx_queues_per_pool = 1;
+
+       if (ixgbe_set_sriov_queues(adapter))
+               return;
+
 #ifdef IXGBE_FCOE
        if (ixgbe_set_fcoe_queues(adapter))
                goto done;
@@ -3393,9 +3577,9 @@ static inline bool ixgbe_cache_ring_rss(struct ixgbe_adapter *adapter)
 
        if (adapter->flags & IXGBE_FLAG_RSS_ENABLED) {
                for (i = 0; i < adapter->num_rx_queues; i++)
-                       adapter->rx_ring[i].reg_idx = i;
+                       adapter->rx_ring[i]->reg_idx = i;
                for (i = 0; i < adapter->num_tx_queues; i++)
-                       adapter->tx_ring[i].reg_idx = i;
+                       adapter->tx_ring[i]->reg_idx = i;
                ret = true;
        } else {
                ret = false;
@@ -3422,8 +3606,8 @@ static inline bool ixgbe_cache_ring_dcb(struct ixgbe_adapter *adapter)
                if (adapter->hw.mac.type == ixgbe_mac_82598EB) {
                        /* the number of queues is assumed to be symmetric */
                        for (i = 0; i < dcb_i; i++) {
-                               adapter->rx_ring[i].reg_idx = i << 3;
-                               adapter->tx_ring[i].reg_idx = i << 2;
+                               adapter->rx_ring[i]->reg_idx = i << 3;
+                               adapter->tx_ring[i]->reg_idx = i << 2;
                        }
                        ret = true;
                } else if (adapter->hw.mac.type == ixgbe_mac_82599EB) {
@@ -3441,18 +3625,18 @@ static inline bool ixgbe_cache_ring_dcb(struct ixgbe_adapter *adapter)
                                 * Rx TC0-TC7 are offset by 16 queues each
                                 */
                                for (i = 0; i < 3; i++) {
-                                       adapter->tx_ring[i].reg_idx = i << 5;
-                                       adapter->rx_ring[i].reg_idx = i << 4;
+                                       adapter->tx_ring[i]->reg_idx = i << 5;
+                                       adapter->rx_ring[i]->reg_idx = i << 4;
                                }
                                for ( ; i < 5; i++) {
-                                       adapter->tx_ring[i].reg_idx =
+                                       adapter->tx_ring[i]->reg_idx =
                                                                 ((i + 2) << 4);
-                                       adapter->rx_ring[i].reg_idx = i << 4;
+                                       adapter->rx_ring[i]->reg_idx = i << 4;
                                }
                                for ( ; i < dcb_i; i++) {
-                                       adapter->tx_ring[i].reg_idx =
+                                       adapter->tx_ring[i]->reg_idx =
                                                                 ((i + 8) << 3);
-                                       adapter->rx_ring[i].reg_idx = i << 4;
+                                       adapter->rx_ring[i]->reg_idx = i << 4;
                                }
 
                                ret = true;
@@ -3465,12 +3649,12 @@ static inline bool ixgbe_cache_ring_dcb(struct ixgbe_adapter *adapter)
                                 *
                                 * Rx TC0-TC3 are offset by 32 queues each
                                 */
-                               adapter->tx_ring[0].reg_idx = 0;
-                               adapter->tx_ring[1].reg_idx = 64;
-                               adapter->tx_ring[2].reg_idx = 96;
-                               adapter->tx_ring[3].reg_idx = 112;
+                               adapter->tx_ring[0]->reg_idx = 0;
+                               adapter->tx_ring[1]->reg_idx = 64;
+                               adapter->tx_ring[2]->reg_idx = 96;
+                               adapter->tx_ring[3]->reg_idx = 112;
                                for (i = 0 ; i < dcb_i; i++)
-                                       adapter->rx_ring[i].reg_idx = i << 5;
+                                       adapter->rx_ring[i]->reg_idx = i << 5;
 
                                ret = true;
                        } else {
@@ -3503,9 +3687,9 @@ static bool inline ixgbe_cache_ring_fdir(struct ixgbe_adapter *adapter)
            ((adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE) ||
             (adapter->flags & IXGBE_FLAG_FDIR_PERFECT_CAPABLE))) {
                for (i = 0; i < adapter->num_rx_queues; i++)
-                       adapter->rx_ring[i].reg_idx = i;
+                       adapter->rx_ring[i]->reg_idx = i;
                for (i = 0; i < adapter->num_tx_queues; i++)
-                       adapter->tx_ring[i].reg_idx = i;
+                       adapter->tx_ring[i]->reg_idx = i;
                ret = true;
        }
 
@@ -3533,8 +3717,8 @@ static inline bool ixgbe_cache_ring_fcoe(struct ixgbe_adapter *adapter)
 
                        ixgbe_cache_ring_dcb(adapter);
                        /* find out queues in TC for FCoE */
-                       fcoe_rx_i = adapter->rx_ring[fcoe->tc].reg_idx + 1;
-                       fcoe_tx_i = adapter->tx_ring[fcoe->tc].reg_idx + 1;
+                       fcoe_rx_i = adapter->rx_ring[fcoe->tc]->reg_idx + 1;
+                       fcoe_tx_i = adapter->tx_ring[fcoe->tc]->reg_idx + 1;
                        /*
                         * In 82599, the number of Tx queues for each traffic
                         * class for both 8-TC and 4-TC modes are:
@@ -3565,8 +3749,8 @@ static inline bool ixgbe_cache_ring_fcoe(struct ixgbe_adapter *adapter)
                        fcoe_tx_i = f->mask;
                }
                for (i = 0; i < f->indices; i++, fcoe_rx_i++, fcoe_tx_i++) {
-                       adapter->rx_ring[f->mask + i].reg_idx = fcoe_rx_i;
-                       adapter->tx_ring[f->mask + i].reg_idx = fcoe_tx_i;
+                       adapter->rx_ring[f->mask + i]->reg_idx = fcoe_rx_i;
+                       adapter->tx_ring[f->mask + i]->reg_idx = fcoe_tx_i;
                }
                ret = true;
        }
@@ -3574,6 +3758,24 @@ static inline bool ixgbe_cache_ring_fcoe(struct ixgbe_adapter *adapter)
 }
 
 #endif /* IXGBE_FCOE */
+/**
+ * ixgbe_cache_ring_sriov - Descriptor ring to register mapping for sriov
+ * @adapter: board private structure to initialize
+ *
+ * SR-IOV doesn't use any descriptor rings but changes the default if
+ * no other mapping is used.
+ *
+ */
+static inline bool ixgbe_cache_ring_sriov(struct ixgbe_adapter *adapter)
+{
+       adapter->rx_ring[0]->reg_idx = adapter->num_vfs * 2;
+       adapter->tx_ring[0]->reg_idx = adapter->num_vfs * 2;
+       if (adapter->num_vfs)
+               return true;
+       else
+               return false;
+}
+
 /**
  * ixgbe_cache_ring_register - Descriptor ring to register mapping
  * @adapter: board private structure to initialize
@@ -3588,8 +3790,11 @@ static inline bool ixgbe_cache_ring_fcoe(struct ixgbe_adapter *adapter)
 static void ixgbe_cache_ring_register(struct ixgbe_adapter *adapter)
 {
        /* start with default case */
-       adapter->rx_ring[0].reg_idx = 0;
-       adapter->tx_ring[0].reg_idx = 0;
+       adapter->rx_ring[0]->reg_idx = 0;
+       adapter->tx_ring[0]->reg_idx = 0;
+
+       if (ixgbe_cache_ring_sriov(adapter))
+               return;
 
 #ifdef IXGBE_FCOE
        if (ixgbe_cache_ring_fcoe(adapter))
@@ -3619,33 +3824,63 @@ static void ixgbe_cache_ring_register(struct ixgbe_adapter *adapter)
 static int ixgbe_alloc_queues(struct ixgbe_adapter *adapter)
 {
        int i;
-
-       adapter->tx_ring = kcalloc(adapter->num_tx_queues,
-                                  sizeof(struct ixgbe_ring), GFP_KERNEL);
-       if (!adapter->tx_ring)
-               goto err_tx_ring_allocation;
-
-       adapter->rx_ring = kcalloc(adapter->num_rx_queues,
-                                  sizeof(struct ixgbe_ring), GFP_KERNEL);
-       if (!adapter->rx_ring)
-               goto err_rx_ring_allocation;
+       int orig_node = adapter->node;
 
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               adapter->tx_ring[i].count = adapter->tx_ring_count;
-               adapter->tx_ring[i].queue_index = i;
+               struct ixgbe_ring *ring = adapter->tx_ring[i];
+               if (orig_node == -1) {
+                       int cur_node = next_online_node(adapter->node);
+                       if (cur_node == MAX_NUMNODES)
+                               cur_node = first_online_node;
+                       adapter->node = cur_node;
+               }
+               ring = kzalloc_node(sizeof(struct ixgbe_ring), GFP_KERNEL,
+                                   adapter->node);
+               if (!ring)
+                       ring = kzalloc(sizeof(struct ixgbe_ring), GFP_KERNEL);
+               if (!ring)
+                       goto err_tx_ring_allocation;
+               ring->count = adapter->tx_ring_count;
+               ring->queue_index = i;
+               ring->numa_node = adapter->node;
+
+               adapter->tx_ring[i] = ring;
        }
 
+       /* Restore the adapter's original node */
+       adapter->node = orig_node;
+
        for (i = 0; i < adapter->num_rx_queues; i++) {
-               adapter->rx_ring[i].count = adapter->rx_ring_count;
-               adapter->rx_ring[i].queue_index = i;
+               struct ixgbe_ring *ring = adapter->rx_ring[i];
+               if (orig_node == -1) {
+                       int cur_node = next_online_node(adapter->node);
+                       if (cur_node == MAX_NUMNODES)
+                               cur_node = first_online_node;
+                       adapter->node = cur_node;
+               }
+               ring = kzalloc_node(sizeof(struct ixgbe_ring), GFP_KERNEL,
+                                   adapter->node);
+               if (!ring)
+                       ring = kzalloc(sizeof(struct ixgbe_ring), GFP_KERNEL);
+               if (!ring)
+                       goto err_rx_ring_allocation;
+               ring->count = adapter->rx_ring_count;
+               ring->queue_index = i;
+               ring->numa_node = adapter->node;
+
+               adapter->rx_ring[i] = ring;
        }
 
+       /* Restore the adapter's original node */
+       adapter->node = orig_node;
+
        ixgbe_cache_ring_register(adapter);
 
        return 0;
 
 err_rx_ring_allocation:
-       kfree(adapter->tx_ring);
+       for (i = 0; i < adapter->num_tx_queues; i++)
+               kfree(adapter->tx_ring[i]);
 err_tx_ring_allocation:
        return -ENOMEM;
 }
@@ -3700,6 +3935,9 @@ static int ixgbe_set_interrupt_capability(struct ixgbe_adapter *adapter)
        adapter->flags &= ~IXGBE_FLAG_FDIR_HASH_CAPABLE;
        adapter->flags &= ~IXGBE_FLAG_FDIR_PERFECT_CAPABLE;
        adapter->atr_sample_rate = 0;
+       if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED)
+               ixgbe_disable_sriov(adapter);
+
        ixgbe_set_num_queues(adapter);
 
        err = pci_enable_msi(adapter->pdev);
@@ -3741,7 +3979,11 @@ static int ixgbe_alloc_q_vectors(struct ixgbe_adapter *adapter)
        }
 
        for (q_idx = 0; q_idx < num_q_vectors; q_idx++) {
-               q_vector = kzalloc(sizeof(struct ixgbe_q_vector), GFP_KERNEL);
+               q_vector = kzalloc_node(sizeof(struct ixgbe_q_vector),
+                                       GFP_KERNEL, adapter->node);
+               if (!q_vector)
+                       q_vector = kzalloc(sizeof(struct ixgbe_q_vector),
+                                          GFP_KERNEL);
                if (!q_vector)
                        goto err_out;
                q_vector->adapter = adapter;
@@ -3868,10 +4110,16 @@ err_set_interrupt:
  **/
 void ixgbe_clear_interrupt_scheme(struct ixgbe_adapter *adapter)
 {
-       kfree(adapter->tx_ring);
-       kfree(adapter->rx_ring);
-       adapter->tx_ring = NULL;
-       adapter->rx_ring = NULL;
+       int i;
+
+       for (i = 0; i < adapter->num_tx_queues; i++) {
+               kfree(adapter->tx_ring[i]);
+               adapter->tx_ring[i] = NULL;
+       }
+       for (i = 0; i < adapter->num_rx_queues; i++) {
+               kfree(adapter->rx_ring[i]);
+               adapter->rx_ring[i] = NULL;
+       }
 
        ixgbe_free_q_vectors(adapter);
        ixgbe_reset_interrupt_capability(adapter);
@@ -3942,6 +4190,7 @@ static int __devinit ixgbe_sw_init(struct ixgbe_adapter *adapter)
 {
        struct ixgbe_hw *hw = &adapter->hw;
        struct pci_dev *pdev = adapter->pdev;
+       struct net_device *dev = adapter->netdev;
        unsigned int rss;
 #ifdef CONFIG_IXGBE_DCB
        int j;
@@ -3969,10 +4218,18 @@ static int __devinit ixgbe_sw_init(struct ixgbe_adapter *adapter)
                adapter->max_msix_q_vectors = MAX_MSIX_Q_VECTORS_82599;
                adapter->flags2 |= IXGBE_FLAG2_RSC_CAPABLE;
                adapter->flags2 |= IXGBE_FLAG2_RSC_ENABLED;
-               adapter->flags |= IXGBE_FLAG_FDIR_HASH_CAPABLE;
+               if (dev->features & NETIF_F_NTUPLE) {
+                       /* Flow Director perfect filter enabled */
+                       adapter->flags |= IXGBE_FLAG_FDIR_PERFECT_CAPABLE;
+                       adapter->atr_sample_rate = 0;
+                       spin_lock_init(&adapter->fdir_perfect_lock);
+               } else {
+                       /* Flow Director hash filters enabled */
+                       adapter->flags |= IXGBE_FLAG_FDIR_HASH_CAPABLE;
+                       adapter->atr_sample_rate = 20;
+               }
                adapter->ring_feature[RING_F_FDIR].indices =
                                                         IXGBE_MAX_FDIR_INDICES;
-               adapter->atr_sample_rate = 20;
                adapter->fdir_pballoc = 0;
 #ifdef IXGBE_FCOE
                adapter->flags |= IXGBE_FLAG_FCOE_CAPABLE;
@@ -4041,6 +4298,9 @@ static int __devinit ixgbe_sw_init(struct ixgbe_adapter *adapter)
        /* enable rx csum by default */
        adapter->flags |= IXGBE_FLAG_RX_CSUM_ENABLED;
 
+       /* get assigned NUMA node */
+       adapter->node = dev_to_node(&pdev->dev);
+
        set_bit(__IXGBE_DOWN, &adapter->state);
 
        return 0;
@@ -4060,7 +4320,9 @@ int ixgbe_setup_tx_resources(struct ixgbe_adapter *adapter,
        int size;
 
        size = sizeof(struct ixgbe_tx_buffer) * tx_ring->count;
-       tx_ring->tx_buffer_info = vmalloc(size);
+       tx_ring->tx_buffer_info = vmalloc_node(size, tx_ring->numa_node);
+       if (!tx_ring->tx_buffer_info)
+               tx_ring->tx_buffer_info = vmalloc(size);
        if (!tx_ring->tx_buffer_info)
                goto err;
        memset(tx_ring->tx_buffer_info, 0, size);
@@ -4102,7 +4364,7 @@ static int ixgbe_setup_all_tx_resources(struct ixgbe_adapter *adapter)
        int i, err = 0;
 
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               err = ixgbe_setup_tx_resources(adapter, &adapter->tx_ring[i]);
+               err = ixgbe_setup_tx_resources(adapter, adapter->tx_ring[i]);
                if (!err)
                        continue;
                DPRINTK(PROBE, ERR, "Allocation for Tx Queue %u failed\n", i);
@@ -4126,7 +4388,9 @@ int ixgbe_setup_rx_resources(struct ixgbe_adapter *adapter,
        int size;
 
        size = sizeof(struct ixgbe_rx_buffer) * rx_ring->count;
-       rx_ring->rx_buffer_info = vmalloc(size);
+       rx_ring->rx_buffer_info = vmalloc_node(size, adapter->node);
+       if (!rx_ring->rx_buffer_info)
+               rx_ring->rx_buffer_info = vmalloc(size);
        if (!rx_ring->rx_buffer_info) {
                DPRINTK(PROBE, ERR,
                        "vmalloc allocation failed for the rx desc ring\n");
@@ -4172,7 +4436,7 @@ static int ixgbe_setup_all_rx_resources(struct ixgbe_adapter *adapter)
        int i, err = 0;
 
        for (i = 0; i < adapter->num_rx_queues; i++) {
-               err = ixgbe_setup_rx_resources(adapter, &adapter->rx_ring[i]);
+               err = ixgbe_setup_rx_resources(adapter, adapter->rx_ring[i]);
                if (!err)
                        continue;
                DPRINTK(PROBE, ERR, "Allocation for Rx Queue %u failed\n", i);
@@ -4215,8 +4479,8 @@ static void ixgbe_free_all_tx_resources(struct ixgbe_adapter *adapter)
        int i;
 
        for (i = 0; i < adapter->num_tx_queues; i++)
-               if (adapter->tx_ring[i].desc)
-                       ixgbe_free_tx_resources(adapter, &adapter->tx_ring[i]);
+               if (adapter->tx_ring[i]->desc)
+                       ixgbe_free_tx_resources(adapter, adapter->tx_ring[i]);
 }
 
 /**
@@ -4252,8 +4516,8 @@ static void ixgbe_free_all_rx_resources(struct ixgbe_adapter *adapter)
        int i;
 
        for (i = 0; i < adapter->num_rx_queues; i++)
-               if (adapter->rx_ring[i].desc)
-                       ixgbe_free_rx_resources(adapter, &adapter->rx_ring[i]);
+               if (adapter->rx_ring[i]->desc)
+                       ixgbe_free_rx_resources(adapter, adapter->rx_ring[i]);
 }
 
 /**
@@ -4530,8 +4794,8 @@ void ixgbe_update_stats(struct ixgbe_adapter *adapter)
                        adapter->hw_rx_no_dma_resources +=
                                             IXGBE_READ_REG(hw, IXGBE_QPRDC(i));
                for (i = 0; i < adapter->num_rx_queues; i++) {
-                       rsc_count += adapter->rx_ring[i].rsc_count;
-                       rsc_flush += adapter->rx_ring[i].rsc_flush;
+                       rsc_count += adapter->rx_ring[i]->rsc_count;
+                       rsc_flush += adapter->rx_ring[i]->rsc_flush;
                }
                adapter->rsc_total_count = rsc_count;
                adapter->rsc_total_flush = rsc_flush;
@@ -4539,11 +4803,11 @@ void ixgbe_update_stats(struct ixgbe_adapter *adapter)
 
        /* gather some stats to the adapter struct that are per queue */
        for (i = 0; i < adapter->num_tx_queues; i++)
-               restart_queue += adapter->tx_ring[i].restart_queue;
+               restart_queue += adapter->tx_ring[i]->restart_queue;
        adapter->restart_queue = restart_queue;
 
        for (i = 0; i < adapter->num_rx_queues; i++)
-               non_eop_descs += adapter->rx_ring[i].non_eop_descs;
+               non_eop_descs += adapter->rx_ring[i]->non_eop_descs;
        adapter->non_eop_descs = non_eop_descs;
 
        adapter->stats.crcerrs += IXGBE_READ_REG(hw, IXGBE_CRCERRS);
@@ -4782,7 +5046,7 @@ static void ixgbe_fdir_reinit_task(struct work_struct *work)
        if (ixgbe_reinit_fdir_tables_82599(hw) == 0) {
                for (i = 0; i < adapter->num_tx_queues; i++)
                        set_bit(__IXGBE_FDIR_INIT_DONE,
-                               &(adapter->tx_ring[i].reinit_state));
+                               &(adapter->tx_ring[i]->reinit_state));
        } else {
                DPRINTK(PROBE, ERR, "failed to finish FDIR re-initialization, "
                        "ignored adding FDIR ATR filters \n");
@@ -4791,6 +5055,8 @@ static void ixgbe_fdir_reinit_task(struct work_struct *work)
        netif_tx_start_all_queues(adapter->netdev);
 }
 
+static DEFINE_MUTEX(ixgbe_watchdog_lock);
+
 /**
  * ixgbe_watchdog_task - worker thread to bring link up
  * @work: pointer to work_struct containing our data
@@ -4802,13 +5068,16 @@ static void ixgbe_watchdog_task(struct work_struct *work)
                                                     watchdog_task);
        struct net_device *netdev = adapter->netdev;
        struct ixgbe_hw *hw = &adapter->hw;
-       u32 link_speed = adapter->link_speed;
-       bool link_up = adapter->link_up;
+       u32 link_speed;
+       bool link_up;
        int i;
        struct ixgbe_ring *tx_ring;
        int some_tx_pending = 0;
 
-       adapter->flags |= IXGBE_FLAG_IN_WATCHDOG_TASK;
+       mutex_lock(&ixgbe_watchdog_lock);
+
+       link_up = adapter->link_up;
+       link_speed = adapter->link_speed;
 
        if (adapter->flags & IXGBE_FLAG_NEED_LINK_UPDATE) {
                hw->mac.ops.check_link(hw, &link_speed, &link_up, false);
@@ -4879,7 +5148,7 @@ static void ixgbe_watchdog_task(struct work_struct *work)
 
        if (!netif_carrier_ok(netdev)) {
                for (i = 0; i < adapter->num_tx_queues; i++) {
-                       tx_ring = &adapter->tx_ring[i];
+                       tx_ring = adapter->tx_ring[i];
                        if (tx_ring->next_to_use != tx_ring->next_to_clean) {
                                some_tx_pending = 1;
                                break;
@@ -4897,7 +5166,7 @@ static void ixgbe_watchdog_task(struct work_struct *work)
        }
 
        ixgbe_update_stats(adapter);
-       adapter->flags &= ~IXGBE_FLAG_IN_WATCHDOG_TASK;
+       mutex_unlock(&ixgbe_watchdog_lock);
 }
 
 static int ixgbe_tso(struct ixgbe_adapter *adapter,
@@ -4928,7 +5197,7 @@ static int ixgbe_tso(struct ixgbe_adapter *adapter,
                                                                 iph->daddr, 0,
                                                                 IPPROTO_TCP,
                                                                 0);
-               } else if (skb_shinfo(skb)->gso_type == SKB_GSO_TCPV6) {
+               } else if (skb_is_gso_v6(skb)) {
                        ipv6_hdr(skb)->payload_len = 0;
                        tcp_hdr(skb)->check =
                            ~csum_ipv6_magic(&ipv6_hdr(skb)->saddr,
@@ -5167,19 +5436,19 @@ dma_error:
        tx_buffer_info->dma = 0;
        tx_buffer_info->time_stamp = 0;
        tx_buffer_info->next_to_watch = 0;
-       count--;
+       if (count)
+               count--;
 
        /* clear timestamp and dma mappings for remaining portion of packet */
-       while (count >= 0) {
-               count--;
-               i--;
-               if (i < 0)
+       while (count--) {
+               if (i==0)
                        i += tx_ring->count;
+               i--;
                tx_buffer_info = &tx_ring->tx_buffer_info[i];
                ixgbe_unmap_and_free_tx_resource(adapter, tx_buffer_info);
        }
 
-       return count;
+       return 0;
 }
 
 static void ixgbe_tx_queue(struct ixgbe_adapter *adapter,
@@ -5329,8 +5598,11 @@ static u16 ixgbe_select_queue(struct net_device *dev, struct sk_buff *skb)
        struct ixgbe_adapter *adapter = netdev_priv(dev);
        int txq = smp_processor_id();
 
-       if (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE)
+       if (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE) {
+               while (unlikely(txq >= dev->real_num_tx_queues))
+                       txq -= dev->real_num_tx_queues;
                return txq;
+       }
 
 #ifdef IXGBE_FCOE
        if ((adapter->flags & IXGBE_FLAG_FCOE_ENABLED) &&
@@ -5378,7 +5650,7 @@ static netdev_tx_t ixgbe_xmit_frame(struct sk_buff *skb,
                }
        }
 
-       tx_ring = &adapter->tx_ring[skb->queue_mapping];
+       tx_ring = adapter->tx_ring[skb->queue_mapping];
 
        if ((adapter->flags & IXGBE_FLAG_FCOE_ENABLED) &&
            (skb->protocol == htons(ETH_P_FCOE))) {
@@ -5484,7 +5756,8 @@ static int ixgbe_set_mac(struct net_device *netdev, void *p)
        memcpy(netdev->dev_addr, addr->sa_data, netdev->addr_len);
        memcpy(hw->mac.addr, addr->sa_data, netdev->addr_len);
 
-       hw->mac.ops.set_rar(hw, 0, hw->mac.addr, 0, IXGBE_RAH_AV);
+       hw->mac.ops.set_rar(hw, 0, hw->mac.addr, adapter->num_vfs,
+                           IXGBE_RAH_AV);
 
        return 0;
 }
@@ -5621,6 +5894,61 @@ static const struct net_device_ops ixgbe_netdev_ops = {
 #endif /* IXGBE_FCOE */
 };
 
+static void __devinit ixgbe_probe_vf(struct ixgbe_adapter *adapter,
+                          const struct ixgbe_info *ii)
+{
+#ifdef CONFIG_PCI_IOV
+       struct ixgbe_hw *hw = &adapter->hw;
+       int err;
+
+       if (hw->mac.type != ixgbe_mac_82599EB || !max_vfs)
+               return;
+
+       /* The 82599 supports up to 64 VFs per physical function
+        * but this implementation limits allocation to 63 so that
+        * basic networking resources are still available to the
+        * physical function
+        */
+       adapter->num_vfs = (max_vfs > 63) ? 63 : max_vfs;
+       adapter->flags |= IXGBE_FLAG_SRIOV_ENABLED;
+       err = pci_enable_sriov(adapter->pdev, adapter->num_vfs);
+       if (err) {
+               DPRINTK(PROBE, ERR,
+                       "Failed to enable PCI sriov: %d\n", err);
+               goto err_novfs;
+       }
+       /* If call to enable VFs succeeded then allocate memory
+        * for per VF control structures.
+        */
+       adapter->vfinfo =
+               kcalloc(adapter->num_vfs,
+                       sizeof(struct vf_data_storage), GFP_KERNEL);
+       if (adapter->vfinfo) {
+               /* Now that we're sure SR-IOV is enabled
+                * and memory allocated set up the mailbox parameters
+                */
+               ixgbe_init_mbx_params_pf(hw);
+               memcpy(&hw->mbx.ops, ii->mbx_ops,
+                      sizeof(hw->mbx.ops));
+
+               /* Disable RSC when in SR-IOV mode */
+               adapter->flags2 &= ~(IXGBE_FLAG2_RSC_CAPABLE |
+                                    IXGBE_FLAG2_RSC_ENABLED);
+               return;
+       }
+
+       /* Oh oh */
+       DPRINTK(PROBE, ERR,
+               "Unable to allocate memory for VF "
+               "Data Storage - SRIOV disabled\n");
+       pci_disable_sriov(adapter->pdev);
+
+err_novfs:
+       adapter->flags &= ~IXGBE_FLAG_SRIOV_ENABLED;
+       adapter->num_vfs = 0;
+#endif /* CONFIG_PCI_IOV */
+}
+
 /**
  * ixgbe_probe - Device Initialization Routine
  * @pdev: PCI device information struct
@@ -5795,6 +6123,8 @@ static int __devinit ixgbe_probe(struct pci_dev *pdev,
                goto err_sw_init;
        }
 
+       ixgbe_probe_vf(adapter, ii);
+
        netdev->features = NETIF_F_SG |
                           NETIF_F_IP_CSUM |
                           NETIF_F_HW_VLAN_TX |
@@ -5815,6 +6145,9 @@ static int __devinit ixgbe_probe(struct pci_dev *pdev,
        netdev->vlan_features |= NETIF_F_IPV6_CSUM;
        netdev->vlan_features |= NETIF_F_SG;
 
+       if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED)
+               adapter->flags &= ~(IXGBE_FLAG_RSS_ENABLED |
+                                   IXGBE_FLAG_DCB_ENABLED);
        if (adapter->flags & IXGBE_FLAG_DCB_ENABLED)
                adapter->flags &= ~IXGBE_FLAG_RSS_ENABLED;
 
@@ -5941,6 +6274,13 @@ static int __devinit ixgbe_probe(struct pci_dev *pdev,
                ixgbe_setup_dca(adapter);
        }
 #endif
+       if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED) {
+               DPRINTK(PROBE, INFO, "IOV is enabled with %d VFs\n",
+                       adapter->num_vfs);
+               for (i = 0; i < adapter->num_vfs; i++)
+                       ixgbe_vf_configuration(pdev, (i | 0x10000000));
+       }
+
        /* add san mac addr to netdev */
        ixgbe_add_sanmac_netdev(netdev);
 
@@ -5953,6 +6293,8 @@ err_register:
        ixgbe_clear_interrupt_scheme(adapter);
 err_sw_init:
 err_eeprom:
+       if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED)
+               ixgbe_disable_sriov(adapter);
        clear_bit(__IXGBE_SFP_MODULE_NOT_FOUND, &adapter->state);
        del_timer_sync(&adapter->sfp_timer);
        cancel_work_sync(&adapter->sfp_task);
@@ -6021,6 +6363,9 @@ static void __devexit ixgbe_remove(struct pci_dev *pdev)
        if (netdev->reg_state == NETREG_REGISTERED)
                unregister_netdev(netdev);
 
+       if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED)
+               ixgbe_disable_sriov(adapter);
+
        ixgbe_clear_interrupt_scheme(adapter);
 
        ixgbe_release_hw_control(adapter);