These changes are the raw update to linux-4.4.6-rt14. Kernel sources
[kvmfornfv.git] / kernel / drivers / net / ethernet / mellanox / mlx4 / main.c
index ced5eca..31c491e 100644 (file)
@@ -405,6 +405,21 @@ static int mlx4_dev_cap(struct mlx4_dev *dev, struct mlx4_dev_cap *dev_cap)
        dev->caps.max_gso_sz         = dev_cap->max_gso_sz;
        dev->caps.max_rss_tbl_sz     = dev_cap->max_rss_tbl_sz;
 
+       if (dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_PHV_EN) {
+               struct mlx4_init_hca_param hca_param;
+
+               memset(&hca_param, 0, sizeof(hca_param));
+               err = mlx4_QUERY_HCA(dev, &hca_param);
+               /* Turn off PHV_EN flag in case phv_check_en is set.
+                * phv_check_en is a HW check that parse the packet and verify
+                * phv bit was reported correctly in the wqe. To allow QinQ
+                * PHV_EN flag should be set and phv_check_en must be cleared
+                * otherwise QinQ packets will be drop by the HW.
+                */
+               if (err || hca_param.phv_check_en)
+                       dev->caps.flags2 &= ~MLX4_DEV_CAP_FLAG2_PHV_EN;
+       }
+
        /* Sense port always allowed on supported devices for ConnectX-1 and -2 */
        if (mlx4_priv(dev)->pci_dev_data & MLX4_PCI_DEV_FORCE_SENSE_PORT)
                dev->caps.flags |= MLX4_DEV_CAP_FLAG_SENSE_SUPPORT;
@@ -479,7 +494,15 @@ static int mlx4_dev_cap(struct mlx4_dev *dev, struct mlx4_dev_cap *dev_cap)
                }
        }
 
-       dev->caps.max_counters = 1 << ilog2(dev_cap->max_counters);
+       if (mlx4_is_master(dev) && (dev->caps.num_ports == 2) &&
+           (port_type_array[0] == MLX4_PORT_TYPE_IB) &&
+           (port_type_array[1] == MLX4_PORT_TYPE_ETH)) {
+               mlx4_warn(dev,
+                         "Granular QoS per VF not supported with IB/Eth configuration\n");
+               dev->caps.flags2 &= ~MLX4_DEV_CAP_FLAG2_QOS_VPP;
+       }
+
+       dev->caps.max_counters = dev_cap->max_counters;
 
        dev->caps.reserved_qps_cnt[MLX4_QP_REGION_FW] = dev_cap->reserved_qps;
        dev->caps.reserved_qps_cnt[MLX4_QP_REGION_ETH_ADDR] =
@@ -840,6 +863,8 @@ static int mlx4_slave_cap(struct mlx4_dev *dev)
                return -ENODEV;
        }
 
+       mlx4_replace_zero_macs(dev);
+
        dev->caps.qp0_qkey = kcalloc(dev->caps.num_ports, sizeof(u32), GFP_KERNEL);
        dev->caps.qp0_tunnel = kcalloc(dev->caps.num_ports, sizeof (u32), GFP_KERNEL);
        dev->caps.qp0_proxy = kcalloc(dev->caps.num_ports, sizeof (u32), GFP_KERNEL);
@@ -867,9 +892,10 @@ static int mlx4_slave_cap(struct mlx4_dev *dev)
                dev->caps.qp1_proxy[i - 1] = func_cap.qp1_proxy_qpn;
                dev->caps.port_mask[i] = dev->caps.port_type[i];
                dev->caps.phys_port_id[i] = func_cap.phys_port_id;
-               if (mlx4_get_slave_pkey_gid_tbl_len(dev, i,
-                                                   &dev->caps.gid_table_len[i],
-                                                   &dev->caps.pkey_table_len[i]))
+               err = mlx4_get_slave_pkey_gid_tbl_len(dev, i,
+                                                     &dev->caps.gid_table_len[i],
+                                                     &dev->caps.pkey_table_len[i]);
+               if (err)
                        goto err_mem;
        }
 
@@ -881,6 +907,7 @@ static int mlx4_slave_cap(struct mlx4_dev *dev)
                         dev->caps.uar_page_size * dev->caps.num_uars,
                         (unsigned long long)
                         pci_resource_len(dev->persist->pdev, 2));
+               err = -ENOMEM;
                goto err_mem;
        }
 
@@ -1674,6 +1701,25 @@ static int map_internal_clock(struct mlx4_dev *dev)
        return 0;
 }
 
+int mlx4_get_internal_clock_params(struct mlx4_dev *dev,
+                                  struct mlx4_clock_params *params)
+{
+       struct mlx4_priv *priv = mlx4_priv(dev);
+
+       if (mlx4_is_slave(dev))
+               return -ENOTSUPP;
+
+       if (!params)
+               return -EINVAL;
+
+       params->bar = priv->fw.clock_bar;
+       params->offset = priv->fw.clock_offset;
+       params->size = MLX4_CLOCK_SIZE;
+
+       return 0;
+}
+EXPORT_SYMBOL_GPL(mlx4_get_internal_clock_params);
+
 static void unmap_internal_clock(struct mlx4_dev *dev)
 {
        struct mlx4_priv *priv = mlx4_priv(dev);
@@ -2193,20 +2239,78 @@ err_free_icm:
 static int mlx4_init_counters_table(struct mlx4_dev *dev)
 {
        struct mlx4_priv *priv = mlx4_priv(dev);
-       int nent;
+       int nent_pow2;
 
        if (!(dev->caps.flags & MLX4_DEV_CAP_FLAG_COUNTERS))
                return -ENOENT;
 
-       nent = dev->caps.max_counters;
-       return mlx4_bitmap_init(&priv->counters_bitmap, nent, nent - 1, 0, 0);
+       if (!dev->caps.max_counters)
+               return -ENOSPC;
+
+       nent_pow2 = roundup_pow_of_two(dev->caps.max_counters);
+       /* reserve last counter index for sink counter */
+       return mlx4_bitmap_init(&priv->counters_bitmap, nent_pow2,
+                               nent_pow2 - 1, 0,
+                               nent_pow2 - dev->caps.max_counters + 1);
 }
 
 static void mlx4_cleanup_counters_table(struct mlx4_dev *dev)
 {
+       if (!(dev->caps.flags & MLX4_DEV_CAP_FLAG_COUNTERS))
+               return;
+
+       if (!dev->caps.max_counters)
+               return;
+
        mlx4_bitmap_cleanup(&mlx4_priv(dev)->counters_bitmap);
 }
 
+static void mlx4_cleanup_default_counters(struct mlx4_dev *dev)
+{
+       struct mlx4_priv *priv = mlx4_priv(dev);
+       int port;
+
+       for (port = 0; port < dev->caps.num_ports; port++)
+               if (priv->def_counter[port] != -1)
+                       mlx4_counter_free(dev,  priv->def_counter[port]);
+}
+
+static int mlx4_allocate_default_counters(struct mlx4_dev *dev)
+{
+       struct mlx4_priv *priv = mlx4_priv(dev);
+       int port, err = 0;
+       u32 idx;
+
+       for (port = 0; port < dev->caps.num_ports; port++)
+               priv->def_counter[port] = -1;
+
+       for (port = 0; port < dev->caps.num_ports; port++) {
+               err = mlx4_counter_alloc(dev, &idx);
+
+               if (!err || err == -ENOSPC) {
+                       priv->def_counter[port] = idx;
+               } else if (err == -ENOENT) {
+                       err = 0;
+                       continue;
+               } else if (mlx4_is_slave(dev) && err == -EINVAL) {
+                       priv->def_counter[port] = MLX4_SINK_COUNTER_INDEX(dev);
+                       mlx4_warn(dev, "can't allocate counter from old PF driver, using index %d\n",
+                                 MLX4_SINK_COUNTER_INDEX(dev));
+                       err = 0;
+               } else {
+                       mlx4_err(dev, "%s: failed to allocate default counter port %d err %d\n",
+                                __func__, port + 1, err);
+                       mlx4_cleanup_default_counters(dev);
+                       return err;
+               }
+
+               mlx4_dbg(dev, "%s: default counter index %d for port %d\n",
+                        __func__, priv->def_counter[port], port + 1);
+       }
+
+       return err;
+}
+
 int __mlx4_counter_alloc(struct mlx4_dev *dev, u32 *idx)
 {
        struct mlx4_priv *priv = mlx4_priv(dev);
@@ -2215,8 +2319,10 @@ int __mlx4_counter_alloc(struct mlx4_dev *dev, u32 *idx)
                return -ENOENT;
 
        *idx = mlx4_bitmap_alloc(&priv->counters_bitmap);
-       if (*idx == -1)
-               return -ENOMEM;
+       if (*idx == -1) {
+               *idx = MLX4_SINK_COUNTER_INDEX(dev);
+               return -ENOSPC;
+       }
 
        return 0;
 }
@@ -2239,8 +2345,35 @@ int mlx4_counter_alloc(struct mlx4_dev *dev, u32 *idx)
 }
 EXPORT_SYMBOL_GPL(mlx4_counter_alloc);
 
+static int __mlx4_clear_if_stat(struct mlx4_dev *dev,
+                               u8 counter_index)
+{
+       struct mlx4_cmd_mailbox *if_stat_mailbox;
+       int err;
+       u32 if_stat_in_mod = (counter_index & 0xff) | MLX4_QUERY_IF_STAT_RESET;
+
+       if_stat_mailbox = mlx4_alloc_cmd_mailbox(dev);
+       if (IS_ERR(if_stat_mailbox))
+               return PTR_ERR(if_stat_mailbox);
+
+       err = mlx4_cmd_box(dev, 0, if_stat_mailbox->dma, if_stat_in_mod, 0,
+                          MLX4_CMD_QUERY_IF_STAT, MLX4_CMD_TIME_CLASS_C,
+                          MLX4_CMD_NATIVE);
+
+       mlx4_free_cmd_mailbox(dev, if_stat_mailbox);
+       return err;
+}
+
 void __mlx4_counter_free(struct mlx4_dev *dev, u32 idx)
 {
+       if (!(dev->caps.flags & MLX4_DEV_CAP_FLAG_COUNTERS))
+               return;
+
+       if (idx == MLX4_SINK_COUNTER_INDEX(dev))
+               return;
+
+       __mlx4_clear_if_stat(dev, idx);
+
        mlx4_bitmap_free(&mlx4_priv(dev)->counters_bitmap, idx, MLX4_USE_RR);
        return;
 }
@@ -2260,6 +2393,14 @@ void mlx4_counter_free(struct mlx4_dev *dev, u32 idx)
 }
 EXPORT_SYMBOL_GPL(mlx4_counter_free);
 
+int mlx4_get_default_counter_index(struct mlx4_dev *dev, int port)
+{
+       struct mlx4_priv *priv = mlx4_priv(dev);
+
+       return priv->def_counter[port - 1];
+}
+EXPORT_SYMBOL_GPL(mlx4_get_default_counter_index);
+
 void mlx4_set_admin_guid(struct mlx4_dev *dev, __be64 guid, int entry, int port)
 {
        struct mlx4_priv *priv = mlx4_priv(dev);
@@ -2364,11 +2505,11 @@ static int mlx4_setup_hca(struct mlx4_dev *dev)
        if (err) {
                if (dev->flags & MLX4_FLAG_MSI_X) {
                        mlx4_warn(dev, "NOP command failed to generate MSI-X interrupt IRQ %d)\n",
-                                 priv->eq_table.eq[dev->caps.num_comp_vectors].irq);
+                                 priv->eq_table.eq[MLX4_EQ_ASYNC].irq);
                        mlx4_warn(dev, "Trying again without MSI-X\n");
                } else {
                        mlx4_err(dev, "NOP command failed to generate interrupt (IRQ %d), aborting\n",
-                                priv->eq_table.eq[dev->caps.num_comp_vectors].irq);
+                                priv->eq_table.eq[MLX4_EQ_ASYNC].irq);
                        mlx4_err(dev, "BIOS or ACPI interrupt routing problem?\n");
                }
 
@@ -2395,10 +2536,18 @@ static int mlx4_setup_hca(struct mlx4_dev *dev)
                goto err_srq_table_free;
        }
 
-       err = mlx4_init_counters_table(dev);
-       if (err && err != -ENOENT) {
-               mlx4_err(dev, "Failed to initialize counters table, aborting\n");
-               goto err_qp_table_free;
+       if (!mlx4_is_slave(dev)) {
+               err = mlx4_init_counters_table(dev);
+               if (err && err != -ENOENT) {
+                       mlx4_err(dev, "Failed to initialize counters table, aborting\n");
+                       goto err_qp_table_free;
+               }
+       }
+
+       err = mlx4_allocate_default_counters(dev);
+       if (err) {
+               mlx4_err(dev, "Failed to allocate default counters, aborting\n");
+               goto err_counters_table_free;
        }
 
        if (!mlx4_is_slave(dev)) {
@@ -2432,15 +2581,19 @@ static int mlx4_setup_hca(struct mlx4_dev *dev)
                        if (err) {
                                mlx4_err(dev, "Failed to set port %d, aborting\n",
                                         port);
-                               goto err_counters_table_free;
+                               goto err_default_countes_free;
                        }
                }
        }
 
        return 0;
 
+err_default_countes_free:
+       mlx4_cleanup_default_counters(dev);
+
 err_counters_table_free:
-       mlx4_cleanup_counters_table(dev);
+       if (!mlx4_is_slave(dev))
+               mlx4_cleanup_counters_table(dev);
 
 err_qp_table_free:
        mlx4_cleanup_qp_table(dev);
@@ -2481,17 +2634,50 @@ err_uar_table_free:
        return err;
 }
 
+static int mlx4_init_affinity_hint(struct mlx4_dev *dev, int port, int eqn)
+{
+       int requested_cpu = 0;
+       struct mlx4_priv *priv = mlx4_priv(dev);
+       struct mlx4_eq *eq;
+       int off = 0;
+       int i;
+
+       if (eqn > dev->caps.num_comp_vectors)
+               return -EINVAL;
+
+       for (i = 1; i < port; i++)
+               off += mlx4_get_eqs_per_port(dev, i);
+
+       requested_cpu = eqn - off - !!(eqn > MLX4_EQ_ASYNC);
+
+       /* Meaning EQs are shared, and this call comes from the second port */
+       if (requested_cpu < 0)
+               return 0;
+
+       eq = &priv->eq_table.eq[eqn];
+
+       if (!zalloc_cpumask_var(&eq->affinity_mask, GFP_KERNEL))
+               return -ENOMEM;
+
+       cpumask_set_cpu(requested_cpu, eq->affinity_mask);
+
+       return 0;
+}
+
 static void mlx4_enable_msi_x(struct mlx4_dev *dev)
 {
        struct mlx4_priv *priv = mlx4_priv(dev);
        struct msix_entry *entries;
        int i;
+       int port = 0;
 
        if (msi_x) {
-               int nreq = dev->caps.num_ports * num_online_cpus() + MSIX_LEGACY_SZ;
+               int nreq = dev->caps.num_ports * num_online_cpus() + 1;
 
                nreq = min_t(int, dev->caps.num_eqs - dev->caps.reserved_eqs,
                             nreq);
+               if (nreq > MAX_MSIX)
+                       nreq = MAX_MSIX;
 
                entries = kcalloc(nreq, sizeof *entries, GFP_KERNEL);
                if (!entries)
@@ -2503,20 +2689,55 @@ static void mlx4_enable_msi_x(struct mlx4_dev *dev)
                nreq = pci_enable_msix_range(dev->persist->pdev, entries, 2,
                                             nreq);
 
-               if (nreq < 0) {
+               if (nreq < 0 || nreq < MLX4_EQ_ASYNC) {
                        kfree(entries);
                        goto no_msi;
-               } else if (nreq < MSIX_LEGACY_SZ +
-                          dev->caps.num_ports * MIN_MSIX_P_PORT) {
-                       /*Working in legacy mode , all EQ's shared*/
-                       dev->caps.comp_pool           = 0;
-                       dev->caps.num_comp_vectors = nreq - 1;
-               } else {
-                       dev->caps.comp_pool           = nreq - MSIX_LEGACY_SZ;
-                       dev->caps.num_comp_vectors = MSIX_LEGACY_SZ - 1;
                }
-               for (i = 0; i < nreq; ++i)
-                       priv->eq_table.eq[i].irq = entries[i].vector;
+               /* 1 is reserved for events (asyncrounous EQ) */
+               dev->caps.num_comp_vectors = nreq - 1;
+
+               priv->eq_table.eq[MLX4_EQ_ASYNC].irq = entries[0].vector;
+               bitmap_zero(priv->eq_table.eq[MLX4_EQ_ASYNC].actv_ports.ports,
+                           dev->caps.num_ports);
+
+               for (i = 0; i < dev->caps.num_comp_vectors + 1; i++) {
+                       if (i == MLX4_EQ_ASYNC)
+                               continue;
+
+                       priv->eq_table.eq[i].irq =
+                               entries[i + 1 - !!(i > MLX4_EQ_ASYNC)].vector;
+
+                       if (MLX4_IS_LEGACY_EQ_MODE(dev->caps)) {
+                               bitmap_fill(priv->eq_table.eq[i].actv_ports.ports,
+                                           dev->caps.num_ports);
+                               /* We don't set affinity hint when there
+                                * aren't enough EQs
+                                */
+                       } else {
+                               set_bit(port,
+                                       priv->eq_table.eq[i].actv_ports.ports);
+                               if (mlx4_init_affinity_hint(dev, port + 1, i))
+                                       mlx4_warn(dev, "Couldn't init hint cpumask for EQ %d\n",
+                                                 i);
+                       }
+                       /* We divide the Eqs evenly between the two ports.
+                        * (dev->caps.num_comp_vectors / dev->caps.num_ports)
+                        * refers to the number of Eqs per port
+                        * (i.e eqs_per_port). Theoretically, we would like to
+                        * write something like (i + 1) % eqs_per_port == 0.
+                        * However, since there's an asynchronous Eq, we have
+                        * to skip over it by comparing this condition to
+                        * !!((i + 1) > MLX4_EQ_ASYNC).
+                        */
+                       if ((dev->caps.num_comp_vectors > dev->caps.num_ports) &&
+                           ((i + 1) %
+                            (dev->caps.num_comp_vectors / dev->caps.num_ports)) ==
+                           !!((i + 1) > MLX4_EQ_ASYNC))
+                               /* If dev->caps.num_comp_vectors < dev->caps.num_ports,
+                                * everything is shared anyway.
+                                */
+                               port++;
+               }
 
                dev->flags |= MLX4_FLAG_MSI_X;
 
@@ -2526,10 +2747,15 @@ static void mlx4_enable_msi_x(struct mlx4_dev *dev)
 
 no_msi:
        dev->caps.num_comp_vectors = 1;
-       dev->caps.comp_pool        = 0;
 
-       for (i = 0; i < 2; ++i)
+       BUG_ON(MLX4_EQ_ASYNC >= 2);
+       for (i = 0; i < 2; ++i) {
                priv->eq_table.eq[i].irq = dev->persist->pdev->irq;
+               if (i != MLX4_EQ_ASYNC) {
+                       bitmap_fill(priv->eq_table.eq[i].actv_ports.ports,
+                                   dev->caps.num_ports);
+               }
+       }
 }
 
 static int mlx4_init_port_info(struct mlx4_dev *dev, int port)
@@ -2594,6 +2820,10 @@ static void mlx4_cleanup_port_info(struct mlx4_port_info *info)
        device_remove_file(&info->dev->persist->pdev->dev, &info->port_attr);
        device_remove_file(&info->dev->persist->pdev->dev,
                           &info->port_mtu_attr);
+#ifdef CONFIG_RFS_ACCEL
+       free_irq_cpu_rmap(info->rmap);
+       info->rmap = NULL;
+#endif
 }
 
 static int mlx4_init_steering(struct mlx4_dev *dev)
@@ -2703,6 +2933,8 @@ static u64 mlx4_enable_sriov(struct mlx4_dev *dev, struct pci_dev *pdev,
 {
        u64 dev_flags = dev->flags;
        int err = 0;
+       int fw_enabled_sriov_vfs = min(pci_sriov_get_totalvfs(pdev),
+                                       MLX4_MAX_NUM_VF);
 
        if (reset_flow) {
                dev->dev_vfs = kcalloc(total_vfs, sizeof(*dev->dev_vfs),
@@ -2728,6 +2960,12 @@ static u64 mlx4_enable_sriov(struct mlx4_dev *dev, struct pci_dev *pdev,
        }
 
        if (!(dev->flags &  MLX4_FLAG_SRIOV)) {
+               if (total_vfs > fw_enabled_sriov_vfs) {
+                       mlx4_err(dev, "requested vfs (%d) > available vfs (%d). Continuing without SR_IOV\n",
+                                total_vfs, fw_enabled_sriov_vfs);
+                       err = -ENOMEM;
+                       goto disable_sriov;
+               }
                mlx4_warn(dev, "Enabling SR-IOV with %d VFs\n", total_vfs);
                err = pci_enable_sriov(pdev, total_vfs);
        }
@@ -2749,6 +2987,7 @@ disable_sriov:
 free_mem:
        dev->persist->num_vfs = 0;
        kfree(dev->dev_vfs);
+        dev->dev_vfs = NULL;
        return dev_flags & ~MLX4_FLAG_MASTER;
 }
 
@@ -2900,6 +3139,7 @@ slave_start:
                                                                  existing_vfs,
                                                                  reset_flow);
 
+                               mlx4_close_fw(dev);
                                mlx4_cmd_cleanup(dev, MLX4_CMD_CLEANUP_ALL);
                                dev->flags = dev_flags;
                                if (!SRIOV_VALID_STATE(dev->flags)) {
@@ -2988,18 +3228,6 @@ slave_start:
        /* In master functions, the communication channel must be initialized
         * after obtaining its address from fw */
        if (mlx4_is_master(dev)) {
-               int ib_ports = 0;
-
-               mlx4_foreach_port(i, dev, MLX4_PORT_TYPE_IB)
-                       ib_ports++;
-
-               if (ib_ports &&
-                   (num_vfs_argc > 1 || probe_vfs_argc > 1)) {
-                       mlx4_err(dev,
-                                "Invalid syntax of num_vfs/probe_vfs with IB port - single port VFs syntax is only supported when all ports are configured as ethernet\n");
-                       err = -EINVAL;
-                       goto err_close;
-               }
                if (dev->caps.num_ports < 2 &&
                    num_vfs_argc > 1) {
                        err = -EINVAL;
@@ -3036,7 +3264,7 @@ slave_start:
        if (err)
                goto err_master_mfunc;
 
-       priv->msix_ctl.pool_bm = 0;
+       bitmap_zero(priv->msix_ctl.pool_bm, MAX_MSIX);
        mutex_init(&priv->msix_ctl.pool_lock);
 
        mlx4_enable_msi_x(dev);
@@ -3058,7 +3286,6 @@ slave_start:
            !mlx4_is_mfunc(dev)) {
                dev->flags &= ~MLX4_FLAG_MSI_X;
                dev->caps.num_comp_vectors = 1;
-               dev->caps.comp_pool        = 0;
                pci_disable_msix(pdev);
                err = mlx4_setup_hca(dev);
        }
@@ -3109,7 +3336,9 @@ err_port:
        for (--port; port >= 1; --port)
                mlx4_cleanup_port_info(&priv->port[port]);
 
-       mlx4_cleanup_counters_table(dev);
+       mlx4_cleanup_default_counters(dev);
+       if (!mlx4_is_slave(dev))
+               mlx4_cleanup_counters_table(dev);
        mlx4_cleanup_qp_table(dev);
        mlx4_cleanup_srq_table(dev);
        mlx4_cleanup_cq_table(dev);
@@ -3218,20 +3447,20 @@ static int __mlx4_init_one(struct pci_dev *pdev, int pci_dev_data,
                        goto err_disable_pdev;
                }
        }
-       if (total_vfs >= MLX4_MAX_NUM_VF) {
+       if (total_vfs > MLX4_MAX_NUM_VF) {
                dev_err(&pdev->dev,
-                       "Requested more VF's (%d) than allowed (%d)\n",
-                       total_vfs, MLX4_MAX_NUM_VF - 1);
+                       "Requested more VF's (%d) than allowed by hw (%d)\n",
+                       total_vfs, MLX4_MAX_NUM_VF);
                err = -EINVAL;
                goto err_disable_pdev;
        }
 
        for (i = 0; i < MLX4_MAX_PORTS; i++) {
-               if (nvfs[i] + nvfs[2] >= MLX4_MAX_NUM_VF_P_PORT) {
+               if (nvfs[i] + nvfs[2] > MLX4_MAX_NUM_VF_P_PORT) {
                        dev_err(&pdev->dev,
-                               "Requested more VF's (%d) for port (%d) than allowed (%d)\n",
+                               "Requested more VF's (%d) for port (%d) than allowed by driver (%d)\n",
                                nvfs[i] + nvfs[2], i + 1,
-                               MLX4_MAX_NUM_VF_P_PORT - 1);
+                               MLX4_MAX_NUM_VF_P_PORT);
                        err = -EINVAL;
                        goto err_disable_pdev;
                }
@@ -3407,7 +3636,9 @@ static void mlx4_unload_one(struct pci_dev *pdev)
                mlx4_free_resource_tracker(dev,
                                           RES_TR_FREE_SLAVES_ONLY);
 
-       mlx4_cleanup_counters_table(dev);
+       mlx4_cleanup_default_counters(dev);
+       if (!mlx4_is_slave(dev))
+               mlx4_cleanup_counters_table(dev);
        mlx4_cleanup_qp_table(dev);
        mlx4_cleanup_srq_table(dev);
        mlx4_cleanup_cq_table(dev);