root/drivers/net/ethernet/netronome/nfp/nfp_netvf_main.c

/* [<][>][^][v][top][bottom][index][help] */

DEFINITIONS

This source file includes following definitions.
  1. nfp_netvf_get_mac_addr
  2. nfp_netvf_pci_probe
  3. nfp_netvf_pci_remove

   1 // SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
   2 /* Copyright (C) 2015-2018 Netronome Systems, Inc. */
   3 
   4 /*
   5  * nfp_netvf_main.c
   6  * Netronome virtual function network device driver: Main entry point
   7  * Author: Jason McMullan <jason.mcmullan@netronome.com>
   8  *         Rolf Neugebauer <rolf.neugebauer@netronome.com>
   9  */
  10 
  11 #include <linux/module.h>
  12 #include <linux/kernel.h>
  13 #include <linux/init.h>
  14 #include <linux/etherdevice.h>
  15 
  16 #include "nfp_net_ctrl.h"
  17 #include "nfp_net.h"
  18 #include "nfp_main.h"
  19 
  20 /**
  21  * struct nfp_net_vf - NFP VF-specific device structure
  22  * @nn:         NFP Net structure for this device
  23  * @irq_entries: Pre-allocated array of MSI-X entries
  24  * @q_bar:      Pointer to mapped QC memory (NULL if TX/RX mapped directly)
  25  * @ddir:       Per-device debugfs directory
  26  */
  27 struct nfp_net_vf {
  28         struct nfp_net *nn;
  29 
  30         struct msix_entry irq_entries[NFP_NET_NON_Q_VECTORS +
  31                                       NFP_NET_MAX_TX_RINGS];
  32         u8 __iomem *q_bar;
  33 
  34         struct dentry *ddir;
  35 };
  36 
  37 static const char nfp_net_driver_name[] = "nfp_netvf";
  38 
  39 #define PCI_DEVICE_NFP6000VF            0x6003
  40 static const struct pci_device_id nfp_netvf_pci_device_ids[] = {
  41         { PCI_VENDOR_ID_NETRONOME, PCI_DEVICE_NFP6000VF,
  42           PCI_VENDOR_ID_NETRONOME, PCI_ANY_ID,
  43           PCI_ANY_ID, 0,
  44         },
  45         { 0, } /* Required last entry. */
  46 };
  47 MODULE_DEVICE_TABLE(pci, nfp_netvf_pci_device_ids);
  48 
  49 static void nfp_netvf_get_mac_addr(struct nfp_net *nn)
  50 {
  51         u8 mac_addr[ETH_ALEN];
  52 
  53         put_unaligned_be32(nn_readl(nn, NFP_NET_CFG_MACADDR + 0), &mac_addr[0]);
  54         put_unaligned_be16(nn_readw(nn, NFP_NET_CFG_MACADDR + 6), &mac_addr[4]);
  55 
  56         if (!is_valid_ether_addr(mac_addr)) {
  57                 eth_hw_addr_random(nn->dp.netdev);
  58                 return;
  59         }
  60 
  61         ether_addr_copy(nn->dp.netdev->dev_addr, mac_addr);
  62         ether_addr_copy(nn->dp.netdev->perm_addr, mac_addr);
  63 }
  64 
  65 static int nfp_netvf_pci_probe(struct pci_dev *pdev,
  66                                const struct pci_device_id *pci_id)
  67 {
  68         struct nfp_net_fw_version fw_ver;
  69         int max_tx_rings, max_rx_rings;
  70         u32 tx_bar_off, rx_bar_off;
  71         u32 tx_bar_sz, rx_bar_sz;
  72         int tx_bar_no, rx_bar_no;
  73         struct nfp_net_vf *vf;
  74         unsigned int num_irqs;
  75         u8 __iomem *ctrl_bar;
  76         struct nfp_net *nn;
  77         u32 startq;
  78         int stride;
  79         int err;
  80 
  81         vf = kzalloc(sizeof(*vf), GFP_KERNEL);
  82         if (!vf)
  83                 return -ENOMEM;
  84         pci_set_drvdata(pdev, vf);
  85 
  86         err = pci_enable_device_mem(pdev);
  87         if (err)
  88                 goto err_free_vf;
  89 
  90         err = pci_request_regions(pdev, nfp_net_driver_name);
  91         if (err) {
  92                 dev_err(&pdev->dev, "Unable to allocate device memory.\n");
  93                 goto err_pci_disable;
  94         }
  95 
  96         pci_set_master(pdev);
  97 
  98         err = dma_set_mask_and_coherent(&pdev->dev,
  99                                         DMA_BIT_MASK(NFP_NET_MAX_DMA_BITS));
 100         if (err)
 101                 goto err_pci_regions;
 102 
 103         /* Map the Control BAR.
 104          *
 105          * Irrespective of the advertised BAR size we only map the
 106          * first NFP_NET_CFG_BAR_SZ of the BAR.  This keeps the code
 107          * the identical for PF and VF drivers.
 108          */
 109         ctrl_bar = ioremap_nocache(pci_resource_start(pdev, NFP_NET_CTRL_BAR),
 110                                    NFP_NET_CFG_BAR_SZ);
 111         if (!ctrl_bar) {
 112                 dev_err(&pdev->dev,
 113                         "Failed to map resource %d\n", NFP_NET_CTRL_BAR);
 114                 err = -EIO;
 115                 goto err_pci_regions;
 116         }
 117 
 118         nfp_net_get_fw_version(&fw_ver, ctrl_bar);
 119         if (fw_ver.resv || fw_ver.class != NFP_NET_CFG_VERSION_CLASS_GENERIC) {
 120                 dev_err(&pdev->dev, "Unknown Firmware ABI %d.%d.%d.%d\n",
 121                         fw_ver.resv, fw_ver.class, fw_ver.major, fw_ver.minor);
 122                 err = -EINVAL;
 123                 goto err_ctrl_unmap;
 124         }
 125 
 126         /* Determine stride */
 127         if (nfp_net_fw_ver_eq(&fw_ver, 0, 0, 0, 1)) {
 128                 stride = 2;
 129                 tx_bar_no = NFP_NET_Q0_BAR;
 130                 rx_bar_no = NFP_NET_Q1_BAR;
 131                 dev_warn(&pdev->dev, "OBSOLETE Firmware detected - VF isolation not available\n");
 132         } else {
 133                 switch (fw_ver.major) {
 134                 case 1 ... 5:
 135                         stride = 4;
 136                         tx_bar_no = NFP_NET_Q0_BAR;
 137                         rx_bar_no = tx_bar_no;
 138                         break;
 139                 default:
 140                         dev_err(&pdev->dev, "Unsupported Firmware ABI %d.%d.%d.%d\n",
 141                                 fw_ver.resv, fw_ver.class,
 142                                 fw_ver.major, fw_ver.minor);
 143                         err = -EINVAL;
 144                         goto err_ctrl_unmap;
 145                 }
 146         }
 147 
 148         /* Find out how many rings are supported */
 149         max_tx_rings = readl(ctrl_bar + NFP_NET_CFG_MAX_TXRINGS);
 150         max_rx_rings = readl(ctrl_bar + NFP_NET_CFG_MAX_RXRINGS);
 151 
 152         tx_bar_sz = NFP_QCP_QUEUE_ADDR_SZ * max_tx_rings * stride;
 153         rx_bar_sz = NFP_QCP_QUEUE_ADDR_SZ * max_rx_rings * stride;
 154 
 155         /* Sanity checks */
 156         if (tx_bar_sz > pci_resource_len(pdev, tx_bar_no)) {
 157                 dev_err(&pdev->dev,
 158                         "TX BAR too small for number of TX rings. Adjusting\n");
 159                 tx_bar_sz = pci_resource_len(pdev, tx_bar_no);
 160                 max_tx_rings = (tx_bar_sz / NFP_QCP_QUEUE_ADDR_SZ) / 2;
 161         }
 162         if (rx_bar_sz > pci_resource_len(pdev, rx_bar_no)) {
 163                 dev_err(&pdev->dev,
 164                         "RX BAR too small for number of RX rings. Adjusting\n");
 165                 rx_bar_sz = pci_resource_len(pdev, rx_bar_no);
 166                 max_rx_rings = (rx_bar_sz / NFP_QCP_QUEUE_ADDR_SZ) / 2;
 167         }
 168 
 169         startq = readl(ctrl_bar + NFP_NET_CFG_START_TXQ);
 170         tx_bar_off = NFP_PCIE_QUEUE(startq);
 171         startq = readl(ctrl_bar + NFP_NET_CFG_START_RXQ);
 172         rx_bar_off = NFP_PCIE_QUEUE(startq);
 173 
 174         /* Allocate and initialise the netdev */
 175         nn = nfp_net_alloc(pdev, ctrl_bar, true, max_tx_rings, max_rx_rings);
 176         if (IS_ERR(nn)) {
 177                 err = PTR_ERR(nn);
 178                 goto err_ctrl_unmap;
 179         }
 180         vf->nn = nn;
 181 
 182         nn->fw_ver = fw_ver;
 183         nn->dp.is_vf = 1;
 184         nn->stride_tx = stride;
 185         nn->stride_rx = stride;
 186 
 187         if (rx_bar_no == tx_bar_no) {
 188                 u32 bar_off, bar_sz;
 189                 resource_size_t map_addr;
 190 
 191                 /* Make a single overlapping BAR mapping */
 192                 if (tx_bar_off < rx_bar_off)
 193                         bar_off = tx_bar_off;
 194                 else
 195                         bar_off = rx_bar_off;
 196 
 197                 if ((tx_bar_off + tx_bar_sz) > (rx_bar_off + rx_bar_sz))
 198                         bar_sz = (tx_bar_off + tx_bar_sz) - bar_off;
 199                 else
 200                         bar_sz = (rx_bar_off + rx_bar_sz) - bar_off;
 201 
 202                 map_addr = pci_resource_start(pdev, tx_bar_no) + bar_off;
 203                 vf->q_bar = ioremap_nocache(map_addr, bar_sz);
 204                 if (!vf->q_bar) {
 205                         nn_err(nn, "Failed to map resource %d\n", tx_bar_no);
 206                         err = -EIO;
 207                         goto err_netdev_free;
 208                 }
 209 
 210                 /* TX queues */
 211                 nn->tx_bar = vf->q_bar + (tx_bar_off - bar_off);
 212                 /* RX queues */
 213                 nn->rx_bar = vf->q_bar + (rx_bar_off - bar_off);
 214         } else {
 215                 resource_size_t map_addr;
 216 
 217                 /* TX queues */
 218                 map_addr = pci_resource_start(pdev, tx_bar_no) + tx_bar_off;
 219                 nn->tx_bar = ioremap_nocache(map_addr, tx_bar_sz);
 220                 if (!nn->tx_bar) {
 221                         nn_err(nn, "Failed to map resource %d\n", tx_bar_no);
 222                         err = -EIO;
 223                         goto err_netdev_free;
 224                 }
 225 
 226                 /* RX queues */
 227                 map_addr = pci_resource_start(pdev, rx_bar_no) + rx_bar_off;
 228                 nn->rx_bar = ioremap_nocache(map_addr, rx_bar_sz);
 229                 if (!nn->rx_bar) {
 230                         nn_err(nn, "Failed to map resource %d\n", rx_bar_no);
 231                         err = -EIO;
 232                         goto err_unmap_tx;
 233                 }
 234         }
 235 
 236         nfp_netvf_get_mac_addr(nn);
 237 
 238         num_irqs = nfp_net_irqs_alloc(pdev, vf->irq_entries,
 239                                       NFP_NET_MIN_VNIC_IRQS,
 240                                       NFP_NET_NON_Q_VECTORS +
 241                                       nn->dp.num_r_vecs);
 242         if (!num_irqs) {
 243                 nn_warn(nn, "Unable to allocate MSI-X Vectors. Exiting\n");
 244                 err = -EIO;
 245                 goto err_unmap_rx;
 246         }
 247         nfp_net_irqs_assign(nn, vf->irq_entries, num_irqs);
 248 
 249         err = nfp_net_init(nn);
 250         if (err)
 251                 goto err_irqs_disable;
 252 
 253         nfp_net_info(nn);
 254         vf->ddir = nfp_net_debugfs_device_add(pdev);
 255         nfp_net_debugfs_vnic_add(nn, vf->ddir);
 256 
 257         return 0;
 258 
 259 err_irqs_disable:
 260         nfp_net_irqs_disable(pdev);
 261 err_unmap_rx:
 262         if (!vf->q_bar)
 263                 iounmap(nn->rx_bar);
 264 err_unmap_tx:
 265         if (!vf->q_bar)
 266                 iounmap(nn->tx_bar);
 267         else
 268                 iounmap(vf->q_bar);
 269 err_netdev_free:
 270         nfp_net_free(nn);
 271 err_ctrl_unmap:
 272         iounmap(ctrl_bar);
 273 err_pci_regions:
 274         pci_release_regions(pdev);
 275 err_pci_disable:
 276         pci_disable_device(pdev);
 277 err_free_vf:
 278         pci_set_drvdata(pdev, NULL);
 279         kfree(vf);
 280         return err;
 281 }
 282 
 283 static void nfp_netvf_pci_remove(struct pci_dev *pdev)
 284 {
 285         struct nfp_net_vf *vf;
 286         struct nfp_net *nn;
 287 
 288         vf = pci_get_drvdata(pdev);
 289         if (!vf)
 290                 return;
 291 
 292         nn = vf->nn;
 293 
 294         /* Note, the order is slightly different from above as we need
 295          * to keep the nn pointer around till we have freed everything.
 296          */
 297         nfp_net_debugfs_dir_clean(&nn->debugfs_dir);
 298         nfp_net_debugfs_dir_clean(&vf->ddir);
 299 
 300         nfp_net_clean(nn);
 301 
 302         nfp_net_irqs_disable(pdev);
 303 
 304         if (!vf->q_bar) {
 305                 iounmap(nn->rx_bar);
 306                 iounmap(nn->tx_bar);
 307         } else {
 308                 iounmap(vf->q_bar);
 309         }
 310         iounmap(nn->dp.ctrl_bar);
 311 
 312         nfp_net_free(nn);
 313 
 314         pci_release_regions(pdev);
 315         pci_disable_device(pdev);
 316 
 317         pci_set_drvdata(pdev, NULL);
 318         kfree(vf);
 319 }
 320 
 321 struct pci_driver nfp_netvf_pci_driver = {
 322         .name        = nfp_net_driver_name,
 323         .id_table    = nfp_netvf_pci_device_ids,
 324         .probe       = nfp_netvf_pci_probe,
 325         .remove      = nfp_netvf_pci_remove,
 326         .shutdown    = nfp_netvf_pci_remove,
 327 };

/* [<][>][^][v][top][bottom][index][help] */