mirror of
https://github.com/tbsdtv/linux_media.git
synced 2025-07-24 05:01:03 +02:00
Some DFL FPGA PCIe cards (e.g. Intel FPGA Programmable Acceleration Card) support MSI-X based interrupts. This patch allows PCIe driver to prepare and pass interrupt resources to DFL via enumeration API. These interrupt resources could then be assigned to actual features which use them. Signed-off-by: Luwei Kang <luwei.kang@intel.com> Signed-off-by: Wu Hao <hao.wu@intel.com> Signed-off-by: Xu Yilun <yilun.xu@intel.com> Signed-off-by: Tom Rix <trix@redhat.com> Reviewed-by: Marcelo Tosatti <mtosatti@redhat.com> Acked-by: Wu Hao <hao.wu@intel.com> Signed-off-by: Moritz Fischer <mdf@kernel.org>
340 lines
8.1 KiB
C
340 lines
8.1 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Driver for FPGA Device Feature List (DFL) PCIe device
|
|
*
|
|
* Copyright (C) 2017-2018 Intel Corporation, Inc.
|
|
*
|
|
* Authors:
|
|
* Zhang Yi <Yi.Z.Zhang@intel.com>
|
|
* Xiao Guangrong <guangrong.xiao@linux.intel.com>
|
|
* Joseph Grecco <joe.grecco@intel.com>
|
|
* Enno Luebbers <enno.luebbers@intel.com>
|
|
* Tim Whisonant <tim.whisonant@intel.com>
|
|
* Ananda Ravuri <ananda.ravuri@intel.com>
|
|
* Henry Mitchel <henry.mitchel@intel.com>
|
|
*/
|
|
|
|
#include <linux/pci.h>
|
|
#include <linux/types.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/stddef.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/aer.h>
|
|
|
|
#include "dfl.h"
|
|
|
|
#define DRV_VERSION "0.8"
|
|
#define DRV_NAME "dfl-pci"
|
|
|
|
struct cci_drvdata {
|
|
struct dfl_fpga_cdev *cdev; /* container device */
|
|
};
|
|
|
|
static void __iomem *cci_pci_ioremap_bar(struct pci_dev *pcidev, int bar)
|
|
{
|
|
if (pcim_iomap_regions(pcidev, BIT(bar), DRV_NAME))
|
|
return NULL;
|
|
|
|
return pcim_iomap_table(pcidev)[bar];
|
|
}
|
|
|
|
static int cci_pci_alloc_irq(struct pci_dev *pcidev)
|
|
{
|
|
int ret, nvec = pci_msix_vec_count(pcidev);
|
|
|
|
if (nvec <= 0) {
|
|
dev_dbg(&pcidev->dev, "fpga interrupt not supported\n");
|
|
return 0;
|
|
}
|
|
|
|
ret = pci_alloc_irq_vectors(pcidev, nvec, nvec, PCI_IRQ_MSIX);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
return nvec;
|
|
}
|
|
|
|
static void cci_pci_free_irq(struct pci_dev *pcidev)
|
|
{
|
|
pci_free_irq_vectors(pcidev);
|
|
}
|
|
|
|
/* PCI Device ID */
|
|
#define PCIE_DEVICE_ID_PF_INT_5_X 0xBCBD
|
|
#define PCIE_DEVICE_ID_PF_INT_6_X 0xBCC0
|
|
#define PCIE_DEVICE_ID_PF_DSC_1_X 0x09C4
|
|
/* VF Device */
|
|
#define PCIE_DEVICE_ID_VF_INT_5_X 0xBCBF
|
|
#define PCIE_DEVICE_ID_VF_INT_6_X 0xBCC1
|
|
#define PCIE_DEVICE_ID_VF_DSC_1_X 0x09C5
|
|
|
|
static struct pci_device_id cci_pcie_id_tbl[] = {
|
|
{PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_PF_INT_5_X),},
|
|
{PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_VF_INT_5_X),},
|
|
{PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_PF_INT_6_X),},
|
|
{PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_VF_INT_6_X),},
|
|
{PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_PF_DSC_1_X),},
|
|
{PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_VF_DSC_1_X),},
|
|
{0,}
|
|
};
|
|
MODULE_DEVICE_TABLE(pci, cci_pcie_id_tbl);
|
|
|
|
static int cci_init_drvdata(struct pci_dev *pcidev)
|
|
{
|
|
struct cci_drvdata *drvdata;
|
|
|
|
drvdata = devm_kzalloc(&pcidev->dev, sizeof(*drvdata), GFP_KERNEL);
|
|
if (!drvdata)
|
|
return -ENOMEM;
|
|
|
|
pci_set_drvdata(pcidev, drvdata);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void cci_remove_feature_devs(struct pci_dev *pcidev)
|
|
{
|
|
struct cci_drvdata *drvdata = pci_get_drvdata(pcidev);
|
|
|
|
/* remove all children feature devices */
|
|
dfl_fpga_feature_devs_remove(drvdata->cdev);
|
|
cci_pci_free_irq(pcidev);
|
|
}
|
|
|
|
static int *cci_pci_create_irq_table(struct pci_dev *pcidev, unsigned int nvec)
|
|
{
|
|
unsigned int i;
|
|
int *table;
|
|
|
|
table = kcalloc(nvec, sizeof(int), GFP_KERNEL);
|
|
if (!table)
|
|
return table;
|
|
|
|
for (i = 0; i < nvec; i++)
|
|
table[i] = pci_irq_vector(pcidev, i);
|
|
|
|
return table;
|
|
}
|
|
|
|
/* enumerate feature devices under pci device */
|
|
static int cci_enumerate_feature_devs(struct pci_dev *pcidev)
|
|
{
|
|
struct cci_drvdata *drvdata = pci_get_drvdata(pcidev);
|
|
int port_num, bar, i, nvec, ret = 0;
|
|
struct dfl_fpga_enum_info *info;
|
|
struct dfl_fpga_cdev *cdev;
|
|
resource_size_t start, len;
|
|
void __iomem *base;
|
|
int *irq_table;
|
|
u32 offset;
|
|
u64 v;
|
|
|
|
/* allocate enumeration info via pci_dev */
|
|
info = dfl_fpga_enum_info_alloc(&pcidev->dev);
|
|
if (!info)
|
|
return -ENOMEM;
|
|
|
|
/* add irq info for enumeration if the device support irq */
|
|
nvec = cci_pci_alloc_irq(pcidev);
|
|
if (nvec < 0) {
|
|
dev_err(&pcidev->dev, "Fail to alloc irq %d.\n", nvec);
|
|
ret = nvec;
|
|
goto enum_info_free_exit;
|
|
} else if (nvec) {
|
|
irq_table = cci_pci_create_irq_table(pcidev, nvec);
|
|
if (!irq_table) {
|
|
ret = -ENOMEM;
|
|
goto irq_free_exit;
|
|
}
|
|
|
|
ret = dfl_fpga_enum_info_add_irq(info, nvec, irq_table);
|
|
kfree(irq_table);
|
|
if (ret)
|
|
goto irq_free_exit;
|
|
}
|
|
|
|
/* start to find Device Feature List from Bar 0 */
|
|
base = cci_pci_ioremap_bar(pcidev, 0);
|
|
if (!base) {
|
|
ret = -ENOMEM;
|
|
goto irq_free_exit;
|
|
}
|
|
|
|
/*
|
|
* PF device has FME and Ports/AFUs, and VF device only has one
|
|
* Port/AFU. Check them and add related "Device Feature List" info
|
|
* for the next step enumeration.
|
|
*/
|
|
if (dfl_feature_is_fme(base)) {
|
|
start = pci_resource_start(pcidev, 0);
|
|
len = pci_resource_len(pcidev, 0);
|
|
|
|
dfl_fpga_enum_info_add_dfl(info, start, len, base);
|
|
|
|
/*
|
|
* find more Device Feature Lists (e.g. Ports) per information
|
|
* indicated by FME module.
|
|
*/
|
|
v = readq(base + FME_HDR_CAP);
|
|
port_num = FIELD_GET(FME_CAP_NUM_PORTS, v);
|
|
|
|
WARN_ON(port_num > MAX_DFL_FPGA_PORT_NUM);
|
|
|
|
for (i = 0; i < port_num; i++) {
|
|
v = readq(base + FME_HDR_PORT_OFST(i));
|
|
|
|
/* skip ports which are not implemented. */
|
|
if (!(v & FME_PORT_OFST_IMP))
|
|
continue;
|
|
|
|
/*
|
|
* add Port's Device Feature List information for next
|
|
* step enumeration.
|
|
*/
|
|
bar = FIELD_GET(FME_PORT_OFST_BAR_ID, v);
|
|
offset = FIELD_GET(FME_PORT_OFST_DFH_OFST, v);
|
|
base = cci_pci_ioremap_bar(pcidev, bar);
|
|
if (!base)
|
|
continue;
|
|
|
|
start = pci_resource_start(pcidev, bar) + offset;
|
|
len = pci_resource_len(pcidev, bar) - offset;
|
|
|
|
dfl_fpga_enum_info_add_dfl(info, start, len,
|
|
base + offset);
|
|
}
|
|
} else if (dfl_feature_is_port(base)) {
|
|
start = pci_resource_start(pcidev, 0);
|
|
len = pci_resource_len(pcidev, 0);
|
|
|
|
dfl_fpga_enum_info_add_dfl(info, start, len, base);
|
|
} else {
|
|
ret = -ENODEV;
|
|
goto irq_free_exit;
|
|
}
|
|
|
|
/* start enumeration with prepared enumeration information */
|
|
cdev = dfl_fpga_feature_devs_enumerate(info);
|
|
if (IS_ERR(cdev)) {
|
|
dev_err(&pcidev->dev, "Enumeration failure\n");
|
|
ret = PTR_ERR(cdev);
|
|
goto irq_free_exit;
|
|
}
|
|
|
|
drvdata->cdev = cdev;
|
|
|
|
irq_free_exit:
|
|
if (ret)
|
|
cci_pci_free_irq(pcidev);
|
|
enum_info_free_exit:
|
|
dfl_fpga_enum_info_free(info);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static
|
|
int cci_pci_probe(struct pci_dev *pcidev, const struct pci_device_id *pcidevid)
|
|
{
|
|
int ret;
|
|
|
|
ret = pcim_enable_device(pcidev);
|
|
if (ret < 0) {
|
|
dev_err(&pcidev->dev, "Failed to enable device %d.\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
ret = pci_enable_pcie_error_reporting(pcidev);
|
|
if (ret && ret != -EINVAL)
|
|
dev_info(&pcidev->dev, "PCIE AER unavailable %d.\n", ret);
|
|
|
|
pci_set_master(pcidev);
|
|
|
|
if (!pci_set_dma_mask(pcidev, DMA_BIT_MASK(64))) {
|
|
ret = pci_set_consistent_dma_mask(pcidev, DMA_BIT_MASK(64));
|
|
if (ret)
|
|
goto disable_error_report_exit;
|
|
} else if (!pci_set_dma_mask(pcidev, DMA_BIT_MASK(32))) {
|
|
ret = pci_set_consistent_dma_mask(pcidev, DMA_BIT_MASK(32));
|
|
if (ret)
|
|
goto disable_error_report_exit;
|
|
} else {
|
|
ret = -EIO;
|
|
dev_err(&pcidev->dev, "No suitable DMA support available.\n");
|
|
goto disable_error_report_exit;
|
|
}
|
|
|
|
ret = cci_init_drvdata(pcidev);
|
|
if (ret) {
|
|
dev_err(&pcidev->dev, "Fail to init drvdata %d.\n", ret);
|
|
goto disable_error_report_exit;
|
|
}
|
|
|
|
ret = cci_enumerate_feature_devs(pcidev);
|
|
if (!ret)
|
|
return ret;
|
|
|
|
dev_err(&pcidev->dev, "enumeration failure %d.\n", ret);
|
|
|
|
disable_error_report_exit:
|
|
pci_disable_pcie_error_reporting(pcidev);
|
|
return ret;
|
|
}
|
|
|
|
static int cci_pci_sriov_configure(struct pci_dev *pcidev, int num_vfs)
|
|
{
|
|
struct cci_drvdata *drvdata = pci_get_drvdata(pcidev);
|
|
struct dfl_fpga_cdev *cdev = drvdata->cdev;
|
|
int ret = 0;
|
|
|
|
if (!num_vfs) {
|
|
/*
|
|
* disable SRIOV and then put released ports back to default
|
|
* PF access mode.
|
|
*/
|
|
pci_disable_sriov(pcidev);
|
|
|
|
dfl_fpga_cdev_config_ports_pf(cdev);
|
|
|
|
} else {
|
|
/*
|
|
* before enable SRIOV, put released ports into VF access mode
|
|
* first of all.
|
|
*/
|
|
ret = dfl_fpga_cdev_config_ports_vf(cdev, num_vfs);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = pci_enable_sriov(pcidev, num_vfs);
|
|
if (ret) {
|
|
dfl_fpga_cdev_config_ports_pf(cdev);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
return num_vfs;
|
|
}
|
|
|
|
static void cci_pci_remove(struct pci_dev *pcidev)
|
|
{
|
|
if (dev_is_pf(&pcidev->dev))
|
|
cci_pci_sriov_configure(pcidev, 0);
|
|
|
|
cci_remove_feature_devs(pcidev);
|
|
pci_disable_pcie_error_reporting(pcidev);
|
|
}
|
|
|
|
static struct pci_driver cci_pci_driver = {
|
|
.name = DRV_NAME,
|
|
.id_table = cci_pcie_id_tbl,
|
|
.probe = cci_pci_probe,
|
|
.remove = cci_pci_remove,
|
|
.sriov_configure = cci_pci_sriov_configure,
|
|
};
|
|
|
|
module_pci_driver(cci_pci_driver);
|
|
|
|
MODULE_DESCRIPTION("FPGA DFL PCIe Device Driver");
|
|
MODULE_AUTHOR("Intel Corporation");
|
|
MODULE_LICENSE("GPL v2");
|