| /* |
| * Copyright 2014 IBM Corp. |
| * |
| * This program is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU General Public License |
| * as published by the Free Software Foundation; either version |
| * 2 of the License, or (at your option) any later version. |
| */ |
| |
| #include <linux/pci.h> |
| #include <misc/cxl.h> |
| #include "cxl.h" |
| |
| static int cxl_dma_set_mask(struct pci_dev *pdev, u64 dma_mask) |
| { |
| if (dma_mask < DMA_BIT_MASK(64)) { |
| pr_info("%s only 64bit DMA supported on CXL", __func__); |
| return -EIO; |
| } |
| |
| *(pdev->dev.dma_mask) = dma_mask; |
| return 0; |
| } |
| |
| static int cxl_pci_probe_mode(struct pci_bus *bus) |
| { |
| return PCI_PROBE_NORMAL; |
| } |
| |
| static int cxl_setup_msi_irqs(struct pci_dev *pdev, int nvec, int type) |
| { |
| return -ENODEV; |
| } |
| |
| static void cxl_teardown_msi_irqs(struct pci_dev *pdev) |
| { |
| /* |
| * MSI should never be set but need still need to provide this call |
| * back. |
| */ |
| } |
| |
| static bool cxl_pci_enable_device_hook(struct pci_dev *dev) |
| { |
| struct pci_controller *phb; |
| struct cxl_afu *afu; |
| struct cxl_context *ctx; |
| |
| phb = pci_bus_to_host(dev->bus); |
| afu = (struct cxl_afu *)phb->private_data; |
| |
| if (!cxl_ops->link_ok(afu->adapter, afu)) { |
| dev_warn(&dev->dev, "%s: Device link is down, refusing to enable AFU\n", __func__); |
| return false; |
| } |
| |
| set_dma_ops(&dev->dev, &dma_nommu_ops); |
| set_dma_offset(&dev->dev, PAGE_OFFSET); |
| |
| /* |
| * Allocate a context to do cxl things too. If we eventually do real |
| * DMA ops, we'll need a default context to attach them to |
| */ |
| ctx = cxl_dev_context_init(dev); |
| if (IS_ERR(ctx)) |
| return false; |
| dev->dev.archdata.cxl_ctx = ctx; |
| |
| return (cxl_ops->afu_check_and_enable(afu) == 0); |
| } |
| |
| static void cxl_pci_disable_device(struct pci_dev *dev) |
| { |
| struct cxl_context *ctx = cxl_get_context(dev); |
| |
| if (ctx) { |
| if (ctx->status == STARTED) { |
| dev_err(&dev->dev, "Default context started\n"); |
| return; |
| } |
| dev->dev.archdata.cxl_ctx = NULL; |
| cxl_release_context(ctx); |
| } |
| } |
| |
| static resource_size_t cxl_pci_window_alignment(struct pci_bus *bus, |
| unsigned long type) |
| { |
| return 1; |
| } |
| |
| static void cxl_pci_reset_secondary_bus(struct pci_dev *dev) |
| { |
| /* Should we do an AFU reset here ? */ |
| } |
| |
| static int cxl_pcie_cfg_record(u8 bus, u8 devfn) |
| { |
| return (bus << 8) + devfn; |
| } |
| |
| static inline struct cxl_afu *pci_bus_to_afu(struct pci_bus *bus) |
| { |
| struct pci_controller *phb = bus ? pci_bus_to_host(bus) : NULL; |
| |
| return phb ? phb->private_data : NULL; |
| } |
| |
| static void cxl_afu_configured_put(struct cxl_afu *afu) |
| { |
| atomic_dec_if_positive(&afu->configured_state); |
| } |
| |
| static bool cxl_afu_configured_get(struct cxl_afu *afu) |
| { |
| return atomic_inc_unless_negative(&afu->configured_state); |
| } |
| |
| static inline int cxl_pcie_config_info(struct pci_bus *bus, unsigned int devfn, |
| struct cxl_afu *afu, int *_record) |
| { |
| int record; |
| |
| record = cxl_pcie_cfg_record(bus->number, devfn); |
| if (record > afu->crs_num) |
| return PCIBIOS_DEVICE_NOT_FOUND; |
| |
| *_record = record; |
| return 0; |
| } |
| |
| static int cxl_pcie_read_config(struct pci_bus *bus, unsigned int devfn, |
| int offset, int len, u32 *val) |
| { |
| int rc, record; |
| struct cxl_afu *afu; |
| u8 val8; |
| u16 val16; |
| u32 val32; |
| |
| afu = pci_bus_to_afu(bus); |
| /* Grab a reader lock on afu. */ |
| if (afu == NULL || !cxl_afu_configured_get(afu)) |
| return PCIBIOS_DEVICE_NOT_FOUND; |
| |
| rc = cxl_pcie_config_info(bus, devfn, afu, &record); |
| if (rc) |
| goto out; |
| |
| switch (len) { |
| case 1: |
| rc = cxl_ops->afu_cr_read8(afu, record, offset, &val8); |
| *val = val8; |
| break; |
| case 2: |
| rc = cxl_ops->afu_cr_read16(afu, record, offset, &val16); |
| *val = val16; |
| break; |
| case 4: |
| rc = cxl_ops->afu_cr_read32(afu, record, offset, &val32); |
| *val = val32; |
| break; |
| default: |
| WARN_ON(1); |
| } |
| |
| out: |
| cxl_afu_configured_put(afu); |
| return rc ? PCIBIOS_DEVICE_NOT_FOUND : PCIBIOS_SUCCESSFUL; |
| } |
| |
| static int cxl_pcie_write_config(struct pci_bus *bus, unsigned int devfn, |
| int offset, int len, u32 val) |
| { |
| int rc, record; |
| struct cxl_afu *afu; |
| |
| afu = pci_bus_to_afu(bus); |
| /* Grab a reader lock on afu. */ |
| if (afu == NULL || !cxl_afu_configured_get(afu)) |
| return PCIBIOS_DEVICE_NOT_FOUND; |
| |
| rc = cxl_pcie_config_info(bus, devfn, afu, &record); |
| if (rc) |
| goto out; |
| |
| switch (len) { |
| case 1: |
| rc = cxl_ops->afu_cr_write8(afu, record, offset, val & 0xff); |
| break; |
| case 2: |
| rc = cxl_ops->afu_cr_write16(afu, record, offset, val & 0xffff); |
| break; |
| case 4: |
| rc = cxl_ops->afu_cr_write32(afu, record, offset, val); |
| break; |
| default: |
| WARN_ON(1); |
| } |
| |
| out: |
| cxl_afu_configured_put(afu); |
| return rc ? PCIBIOS_SET_FAILED : PCIBIOS_SUCCESSFUL; |
| } |
| |
| static struct pci_ops cxl_pcie_pci_ops = |
| { |
| .read = cxl_pcie_read_config, |
| .write = cxl_pcie_write_config, |
| }; |
| |
| |
| static struct pci_controller_ops cxl_pci_controller_ops = |
| { |
| .probe_mode = cxl_pci_probe_mode, |
| .enable_device_hook = cxl_pci_enable_device_hook, |
| .disable_device = cxl_pci_disable_device, |
| .release_device = cxl_pci_disable_device, |
| .window_alignment = cxl_pci_window_alignment, |
| .reset_secondary_bus = cxl_pci_reset_secondary_bus, |
| .setup_msi_irqs = cxl_setup_msi_irqs, |
| .teardown_msi_irqs = cxl_teardown_msi_irqs, |
| .dma_set_mask = cxl_dma_set_mask, |
| }; |
| |
| int cxl_pci_vphb_add(struct cxl_afu *afu) |
| { |
| struct pci_controller *phb; |
| struct device_node *vphb_dn; |
| struct device *parent; |
| |
| /* |
| * If there are no AFU configuration records we won't have anything to |
| * expose under the vPHB, so skip creating one, returning success since |
| * this is still a valid case. This will also opt us out of EEH |
| * handling since we won't have anything special to do if there are no |
| * kernel drivers attached to the vPHB, and EEH handling is not yet |
| * supported in the peer model. |
| */ |
| if (!afu->crs_num) |
| return 0; |
| |
| /* The parent device is the adapter. Reuse the device node of |
| * the adapter. |
| * We don't seem to care what device node is used for the vPHB, |
| * but tools such as lsvpd walk up the device parents looking |
| * for a valid location code, so we might as well show devices |
| * attached to the adapter as being located on that adapter. |
| */ |
| parent = afu->adapter->dev.parent; |
| vphb_dn = parent->of_node; |
| |
| /* Alloc and setup PHB data structure */ |
| phb = pcibios_alloc_controller(vphb_dn); |
| if (!phb) |
| return -ENODEV; |
| |
| /* Setup parent in sysfs */ |
| phb->parent = parent; |
| |
| /* Setup the PHB using arch provided callback */ |
| phb->ops = &cxl_pcie_pci_ops; |
| phb->cfg_addr = NULL; |
| phb->cfg_data = NULL; |
| phb->private_data = afu; |
| phb->controller_ops = cxl_pci_controller_ops; |
| |
| /* Scan the bus */ |
| pcibios_scan_phb(phb); |
| if (phb->bus == NULL) |
| return -ENXIO; |
| |
| /* Set release hook on root bus */ |
| pci_set_host_bridge_release(to_pci_host_bridge(phb->bus->bridge), |
| pcibios_free_controller_deferred, |
| (void *) phb); |
| |
| /* Claim resources. This might need some rework as well depending |
| * whether we are doing probe-only or not, like assigning unassigned |
| * resources etc... |
| */ |
| pcibios_claim_one_bus(phb->bus); |
| |
| /* Add probed PCI devices to the device model */ |
| pci_bus_add_devices(phb->bus); |
| |
| afu->phb = phb; |
| |
| return 0; |
| } |
| |
| void cxl_pci_vphb_remove(struct cxl_afu *afu) |
| { |
| struct pci_controller *phb; |
| |
| /* If there is no configuration record we won't have one of these */ |
| if (!afu || !afu->phb) |
| return; |
| |
| phb = afu->phb; |
| afu->phb = NULL; |
| |
| pci_remove_root_bus(phb->bus); |
| /* |
| * We don't free phb here - that's handled by |
| * pcibios_free_controller_deferred() |
| */ |
| } |
| |
| bool cxl_pci_is_vphb_device(struct pci_dev *dev) |
| { |
| struct pci_controller *phb; |
| |
| phb = pci_bus_to_host(dev->bus); |
| |
| return (phb->ops == &cxl_pcie_pci_ops); |
| } |
| |
| struct cxl_afu *cxl_pci_to_afu(struct pci_dev *dev) |
| { |
| struct pci_controller *phb; |
| |
| phb = pci_bus_to_host(dev->bus); |
| |
| return (struct cxl_afu *)phb->private_data; |
| } |
| EXPORT_SYMBOL_GPL(cxl_pci_to_afu); |
| |
| unsigned int cxl_pci_to_cfg_record(struct pci_dev *dev) |
| { |
| return cxl_pcie_cfg_record(dev->bus->number, dev->devfn); |
| } |
| EXPORT_SYMBOL_GPL(cxl_pci_to_cfg_record); |