blob: 72d4e3f193af1ba5f7ffd704cd353ac1abddfd4b [file] [log] [blame]
Thomas Gleixner3b20eb22019-05-29 16:57:35 -07001// SPDX-License-Identifier: GPL-2.0-only
Weidong Hane28c31a2010-10-27 17:55:04 +01002/*
3 * Copyright (c) 2009, Intel Corporation.
4 *
Weidong Hane28c31a2010-10-27 17:55:04 +01005 * Author: Weidong Han <weidong.han@intel.com>
6 */
7
8#include <linux/pci.h>
Jan Beulich55e901f2011-09-22 09:17:57 +01009#include <linux/acpi.h>
Ross Lagerwall0b97b032015-04-09 08:05:10 +010010#include <linux/pci-acpi.h>
Oleksandr Andrushchenkoa67efff2021-10-28 17:36:20 +030011#include <xen/pci.h>
Weidong Hane28c31a2010-10-27 17:55:04 +010012#include <xen/xen.h>
13#include <xen/interface/physdev.h>
14#include <xen/interface/xen.h>
15
16#include <asm/xen/hypervisor.h>
17#include <asm/xen/hypercall.h>
18#include "../pci/pci.h"
Ben Hutchingsb7ef4a62013-12-31 20:46:27 +010019#ifdef CONFIG_PCI_MMCONFIG
Konrad Rzeszutek Wilk8deb3eb2013-10-25 16:26:02 -040020#include <asm/pci_x86.h>
Igor Druzhinina4098bc2019-09-12 19:31:51 +010021
22static int xen_mcfg_late(void);
Ben Hutchingsb7ef4a62013-12-31 20:46:27 +010023#endif
Weidong Hane28c31a2010-10-27 17:55:04 +010024
Jan Beulich55e901f2011-09-22 09:17:57 +010025static bool __read_mostly pci_seg_supported = true;
26
Weidong Hane28c31a2010-10-27 17:55:04 +010027static int xen_add_device(struct device *dev)
28{
29 int r;
30 struct pci_dev *pci_dev = to_pci_dev(dev);
Jan Beulich55e901f2011-09-22 09:17:57 +010031#ifdef CONFIG_PCI_IOV
32 struct pci_dev *physfn = pci_dev->physfn;
33#endif
Igor Druzhinina4098bc2019-09-12 19:31:51 +010034#ifdef CONFIG_PCI_MMCONFIG
35 static bool pci_mcfg_reserved = false;
36 /*
37 * Reserve MCFG areas in Xen on first invocation due to this being
38 * potentially called from inside of acpi_init immediately after
39 * MCFG table has been finally parsed.
40 */
41 if (!pci_mcfg_reserved) {
42 xen_mcfg_late();
43 pci_mcfg_reserved = true;
44 }
45#endif
Jan Beulich55e901f2011-09-22 09:17:57 +010046 if (pci_seg_supported) {
Boris Ostrovsky486edb22014-08-04 18:17:23 -040047 struct {
48 struct physdev_pci_device_add add;
49 uint32_t pxm;
50 } add_ext = {
51 .add.seg = pci_domain_nr(pci_dev->bus),
52 .add.bus = pci_dev->bus->number,
53 .add.devfn = pci_dev->devfn
Jan Beulich55e901f2011-09-22 09:17:57 +010054 };
Boris Ostrovsky486edb22014-08-04 18:17:23 -040055 struct physdev_pci_device_add *add = &add_ext.add;
56
Jan Beulich55e901f2011-09-22 09:17:57 +010057#ifdef CONFIG_ACPI
58 acpi_handle handle;
59#endif
Weidong Hane28c31a2010-10-27 17:55:04 +010060
61#ifdef CONFIG_PCI_IOV
Jan Beulich55e901f2011-09-22 09:17:57 +010062 if (pci_dev->is_virtfn) {
Boris Ostrovsky486edb22014-08-04 18:17:23 -040063 add->flags = XEN_PCI_DEV_VIRTFN;
64 add->physfn.bus = physfn->bus->number;
65 add->physfn.devfn = physfn->devfn;
Jan Beulich55e901f2011-09-22 09:17:57 +010066 } else
67#endif
68 if (pci_ari_enabled(pci_dev->bus) && PCI_SLOT(pci_dev->devfn))
Boris Ostrovsky486edb22014-08-04 18:17:23 -040069 add->flags = XEN_PCI_DEV_EXTFN;
Jan Beulich55e901f2011-09-22 09:17:57 +010070
71#ifdef CONFIG_ACPI
Rafael J. Wysocki3a83f992013-11-14 23:17:21 +010072 handle = ACPI_HANDLE(&pci_dev->dev);
Jan Beulich55e901f2011-09-22 09:17:57 +010073#ifdef CONFIG_PCI_IOV
74 if (!handle && pci_dev->is_virtfn)
Rafael J. Wysocki3a83f992013-11-14 23:17:21 +010075 handle = ACPI_HANDLE(physfn->bus->bridge);
Jan Beulich55e901f2011-09-22 09:17:57 +010076#endif
Ross Lagerwall0b97b032015-04-09 08:05:10 +010077 if (!handle) {
78 /*
79 * This device was not listed in the ACPI name space at
80 * all. Try to get acpi handle of parent pci bus.
81 */
82 struct pci_bus *pbus;
83 for (pbus = pci_dev->bus; pbus; pbus = pbus->parent) {
84 handle = acpi_pci_get_bridge_handle(pbus);
85 if (handle)
86 break;
87 }
88 }
Jan Beulich55e901f2011-09-22 09:17:57 +010089 if (handle) {
90 acpi_status status;
91
92 do {
93 unsigned long long pxm;
94
95 status = acpi_evaluate_integer(handle, "_PXM",
96 NULL, &pxm);
97 if (ACPI_SUCCESS(status)) {
Boris Ostrovsky486edb22014-08-04 18:17:23 -040098 add->optarr[0] = pxm;
99 add->flags |= XEN_PCI_DEV_PXM;
Jan Beulich55e901f2011-09-22 09:17:57 +0100100 break;
101 }
102 status = acpi_get_parent(handle, &handle);
103 } while (ACPI_SUCCESS(status));
104 }
105#endif /* CONFIG_ACPI */
106
Boris Ostrovsky486edb22014-08-04 18:17:23 -0400107 r = HYPERVISOR_physdev_op(PHYSDEVOP_pci_device_add, add);
Jan Beulich55e901f2011-09-22 09:17:57 +0100108 if (r != -ENOSYS)
109 return r;
110 pci_seg_supported = false;
111 }
112
113 if (pci_domain_nr(pci_dev->bus))
114 r = -ENOSYS;
115#ifdef CONFIG_PCI_IOV
116 else if (pci_dev->is_virtfn) {
Weidong Hane28c31a2010-10-27 17:55:04 +0100117 struct physdev_manage_pci_ext manage_pci_ext = {
118 .bus = pci_dev->bus->number,
119 .devfn = pci_dev->devfn,
120 .is_virtfn = 1,
Jan Beulich55e901f2011-09-22 09:17:57 +0100121 .physfn.bus = physfn->bus->number,
122 .physfn.devfn = physfn->devfn,
Weidong Hane28c31a2010-10-27 17:55:04 +0100123 };
124
125 r = HYPERVISOR_physdev_op(PHYSDEVOP_manage_pci_add_ext,
126 &manage_pci_ext);
Jan Beulich55e901f2011-09-22 09:17:57 +0100127 }
Weidong Hane28c31a2010-10-27 17:55:04 +0100128#endif
Jan Beulich55e901f2011-09-22 09:17:57 +0100129 else if (pci_ari_enabled(pci_dev->bus) && PCI_SLOT(pci_dev->devfn)) {
Weidong Hane28c31a2010-10-27 17:55:04 +0100130 struct physdev_manage_pci_ext manage_pci_ext = {
131 .bus = pci_dev->bus->number,
132 .devfn = pci_dev->devfn,
133 .is_extfn = 1,
134 };
135
136 r = HYPERVISOR_physdev_op(PHYSDEVOP_manage_pci_add_ext,
137 &manage_pci_ext);
138 } else {
139 struct physdev_manage_pci manage_pci = {
Ruslan Pisarev4b010982011-07-26 14:16:38 +0300140 .bus = pci_dev->bus->number,
Weidong Hane28c31a2010-10-27 17:55:04 +0100141 .devfn = pci_dev->devfn,
142 };
143
144 r = HYPERVISOR_physdev_op(PHYSDEVOP_manage_pci_add,
145 &manage_pci);
146 }
147
148 return r;
149}
150
151static int xen_remove_device(struct device *dev)
152{
153 int r;
154 struct pci_dev *pci_dev = to_pci_dev(dev);
Weidong Hane28c31a2010-10-27 17:55:04 +0100155
Jan Beulich55e901f2011-09-22 09:17:57 +0100156 if (pci_seg_supported) {
157 struct physdev_pci_device device = {
158 .seg = pci_domain_nr(pci_dev->bus),
159 .bus = pci_dev->bus->number,
160 .devfn = pci_dev->devfn
161 };
Weidong Hane28c31a2010-10-27 17:55:04 +0100162
Jan Beulich55e901f2011-09-22 09:17:57 +0100163 r = HYPERVISOR_physdev_op(PHYSDEVOP_pci_device_remove,
164 &device);
165 } else if (pci_domain_nr(pci_dev->bus))
166 r = -ENOSYS;
167 else {
168 struct physdev_manage_pci manage_pci = {
169 .bus = pci_dev->bus->number,
170 .devfn = pci_dev->devfn
171 };
172
173 r = HYPERVISOR_physdev_op(PHYSDEVOP_manage_pci_remove,
174 &manage_pci);
175 }
Weidong Hane28c31a2010-10-27 17:55:04 +0100176
177 return r;
178}
179
180static int xen_pci_notifier(struct notifier_block *nb,
181 unsigned long action, void *data)
182{
183 struct device *dev = data;
184 int r = 0;
185
186 switch (action) {
187 case BUS_NOTIFY_ADD_DEVICE:
188 r = xen_add_device(dev);
189 break;
190 case BUS_NOTIFY_DEL_DEVICE:
191 r = xen_remove_device(dev);
192 break;
193 default:
Jan Beulich12e13ac2011-08-17 09:32:32 +0100194 return NOTIFY_DONE;
Weidong Hane28c31a2010-10-27 17:55:04 +0100195 }
Jan Beulich12e13ac2011-08-17 09:32:32 +0100196 if (r)
197 dev_err(dev, "Failed to %s - passthrough or MSI/MSI-X might fail!\n",
198 action == BUS_NOTIFY_ADD_DEVICE ? "add" :
199 (action == BUS_NOTIFY_DEL_DEVICE ? "delete" : "?"));
200 return NOTIFY_OK;
Weidong Hane28c31a2010-10-27 17:55:04 +0100201}
202
Jan Beulich12e13ac2011-08-17 09:32:32 +0100203static struct notifier_block device_nb = {
Weidong Hane28c31a2010-10-27 17:55:04 +0100204 .notifier_call = xen_pci_notifier,
205};
206
207static int __init register_xen_pci_notifier(void)
208{
209 if (!xen_initial_domain())
210 return 0;
211
212 return bus_register_notifier(&pci_bus_type, &device_nb);
213}
214
215arch_initcall(register_xen_pci_notifier);
Konrad Rzeszutek Wilk8deb3eb2013-10-25 16:26:02 -0400216
217#ifdef CONFIG_PCI_MMCONFIG
Igor Druzhinina4098bc2019-09-12 19:31:51 +0100218static int xen_mcfg_late(void)
Konrad Rzeszutek Wilk8deb3eb2013-10-25 16:26:02 -0400219{
220 struct pci_mmcfg_region *cfg;
221 int rc;
222
223 if (!xen_initial_domain())
224 return 0;
225
226 if ((pci_probe & PCI_PROBE_MMCONF) == 0)
227 return 0;
228
229 if (list_empty(&pci_mmcfg_list))
230 return 0;
231
232 /* Check whether they are in the right area. */
233 list_for_each_entry(cfg, &pci_mmcfg_list, list) {
234 struct physdev_pci_mmcfg_reserved r;
235
236 r.address = cfg->address;
237 r.segment = cfg->segment;
238 r.start_bus = cfg->start_bus;
239 r.end_bus = cfg->end_bus;
240 r.flags = XEN_PCI_MMCFG_RESERVED;
241
242 rc = HYPERVISOR_physdev_op(PHYSDEVOP_pci_mmcfg_reserved, &r);
243 switch (rc) {
244 case 0:
245 case -ENOSYS:
246 continue;
247
248 default:
249 pr_warn("Failed to report MMCONFIG reservation"
250 " state for %s to hypervisor"
251 " (%d)\n",
252 cfg->name, rc);
253 }
254 }
255 return 0;
256}
Konrad Rzeszutek Wilk8deb3eb2013-10-25 16:26:02 -0400257#endif
Oleksandr Andrushchenkoa67efff2021-10-28 17:36:20 +0300258
259#ifdef CONFIG_XEN_DOM0
260struct xen_device_domain_owner {
261 domid_t domain;
262 struct pci_dev *dev;
263 struct list_head list;
264};
265
266static DEFINE_SPINLOCK(dev_domain_list_spinlock);
Cai Huoqingf66edf62022-02-09 11:28:41 +0800267static LIST_HEAD(dev_domain_list);
Oleksandr Andrushchenkoa67efff2021-10-28 17:36:20 +0300268
269static struct xen_device_domain_owner *find_device(struct pci_dev *dev)
270{
271 struct xen_device_domain_owner *owner;
272
273 list_for_each_entry(owner, &dev_domain_list, list) {
274 if (owner->dev == dev)
275 return owner;
276 }
277 return NULL;
278}
279
280int xen_find_device_domain_owner(struct pci_dev *dev)
281{
282 struct xen_device_domain_owner *owner;
283 int domain = -ENODEV;
284
285 spin_lock(&dev_domain_list_spinlock);
286 owner = find_device(dev);
287 if (owner)
288 domain = owner->domain;
289 spin_unlock(&dev_domain_list_spinlock);
290 return domain;
291}
292EXPORT_SYMBOL_GPL(xen_find_device_domain_owner);
293
294int xen_register_device_domain_owner(struct pci_dev *dev, uint16_t domain)
295{
296 struct xen_device_domain_owner *owner;
297
298 owner = kzalloc(sizeof(struct xen_device_domain_owner), GFP_KERNEL);
299 if (!owner)
300 return -ENODEV;
301
302 spin_lock(&dev_domain_list_spinlock);
303 if (find_device(dev)) {
304 spin_unlock(&dev_domain_list_spinlock);
305 kfree(owner);
306 return -EEXIST;
307 }
308 owner->domain = domain;
309 owner->dev = dev;
310 list_add_tail(&owner->list, &dev_domain_list);
311 spin_unlock(&dev_domain_list_spinlock);
312 return 0;
313}
314EXPORT_SYMBOL_GPL(xen_register_device_domain_owner);
315
316int xen_unregister_device_domain_owner(struct pci_dev *dev)
317{
318 struct xen_device_domain_owner *owner;
319
320 spin_lock(&dev_domain_list_spinlock);
321 owner = find_device(dev);
322 if (!owner) {
323 spin_unlock(&dev_domain_list_spinlock);
324 return -ENODEV;
325 }
326 list_del(&owner->list);
327 spin_unlock(&dev_domain_list_spinlock);
328 kfree(owner);
329 return 0;
330}
331EXPORT_SYMBOL_GPL(xen_unregister_device_domain_owner);
332#endif