Commit | Line | Data |
---|---|---|
8cfab3cf | 1 | // SPDX-License-Identifier: GPL-2.0 |
1da177e4 | 2 | /* |
2b937303 GKH |
3 | * (C) Copyright 2002-2004, 2007 Greg Kroah-Hartman <greg@kroah.com> |
4 | * (C) Copyright 2007 Novell Inc. | |
1da177e4 LT |
5 | */ |
6 | ||
7 | #include <linux/pci.h> | |
8 | #include <linux/module.h> | |
9 | #include <linux/init.h> | |
10 | #include <linux/device.h> | |
d42c6997 | 11 | #include <linux/mempolicy.h> |
4e57b681 TS |
12 | #include <linux/string.h> |
13 | #include <linux/slab.h> | |
8c65b4a6 | 14 | #include <linux/sched.h> |
69a18b18 | 15 | #include <linux/sched/isolation.h> |
873392ca | 16 | #include <linux/cpu.h> |
6cbf8214 | 17 | #include <linux/pm_runtime.h> |
eea3fc03 | 18 | #include <linux/suspend.h> |
4fc9bbf9 | 19 | #include <linux/kexec.h> |
07397df2 NG |
20 | #include <linux/of_device.h> |
21 | #include <linux/acpi.h> | |
a1fd09e8 | 22 | #include <linux/dma-map-ops.h> |
512881ea | 23 | #include <linux/iommu.h> |
1da177e4 | 24 | #include "pci.h" |
c6c889d9 | 25 | #include "pcie/portdrv.h" |
1da177e4 | 26 | |
75865858 GKH |
27 | struct pci_dynid { |
28 | struct list_head node; | |
29 | struct pci_device_id id; | |
30 | }; | |
1da177e4 | 31 | |
9dba910e TH |
32 | /** |
33 | * pci_add_dynid - add a new PCI device ID to this driver and re-probe devices | |
34 | * @drv: target pci driver | |
35 | * @vendor: PCI vendor ID | |
36 | * @device: PCI device ID | |
37 | * @subvendor: PCI subvendor ID | |
38 | * @subdevice: PCI subdevice ID | |
39 | * @class: PCI class | |
40 | * @class_mask: PCI class mask | |
41 | * @driver_data: private driver data | |
42 | * | |
43 | * Adds a new dynamic pci device ID to this driver and causes the | |
44 | * driver to probe for all devices again. @drv must have been | |
45 | * registered prior to calling this function. | |
46 | * | |
47 | * CONTEXT: | |
48 | * Does GFP_KERNEL allocation. | |
49 | * | |
50 | * RETURNS: | |
51 | * 0 on success, -errno on failure. | |
52 | */ | |
53 | int pci_add_dynid(struct pci_driver *drv, | |
54 | unsigned int vendor, unsigned int device, | |
55 | unsigned int subvendor, unsigned int subdevice, | |
56 | unsigned int class, unsigned int class_mask, | |
57 | unsigned long driver_data) | |
58 | { | |
59 | struct pci_dynid *dynid; | |
9dba910e TH |
60 | |
61 | dynid = kzalloc(sizeof(*dynid), GFP_KERNEL); | |
62 | if (!dynid) | |
63 | return -ENOMEM; | |
64 | ||
65 | dynid->id.vendor = vendor; | |
66 | dynid->id.device = device; | |
67 | dynid->id.subvendor = subvendor; | |
68 | dynid->id.subdevice = subdevice; | |
69 | dynid->id.class = class; | |
70 | dynid->id.class_mask = class_mask; | |
71 | dynid->id.driver_data = driver_data; | |
3d3c2ae1 | 72 | |
9dba910e TH |
73 | spin_lock(&drv->dynids.lock); |
74 | list_add_tail(&dynid->node, &drv->dynids.list); | |
75 | spin_unlock(&drv->dynids.lock); | |
76 | ||
3b7f1016 | 77 | return driver_attach(&drv->driver); |
9dba910e | 78 | } |
b7fe9434 | 79 | EXPORT_SYMBOL_GPL(pci_add_dynid); |
9dba910e TH |
80 | |
81 | static void pci_free_dynids(struct pci_driver *drv) | |
82 | { | |
83 | struct pci_dynid *dynid, *n; | |
3d3c2ae1 | 84 | |
9dba910e TH |
85 | spin_lock(&drv->dynids.lock); |
86 | list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { | |
87 | list_del(&dynid->node); | |
88 | kfree(dynid); | |
89 | } | |
90 | spin_unlock(&drv->dynids.lock); | |
91 | } | |
92 | ||
1f40704b ZD |
93 | /** |
94 | * pci_match_id - See if a PCI device matches a given pci_id table | |
95 | * @ids: array of PCI device ID structures to search in | |
96 | * @dev: the PCI device structure to match against. | |
97 | * | |
98 | * Used by a driver to check whether a PCI device is in its list of | |
99 | * supported devices. Returns the matching pci_device_id structure or | |
100 | * %NULL if there is no match. | |
101 | * | |
102 | * Deprecated; don't use this as it will not catch any dynamic IDs | |
103 | * that a driver might want to check for. | |
104 | */ | |
105 | const struct pci_device_id *pci_match_id(const struct pci_device_id *ids, | |
106 | struct pci_dev *dev) | |
107 | { | |
108 | if (ids) { | |
109 | while (ids->vendor || ids->subvendor || ids->class_mask) { | |
110 | if (pci_match_one_device(ids, dev)) | |
111 | return ids; | |
112 | ids++; | |
113 | } | |
114 | } | |
115 | return NULL; | |
116 | } | |
117 | EXPORT_SYMBOL(pci_match_id); | |
118 | ||
119 | static const struct pci_device_id pci_device_id_any = { | |
120 | .vendor = PCI_ANY_ID, | |
121 | .device = PCI_ANY_ID, | |
122 | .subvendor = PCI_ANY_ID, | |
123 | .subdevice = PCI_ANY_ID, | |
124 | }; | |
125 | ||
126 | /** | |
127 | * pci_match_device - See if a device matches a driver's list of IDs | |
128 | * @drv: the PCI driver to match against | |
129 | * @dev: the PCI device structure to match against | |
130 | * | |
131 | * Used by a driver to check whether a PCI device is in its list of | |
132 | * supported devices or in the dynids list, which may have been augmented | |
133 | * via the sysfs "new_id" file. Returns the matching pci_device_id | |
134 | * structure or %NULL if there is no match. | |
135 | */ | |
136 | static const struct pci_device_id *pci_match_device(struct pci_driver *drv, | |
137 | struct pci_dev *dev) | |
138 | { | |
139 | struct pci_dynid *dynid; | |
343b7258 | 140 | const struct pci_device_id *found_id = NULL, *ids; |
1f40704b ZD |
141 | |
142 | /* When driver_override is set, only bind to the matching driver */ | |
143 | if (dev->driver_override && strcmp(dev->driver_override, drv->name)) | |
144 | return NULL; | |
145 | ||
146 | /* Look at the dynamic ids first, before the static ones */ | |
147 | spin_lock(&drv->dynids.lock); | |
148 | list_for_each_entry(dynid, &drv->dynids.list, node) { | |
149 | if (pci_match_one_device(&dynid->id, dev)) { | |
150 | found_id = &dynid->id; | |
151 | break; | |
152 | } | |
153 | } | |
154 | spin_unlock(&drv->dynids.lock); | |
155 | ||
343b7258 MG |
156 | if (found_id) |
157 | return found_id; | |
158 | ||
159 | for (ids = drv->id_table; (found_id = pci_match_id(ids, dev)); | |
160 | ids = found_id + 1) { | |
161 | /* | |
162 | * The match table is split based on driver_override. | |
163 | * In case override_only was set, enforce driver_override | |
164 | * matching. | |
165 | */ | |
166 | if (found_id->override_only) { | |
167 | if (dev->driver_override) | |
168 | return found_id; | |
169 | } else { | |
170 | return found_id; | |
171 | } | |
172 | } | |
1f40704b ZD |
173 | |
174 | /* driver_override will always match, send a dummy id */ | |
343b7258 MG |
175 | if (dev->driver_override) |
176 | return &pci_device_id_any; | |
177 | return NULL; | |
1f40704b ZD |
178 | } |
179 | ||
1da177e4 | 180 | /** |
2f0cd59c | 181 | * new_id_store - sysfs frontend to pci_add_dynid() |
8f7020d3 RD |
182 | * @driver: target device driver |
183 | * @buf: buffer for scanning device ID data | |
184 | * @count: input size | |
1da177e4 | 185 | * |
9dba910e | 186 | * Allow PCI IDs to be added to an existing driver via sysfs. |
1da177e4 | 187 | */ |
a9427741 | 188 | static ssize_t new_id_store(struct device_driver *driver, const char *buf, |
3c78bc61 | 189 | size_t count) |
1da177e4 | 190 | { |
1da177e4 | 191 | struct pci_driver *pdrv = to_pci_driver(driver); |
b41d6cf3 | 192 | const struct pci_device_id *ids = pdrv->id_table; |
20a796a9 | 193 | u32 vendor, device, subvendor = PCI_ANY_ID, |
3c78bc61 RD |
194 | subdevice = PCI_ANY_ID, class = 0, class_mask = 0; |
195 | unsigned long driver_data = 0; | |
a49287d3 | 196 | int fields; |
8895d3bc | 197 | int retval = 0; |
1da177e4 | 198 | |
b41d6cf3 | 199 | fields = sscanf(buf, "%x %x %x %x %x %x %lx", |
1da177e4 LT |
200 | &vendor, &device, &subvendor, &subdevice, |
201 | &class, &class_mask, &driver_data); | |
6ba18636 | 202 | if (fields < 2) |
1da177e4 LT |
203 | return -EINVAL; |
204 | ||
8895d3bc BD |
205 | if (fields != 7) { |
206 | struct pci_dev *pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); | |
207 | if (!pdev) | |
208 | return -ENOMEM; | |
209 | ||
210 | pdev->vendor = vendor; | |
211 | pdev->device = device; | |
212 | pdev->subsystem_vendor = subvendor; | |
213 | pdev->subsystem_device = subdevice; | |
214 | pdev->class = class; | |
215 | ||
3853f912 | 216 | if (pci_match_device(pdrv, pdev)) |
8895d3bc BD |
217 | retval = -EEXIST; |
218 | ||
219 | kfree(pdev); | |
220 | ||
221 | if (retval) | |
222 | return retval; | |
223 | } | |
224 | ||
b41d6cf3 JD |
225 | /* Only accept driver_data values that match an existing id_table |
226 | entry */ | |
2debb4d2 CW |
227 | if (ids) { |
228 | retval = -EINVAL; | |
229 | while (ids->vendor || ids->subvendor || ids->class_mask) { | |
230 | if (driver_data == ids->driver_data) { | |
231 | retval = 0; | |
232 | break; | |
233 | } | |
234 | ids++; | |
b41d6cf3 | 235 | } |
2debb4d2 CW |
236 | if (retval) /* No match */ |
237 | return retval; | |
b41d6cf3 | 238 | } |
b41d6cf3 | 239 | |
9dba910e TH |
240 | retval = pci_add_dynid(pdrv, vendor, device, subvendor, subdevice, |
241 | class, class_mask, driver_data); | |
b19441af GKH |
242 | if (retval) |
243 | return retval; | |
1da177e4 LT |
244 | return count; |
245 | } | |
a9427741 | 246 | static DRIVER_ATTR_WO(new_id); |
1da177e4 | 247 | |
0994375e | 248 | /** |
2f0cd59c | 249 | * remove_id_store - remove a PCI device ID from this driver |
0994375e CW |
250 | * @driver: target device driver |
251 | * @buf: buffer for scanning device ID data | |
252 | * @count: input size | |
253 | * | |
254 | * Removes a dynamic pci device ID to this driver. | |
255 | */ | |
a9427741 | 256 | static ssize_t remove_id_store(struct device_driver *driver, const char *buf, |
3c78bc61 | 257 | size_t count) |
0994375e CW |
258 | { |
259 | struct pci_dynid *dynid, *n; | |
260 | struct pci_driver *pdrv = to_pci_driver(driver); | |
20a796a9 | 261 | u32 vendor, device, subvendor = PCI_ANY_ID, |
0994375e | 262 | subdevice = PCI_ANY_ID, class = 0, class_mask = 0; |
a49287d3 | 263 | int fields; |
9222097f | 264 | size_t retval = -ENODEV; |
0994375e CW |
265 | |
266 | fields = sscanf(buf, "%x %x %x %x %x %x", | |
267 | &vendor, &device, &subvendor, &subdevice, | |
268 | &class, &class_mask); | |
269 | if (fields < 2) | |
270 | return -EINVAL; | |
271 | ||
272 | spin_lock(&pdrv->dynids.lock); | |
273 | list_for_each_entry_safe(dynid, n, &pdrv->dynids.list, node) { | |
274 | struct pci_device_id *id = &dynid->id; | |
275 | if ((id->vendor == vendor) && | |
276 | (id->device == device) && | |
277 | (subvendor == PCI_ANY_ID || id->subvendor == subvendor) && | |
278 | (subdevice == PCI_ANY_ID || id->subdevice == subdevice) && | |
279 | !((id->class ^ class) & class_mask)) { | |
280 | list_del(&dynid->node); | |
281 | kfree(dynid); | |
9222097f | 282 | retval = count; |
0994375e CW |
283 | break; |
284 | } | |
285 | } | |
286 | spin_unlock(&pdrv->dynids.lock); | |
287 | ||
9222097f | 288 | return retval; |
0994375e | 289 | } |
a9427741 | 290 | static DRIVER_ATTR_WO(remove_id); |
0994375e | 291 | |
2229c1fb GKH |
292 | static struct attribute *pci_drv_attrs[] = { |
293 | &driver_attr_new_id.attr, | |
294 | &driver_attr_remove_id.attr, | |
295 | NULL, | |
bfb09a86 | 296 | }; |
2229c1fb | 297 | ATTRIBUTE_GROUPS(pci_drv); |
0994375e | 298 | |
873392ca RR |
299 | struct drv_dev_and_id { |
300 | struct pci_driver *drv; | |
301 | struct pci_dev *dev; | |
302 | const struct pci_device_id *id; | |
303 | }; | |
304 | ||
305 | static long local_pci_probe(void *_ddi) | |
306 | { | |
307 | struct drv_dev_and_id *ddi = _ddi; | |
967577b0 HY |
308 | struct pci_dev *pci_dev = ddi->dev; |
309 | struct pci_driver *pci_drv = ddi->drv; | |
310 | struct device *dev = &pci_dev->dev; | |
f3ec4f87 AS |
311 | int rc; |
312 | ||
967577b0 HY |
313 | /* |
314 | * Unbound PCI devices are always put in D0, regardless of | |
315 | * runtime PM status. During probe, the device is set to | |
316 | * active and the usage count is incremented. If the driver | |
a8360062 RW |
317 | * supports runtime PM, it should call pm_runtime_put_noidle(), |
318 | * or any other runtime PM helper function decrementing the usage | |
319 | * count, in its probe routine and pm_runtime_get_noresume() in | |
320 | * its remove routine. | |
f3ec4f87 | 321 | */ |
967577b0 | 322 | pm_runtime_get_sync(dev); |
68da4e0e | 323 | pci_dev->driver = pci_drv; |
967577b0 | 324 | rc = pci_drv->probe(pci_dev, ddi->id); |
f92d74c1 SC |
325 | if (!rc) |
326 | return rc; | |
327 | if (rc < 0) { | |
68da4e0e | 328 | pci_dev->driver = NULL; |
967577b0 | 329 | pm_runtime_put_sync(dev); |
f92d74c1 | 330 | return rc; |
f3ec4f87 | 331 | } |
f92d74c1 SC |
332 | /* |
333 | * Probe function should return < 0 for failure, 0 for success | |
334 | * Treat values > 0 as success, but warn. | |
335 | */ | |
6941a0c2 BH |
336 | pci_warn(pci_dev, "Driver probe function unexpectedly returned %d\n", |
337 | rc); | |
f92d74c1 | 338 | return 0; |
873392ca RR |
339 | } |
340 | ||
0b2c2a71 TG |
341 | static bool pci_physfn_is_probed(struct pci_dev *dev) |
342 | { | |
343 | #ifdef CONFIG_PCI_IOV | |
344 | return dev->is_virtfn && dev->physfn->is_probed; | |
345 | #else | |
346 | return false; | |
347 | #endif | |
348 | } | |
349 | ||
d42c6997 AK |
350 | static int pci_call_probe(struct pci_driver *drv, struct pci_dev *dev, |
351 | const struct pci_device_id *id) | |
352 | { | |
0b2c2a71 | 353 | int error, node, cpu; |
873392ca RR |
354 | struct drv_dev_and_id ddi = { drv, dev, id }; |
355 | ||
12c3156f AD |
356 | /* |
357 | * Execute driver initialization on node where the device is | |
358 | * attached. This way the driver likely allocates its local memory | |
359 | * on the right node. | |
360 | */ | |
873392ca | 361 | node = dev_to_node(&dev->dev); |
0b2c2a71 TG |
362 | dev->is_probed = 1; |
363 | ||
364 | cpu_hotplug_disable(); | |
12c3156f AD |
365 | |
366 | /* | |
0b2c2a71 TG |
367 | * Prevent nesting work_on_cpu() for the case where a Virtual Function |
368 | * device is probed from work_on_cpu() of the Physical device. | |
12c3156f | 369 | */ |
0b2c2a71 | 370 | if (node < 0 || node >= MAX_NUMNODES || !node_online(node) || |
9d42ea0d | 371 | pci_physfn_is_probed(dev)) { |
0b2c2a71 | 372 | cpu = nr_cpu_ids; |
9d42ea0d FW |
373 | } else { |
374 | cpumask_var_t wq_domain_mask; | |
375 | ||
376 | if (!zalloc_cpumask_var(&wq_domain_mask, GFP_KERNEL)) { | |
377 | error = -ENOMEM; | |
378 | goto out; | |
379 | } | |
380 | cpumask_and(wq_domain_mask, | |
04d4e665 FW |
381 | housekeeping_cpumask(HK_TYPE_WQ), |
382 | housekeeping_cpumask(HK_TYPE_DOMAIN)); | |
9d42ea0d | 383 | |
69a18b18 | 384 | cpu = cpumask_any_and(cpumask_of_node(node), |
9d42ea0d FW |
385 | wq_domain_mask); |
386 | free_cpumask_var(wq_domain_mask); | |
387 | } | |
0b2c2a71 TG |
388 | |
389 | if (cpu < nr_cpu_ids) | |
390 | error = work_on_cpu(cpu, local_pci_probe, &ddi); | |
391 | else | |
873392ca | 392 | error = local_pci_probe(&ddi); |
9d42ea0d | 393 | out: |
0b2c2a71 TG |
394 | dev->is_probed = 0; |
395 | cpu_hotplug_enable(); | |
d42c6997 AK |
396 | return error; |
397 | } | |
398 | ||
1da177e4 | 399 | /** |
23ea3793 | 400 | * __pci_device_probe - check if a driver wants to claim a specific PCI device |
8f7020d3 RD |
401 | * @drv: driver to call to check if it wants the PCI device |
402 | * @pci_dev: PCI device being probed | |
f7625980 | 403 | * |
8f7020d3 | 404 | * returns 0 on success, else error. |
68da4e0e | 405 | * side-effect: pci_dev->driver is set to drv when drv claims pci_dev. |
1da177e4 | 406 | */ |
3c78bc61 | 407 | static int __pci_device_probe(struct pci_driver *drv, struct pci_dev *pci_dev) |
75865858 GKH |
408 | { |
409 | const struct pci_device_id *id; | |
1da177e4 LT |
410 | int error = 0; |
411 | ||
ae232f09 | 412 | if (drv->probe) { |
75865858 GKH |
413 | error = -ENODEV; |
414 | ||
415 | id = pci_match_device(drv, pci_dev); | |
416 | if (id) | |
d42c6997 | 417 | error = pci_call_probe(drv, pci_dev, id); |
1da177e4 LT |
418 | } |
419 | return error; | |
420 | } | |
421 | ||
890e4847 JL |
422 | int __weak pcibios_alloc_irq(struct pci_dev *dev) |
423 | { | |
424 | return 0; | |
425 | } | |
426 | ||
427 | void __weak pcibios_free_irq(struct pci_dev *dev) | |
428 | { | |
429 | } | |
430 | ||
0e7df224 BW |
431 | #ifdef CONFIG_PCI_IOV |
432 | static inline bool pci_device_can_probe(struct pci_dev *pdev) | |
433 | { | |
2d2f4273 AW |
434 | return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe || |
435 | pdev->driver_override); | |
0e7df224 BW |
436 | } |
437 | #else | |
438 | static inline bool pci_device_can_probe(struct pci_dev *pdev) | |
439 | { | |
440 | return true; | |
441 | } | |
442 | #endif | |
443 | ||
3c78bc61 | 444 | static int pci_device_probe(struct device *dev) |
1da177e4 | 445 | { |
890e4847 JL |
446 | int error; |
447 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
448 | struct pci_driver *drv = to_pci_driver(dev->driver); | |
449 | ||
76002d8b AW |
450 | if (!pci_device_can_probe(pci_dev)) |
451 | return -ENODEV; | |
452 | ||
30fdfb92 MM |
453 | pci_assign_irq(pci_dev); |
454 | ||
890e4847 JL |
455 | error = pcibios_alloc_irq(pci_dev); |
456 | if (error < 0) | |
457 | return error; | |
1da177e4 | 458 | |
1da177e4 | 459 | pci_dev_get(pci_dev); |
76002d8b AW |
460 | error = __pci_device_probe(drv, pci_dev); |
461 | if (error) { | |
462 | pcibios_free_irq(pci_dev); | |
463 | pci_dev_put(pci_dev); | |
890e4847 | 464 | } |
1da177e4 LT |
465 | |
466 | return error; | |
467 | } | |
468 | ||
fc7a6209 | 469 | static void pci_device_remove(struct device *dev) |
1da177e4 | 470 | { |
3c78bc61 | 471 | struct pci_dev *pci_dev = to_pci_dev(dev); |
e0217c5b | 472 | struct pci_driver *drv = pci_dev->driver; |
1da177e4 | 473 | |
097d9d41 UKK |
474 | if (drv->remove) { |
475 | pm_runtime_get_sync(dev); | |
476 | drv->remove(pci_dev); | |
477 | pm_runtime_put_noidle(dev); | |
1da177e4 | 478 | } |
097d9d41 | 479 | pcibios_free_irq(pci_dev); |
68da4e0e | 480 | pci_dev->driver = NULL; |
097d9d41 | 481 | pci_iov_remove(pci_dev); |
1da177e4 | 482 | |
f3ec4f87 | 483 | /* Undo the runtime PM settings in local_pci_probe() */ |
967577b0 | 484 | pm_runtime_put_sync(dev); |
f3ec4f87 | 485 | |
2449e06a SL |
486 | /* |
487 | * If the device is still on, set the power state as "unknown", | |
488 | * since it might change by the next time we load the driver. | |
489 | */ | |
490 | if (pci_dev->current_state == PCI_D0) | |
491 | pci_dev->current_state = PCI_UNKNOWN; | |
492 | ||
1da177e4 LT |
493 | /* |
494 | * We would love to complain here if pci_dev->is_enabled is set, that | |
495 | * the driver should have called pci_disable_device(), but the | |
496 | * unfortunate fact is there are too many odd BIOS and bridge setups | |
f7625980 | 497 | * that don't like drivers doing that all of the time. |
1da177e4 LT |
498 | * Oh well, we can dream of sane hardware when we sleep, no matter how |
499 | * horrible the crap we have to deal with is when we are awake... | |
500 | */ | |
501 | ||
502 | pci_dev_put(pci_dev); | |
1da177e4 LT |
503 | } |
504 | ||
bbb44d9f RW |
505 | static void pci_device_shutdown(struct device *dev) |
506 | { | |
507 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
e0217c5b | 508 | struct pci_driver *drv = pci_dev->driver; |
bbb44d9f | 509 | |
3ff2de9b HY |
510 | pm_runtime_resume(dev); |
511 | ||
bbb44d9f RW |
512 | if (drv && drv->shutdown) |
513 | drv->shutdown(pci_dev); | |
5b415f1e | 514 | |
b566a22c | 515 | /* |
4fc9bbf9 KA |
516 | * If this is a kexec reboot, turn off Bus Master bit on the |
517 | * device to tell it to not continue to do DMA. Don't touch | |
518 | * devices in D3cold or unknown states. | |
519 | * If it is not a kexec reboot, firmware will hit the PCI | |
520 | * devices with big hammer and stop their DMA any way. | |
b566a22c | 521 | */ |
4fc9bbf9 | 522 | if (kexec_in_progress && (pci_dev->current_state <= PCI_D3hot)) |
6e0eda3c | 523 | pci_clear_master(pci_dev); |
bbb44d9f RW |
524 | } |
525 | ||
18a94192 | 526 | #ifdef CONFIG_PM_SLEEP |
6cbf8214 | 527 | |
18a94192 | 528 | /* Auxiliary functions used for system resume */ |
6cbf8214 RW |
529 | |
530 | /** | |
531 | * pci_restore_standard_config - restore standard config registers of PCI device | |
532 | * @pci_dev: PCI device to handle | |
533 | */ | |
534 | static int pci_restore_standard_config(struct pci_dev *pci_dev) | |
535 | { | |
536 | pci_update_current_state(pci_dev, PCI_UNKNOWN); | |
537 | ||
538 | if (pci_dev->current_state != PCI_D0) { | |
539 | int error = pci_set_power_state(pci_dev, PCI_D0); | |
540 | if (error) | |
541 | return error; | |
542 | } | |
543 | ||
1d3c16a8 | 544 | pci_restore_state(pci_dev); |
0ce3fcaf | 545 | pci_pme_restore(pci_dev); |
1d3c16a8 | 546 | return 0; |
6cbf8214 | 547 | } |
18a94192 KK |
548 | #endif /* CONFIG_PM_SLEEP */ |
549 | ||
550 | #ifdef CONFIG_PM | |
551 | ||
552 | /* Auxiliary functions used for system resume and run-time resume */ | |
6cbf8214 | 553 | |
f7b32a86 BH |
554 | static void pci_pm_default_resume(struct pci_dev *pci_dev) |
555 | { | |
556 | pci_fixup_device(pci_fixup_resume, pci_dev); | |
557 | pci_enable_wake(pci_dev, PCI_D0, false); | |
558 | } | |
559 | ||
0f40ac35 | 560 | static void pci_pm_power_up_and_verify_state(struct pci_dev *pci_dev) |
6cbf8214 | 561 | { |
db288c9c | 562 | pci_power_up(pci_dev); |
81cfa590 | 563 | pci_update_current_state(pci_dev, PCI_D0); |
0f40ac35 RW |
564 | } |
565 | ||
566 | static void pci_pm_default_resume_early(struct pci_dev *pci_dev) | |
567 | { | |
568 | pci_pm_power_up_and_verify_state(pci_dev); | |
db288c9c | 569 | pci_restore_state(pci_dev); |
0ce3fcaf | 570 | pci_pme_restore(pci_dev); |
6cbf8214 RW |
571 | } |
572 | ||
730643d3 RW |
573 | static void pci_pm_bridge_power_up_actions(struct pci_dev *pci_dev) |
574 | { | |
e74b2b58 MW |
575 | pci_bridge_wait_for_secondary_bus(pci_dev, "resume"); |
576 | ||
730643d3 RW |
577 | /* |
578 | * When powering on a bridge from D3cold, the whole hierarchy may be | |
579 | * powered on into D0uninitialized state, resume them to give them a | |
580 | * chance to suspend again | |
581 | */ | |
582 | pci_resume_bus(pci_dev->subordinate); | |
583 | } | |
584 | ||
18a94192 KK |
585 | #endif /* CONFIG_PM */ |
586 | ||
587 | #ifdef CONFIG_PM_SLEEP | |
588 | ||
fa58d305 RW |
589 | /* |
590 | * Default "suspend" method for devices that have no driver provided suspend, | |
591 | * or not even a driver at all (second part). | |
bbb44d9f | 592 | */ |
bb808945 | 593 | static void pci_pm_set_unknown_state(struct pci_dev *pci_dev) |
bbb44d9f | 594 | { |
bbb44d9f RW |
595 | /* |
596 | * mark its power state as "unknown", since we don't know if | |
597 | * e.g. the BIOS will change its device state when we suspend. | |
598 | */ | |
599 | if (pci_dev->current_state == PCI_D0) | |
600 | pci_dev->current_state = PCI_UNKNOWN; | |
601 | } | |
602 | ||
355a72d7 RW |
603 | /* |
604 | * Default "resume" method for devices that have no driver provided resume, | |
605 | * or not even a driver at all (second part). | |
606 | */ | |
bb808945 | 607 | static int pci_pm_reenable_device(struct pci_dev *pci_dev) |
355a72d7 RW |
608 | { |
609 | int retval; | |
610 | ||
b2105b9f | 611 | /* if the device was enabled before suspend, re-enable */ |
bbb44d9f RW |
612 | retval = pci_reenable_device(pci_dev); |
613 | /* | |
614 | * if the device was busmaster before the suspend, make it busmaster | |
615 | * again | |
616 | */ | |
617 | if (pci_dev->is_busmaster) | |
618 | pci_set_master(pci_dev); | |
619 | ||
620 | return retval; | |
621 | } | |
622 | ||
623 | static int pci_legacy_suspend(struct device *dev, pm_message_t state) | |
1da177e4 | 624 | { |
3c78bc61 | 625 | struct pci_dev *pci_dev = to_pci_dev(dev); |
e0217c5b | 626 | struct pci_driver *drv = pci_dev->driver; |
46939f8b | 627 | |
02669492 | 628 | if (drv && drv->suspend) { |
99dadce8 | 629 | pci_power_t prev = pci_dev->current_state; |
46939f8b | 630 | int error; |
aa8c6c93 | 631 | |
57ef8026 | 632 | error = drv->suspend(pci_dev, state); |
a759de69 | 633 | suspend_report_result(dev, drv->suspend, error); |
57ef8026 FP |
634 | if (error) |
635 | return error; | |
aa8c6c93 | 636 | |
46939f8b | 637 | if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 |
99dadce8 | 638 | && pci_dev->current_state != PCI_UNKNOWN) { |
12bcae44 BH |
639 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
640 | "PCI PM: Device state not saved by %pS\n", | |
641 | drv->suspend); | |
99dadce8 | 642 | } |
02669492 | 643 | } |
ad8cfa1d RW |
644 | |
645 | pci_fixup_device(pci_fixup_suspend, pci_dev); | |
646 | ||
46939f8b | 647 | return 0; |
1da177e4 LT |
648 | } |
649 | ||
5984de0b | 650 | static int pci_legacy_suspend_late(struct device *dev) |
cbd69dbb | 651 | { |
3c78bc61 | 652 | struct pci_dev *pci_dev = to_pci_dev(dev); |
46939f8b RW |
653 | |
654 | if (!pci_dev->state_saved) | |
655 | pci_save_state(pci_dev); | |
656 | ||
657 | pci_pm_set_unknown_state(pci_dev); | |
658 | ||
7d2a01b8 AN |
659 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); |
660 | ||
46939f8b | 661 | return 0; |
cbd69dbb | 662 | } |
1da177e4 | 663 | |
bbb44d9f | 664 | static int pci_legacy_resume(struct device *dev) |
1da177e4 | 665 | { |
3c78bc61 | 666 | struct pci_dev *pci_dev = to_pci_dev(dev); |
e0217c5b | 667 | struct pci_driver *drv = pci_dev->driver; |
1da177e4 | 668 | |
ad8cfa1d RW |
669 | pci_fixup_device(pci_fixup_resume, pci_dev); |
670 | ||
aa8c6c93 RW |
671 | return drv && drv->resume ? |
672 | drv->resume(pci_dev) : pci_pm_reenable_device(pci_dev); | |
1da177e4 LT |
673 | } |
674 | ||
571ff758 RW |
675 | /* Auxiliary functions used by the new power management framework */ |
676 | ||
5294e256 | 677 | static void pci_pm_default_suspend(struct pci_dev *pci_dev) |
73410429 | 678 | { |
5294e256 | 679 | /* Disable non-bridge devices without PM support */ |
326c1cda | 680 | if (!pci_has_subordinate(pci_dev)) |
cbbc2f6b | 681 | pci_disable_enabled_device(pci_dev); |
73410429 RW |
682 | } |
683 | ||
07e836e8 RW |
684 | static bool pci_has_legacy_pm_support(struct pci_dev *pci_dev) |
685 | { | |
e0217c5b | 686 | struct pci_driver *drv = pci_dev->driver; |
1a1daf09 | 687 | bool ret = drv && (drv->suspend || drv->resume); |
bb808945 RW |
688 | |
689 | /* | |
690 | * Legacy PM support is used by default, so warn if the new framework is | |
691 | * supported as well. Drivers are supposed to support either the | |
692 | * former, or the latter, but not both at the same time. | |
693 | */ | |
12bcae44 BH |
694 | pci_WARN(pci_dev, ret && drv->driver.pm, "device %04x:%04x\n", |
695 | pci_dev->vendor, pci_dev->device); | |
bb808945 RW |
696 | |
697 | return ret; | |
07e836e8 RW |
698 | } |
699 | ||
571ff758 RW |
700 | /* New power management framework */ |
701 | ||
bbb44d9f RW |
702 | static int pci_pm_prepare(struct device *dev) |
703 | { | |
0c7376ad | 704 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 705 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 706 | |
6da2f2cc BH |
707 | if (pm && pm->prepare) { |
708 | int error = pm->prepare(dev); | |
08810a41 | 709 | if (error < 0) |
bac2a909 | 710 | return error; |
08810a41 RW |
711 | |
712 | if (!error && dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_PREPARE)) | |
713 | return 0; | |
bac2a909 | 714 | } |
0c7376ad RW |
715 | if (pci_dev_need_resume(pci_dev)) |
716 | return 0; | |
717 | ||
718 | /* | |
719 | * The PME setting needs to be adjusted here in case the direct-complete | |
720 | * optimization is used with respect to this device. | |
721 | */ | |
722 | pci_dev_adjust_pme(pci_dev); | |
723 | return 1; | |
bbb44d9f RW |
724 | } |
725 | ||
2cef548a RW |
726 | static void pci_pm_complete(struct device *dev) |
727 | { | |
a0d2a959 LW |
728 | struct pci_dev *pci_dev = to_pci_dev(dev); |
729 | ||
730 | pci_dev_complete_resume(pci_dev); | |
731 | pm_generic_complete(dev); | |
732 | ||
733 | /* Resume device if platform firmware has put it in reset-power-on */ | |
bd755d77 | 734 | if (pm_runtime_suspended(dev) && pm_resume_via_firmware()) { |
a0d2a959 LW |
735 | pci_power_t pre_sleep_state = pci_dev->current_state; |
736 | ||
b51033e0 RW |
737 | pci_refresh_power_state(pci_dev); |
738 | /* | |
739 | * On platforms with ACPI this check may also trigger for | |
740 | * devices sharing power resources if one of those power | |
741 | * resources has been activated as a result of a change of the | |
742 | * power state of another device sharing it. However, in that | |
743 | * case it is also better to resume the device, in general. | |
744 | */ | |
a0d2a959 LW |
745 | if (pci_dev->current_state < pre_sleep_state) |
746 | pm_request_resume(dev); | |
747 | } | |
2cef548a | 748 | } |
bbb44d9f | 749 | |
6cbf8214 RW |
750 | #else /* !CONFIG_PM_SLEEP */ |
751 | ||
752 | #define pci_pm_prepare NULL | |
2cef548a | 753 | #define pci_pm_complete NULL |
6cbf8214 RW |
754 | |
755 | #endif /* !CONFIG_PM_SLEEP */ | |
756 | ||
bbb44d9f | 757 | #ifdef CONFIG_SUSPEND |
a39bd851 BH |
758 | static void pcie_pme_root_status_cleanup(struct pci_dev *pci_dev) |
759 | { | |
760 | /* | |
761 | * Some BIOSes forget to clear Root PME Status bits after system | |
762 | * wakeup, which breaks ACPI-based runtime wakeup on PCI Express. | |
763 | * Clear those bits now just in case (shouldn't hurt). | |
764 | */ | |
765 | if (pci_is_pcie(pci_dev) && | |
3620c714 BH |
766 | (pci_pcie_type(pci_dev) == PCI_EXP_TYPE_ROOT_PORT || |
767 | pci_pcie_type(pci_dev) == PCI_EXP_TYPE_RC_EC)) | |
a39bd851 BH |
768 | pcie_clear_root_pme_status(pci_dev); |
769 | } | |
bbb44d9f RW |
770 | |
771 | static int pci_pm_suspend(struct device *dev) | |
772 | { | |
773 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
8150f32b | 774 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 775 | |
d491f2b7 RW |
776 | pci_dev->skip_bus_pm = false; |
777 | ||
c01163db BH |
778 | /* |
779 | * Disabling PTM allows some systems, e.g., Intel mobile chips | |
780 | * since Coffee Lake, to enter a lower-power PM state. | |
781 | */ | |
782 | pci_suspend_ptm(pci_dev); | |
783 | ||
ad8cfa1d RW |
784 | if (pci_has_legacy_pm_support(pci_dev)) |
785 | return pci_legacy_suspend(dev, PMSG_SUSPEND); | |
bb808945 | 786 | |
5294e256 RW |
787 | if (!pm) { |
788 | pci_pm_default_suspend(pci_dev); | |
c4b65157 | 789 | return 0; |
5294e256 RW |
790 | } |
791 | ||
7cd0602d | 792 | /* |
c4b65157 RW |
793 | * PCI devices suspended at run time may need to be resumed at this |
794 | * point, because in general it may be necessary to reconfigure them for | |
795 | * system suspend. Namely, if the device is expected to wake up the | |
796 | * system from the sleep state, it may have to be reconfigured for this | |
797 | * purpose, or if the device is not expected to wake up the system from | |
798 | * the sleep state, it should be prevented from signaling wakeup events | |
799 | * going forward. | |
800 | * | |
801 | * Also if the driver of the device does not indicate that its system | |
802 | * suspend callbacks can cope with runtime-suspended devices, it is | |
803 | * better to resume the device from runtime suspend here. | |
7cd0602d | 804 | */ |
c4b65157 | 805 | if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || |
0c7376ad | 806 | pci_dev_need_resume(pci_dev)) { |
c4b65157 | 807 | pm_runtime_resume(dev); |
656088aa | 808 | pci_dev->state_saved = false; |
0c7376ad RW |
809 | } else { |
810 | pci_dev_adjust_pme(pci_dev); | |
656088aa | 811 | } |
7cd0602d | 812 | |
5294e256 RW |
813 | if (pm->suspend) { |
814 | pci_power_t prev = pci_dev->current_state; | |
815 | int error; | |
816 | ||
ddb7c9d2 | 817 | error = pm->suspend(dev); |
a759de69 | 818 | suspend_report_result(dev, pm->suspend, error); |
5294e256 RW |
819 | if (error) |
820 | return error; | |
821 | ||
46939f8b | 822 | if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 |
5294e256 | 823 | && pci_dev->current_state != PCI_UNKNOWN) { |
12bcae44 BH |
824 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
825 | "PCI PM: State of device not saved by %pS\n", | |
826 | pm->suspend); | |
5294e256 | 827 | } |
bbb44d9f | 828 | } |
fa58d305 | 829 | |
5294e256 | 830 | return 0; |
bbb44d9f RW |
831 | } |
832 | ||
c4b65157 RW |
833 | static int pci_pm_suspend_late(struct device *dev) |
834 | { | |
fa2bfead | 835 | if (dev_pm_skip_suspend(dev)) |
c4b65157 RW |
836 | return 0; |
837 | ||
838 | pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); | |
839 | ||
840 | return pm_generic_suspend_late(dev); | |
841 | } | |
842 | ||
bbb44d9f | 843 | static int pci_pm_suspend_noirq(struct device *dev) |
c8958177 | 844 | { |
355a72d7 | 845 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 846 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
c8958177 | 847 | |
fa2bfead | 848 | if (dev_pm_skip_suspend(dev)) |
c4b65157 RW |
849 | return 0; |
850 | ||
bb808945 | 851 | if (pci_has_legacy_pm_support(pci_dev)) |
5984de0b | 852 | return pci_legacy_suspend_late(dev); |
bb808945 | 853 | |
931ff68a RW |
854 | if (!pm) { |
855 | pci_save_state(pci_dev); | |
7d2a01b8 | 856 | goto Fixup; |
931ff68a | 857 | } |
46939f8b RW |
858 | |
859 | if (pm->suspend_noirq) { | |
860 | pci_power_t prev = pci_dev->current_state; | |
861 | int error; | |
862 | ||
863 | error = pm->suspend_noirq(dev); | |
a759de69 | 864 | suspend_report_result(dev, pm->suspend_noirq, error); |
46939f8b RW |
865 | if (error) |
866 | return error; | |
867 | ||
868 | if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 | |
869 | && pci_dev->current_state != PCI_UNKNOWN) { | |
12bcae44 BH |
870 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
871 | "PCI PM: State of device not saved by %pS\n", | |
872 | pm->suspend_noirq); | |
7d2a01b8 | 873 | goto Fixup; |
46939f8b | 874 | } |
bbb44d9f RW |
875 | } |
876 | ||
4c00cba1 RJ |
877 | if (!pci_dev->state_saved) { |
878 | pci_save_state(pci_dev); | |
879 | ||
d491f2b7 | 880 | /* |
4c00cba1 RJ |
881 | * If the device is a bridge with a child in D0 below it, |
882 | * it needs to stay in D0, so check skip_bus_pm to avoid | |
883 | * putting it into a low-power state in that case. | |
d491f2b7 | 884 | */ |
4c00cba1 | 885 | if (!pci_dev->skip_bus_pm && pci_power_manageable(pci_dev)) |
46939f8b RW |
886 | pci_prepare_to_sleep(pci_dev); |
887 | } | |
d67e37d7 | 888 | |
6941a0c2 | 889 | pci_dbg(pci_dev, "PCI PM: Suspend power state: %s\n", |
ca67ab5c RW |
890 | pci_power_name(pci_dev->current_state)); |
891 | ||
3e26c5fe RW |
892 | if (pci_dev->current_state == PCI_D0) { |
893 | pci_dev->skip_bus_pm = true; | |
894 | /* | |
895 | * Per PCI PM r1.2, table 6-1, a bridge must be in D0 if any | |
896 | * downstream device is in D0, so avoid changing the power state | |
897 | * of the parent bridge by setting the skip_bus_pm flag for it. | |
898 | */ | |
899 | if (pci_dev->bus->self) | |
900 | pci_dev->bus->self->skip_bus_pm = true; | |
901 | } | |
902 | ||
471a739a | 903 | if (pci_dev->skip_bus_pm && pm_suspend_no_platform()) { |
6941a0c2 | 904 | pci_dbg(pci_dev, "PCI PM: Skipped\n"); |
3e26c5fe RW |
905 | goto Fixup; |
906 | } | |
907 | ||
46939f8b RW |
908 | pci_pm_set_unknown_state(pci_dev); |
909 | ||
dbf0e4c7 AS |
910 | /* |
911 | * Some BIOSes from ASUS have a bug: If a USB EHCI host controller's | |
912 | * PCI COMMAND register isn't 0, the BIOS assumes that the controller | |
913 | * hasn't been quiesced and tries to turn it off. If the controller | |
914 | * is already in D3, this can hang or cause memory corruption. | |
915 | * | |
916 | * Since the value of the COMMAND register doesn't matter once the | |
917 | * device has been suspended, we can safely set it to 0 here. | |
918 | */ | |
919 | if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) | |
920 | pci_write_config_word(pci_dev, PCI_COMMAND, 0); | |
921 | ||
7d2a01b8 AN |
922 | Fixup: |
923 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); | |
924 | ||
bd755d77 RW |
925 | /* |
926 | * If the target system sleep state is suspend-to-idle, it is sufficient | |
927 | * to check whether or not the device's wakeup settings are good for | |
928 | * runtime PM. Otherwise, the pm_resume_via_firmware() check will cause | |
929 | * pci_pm_complete() to take care of fixing up the device's state | |
930 | * anyway, if need be. | |
931 | */ | |
0fe8a1be RW |
932 | if (device_can_wakeup(dev) && !device_may_wakeup(dev)) |
933 | dev->power.may_skip_resume = false; | |
bd755d77 | 934 | |
46939f8b | 935 | return 0; |
c8958177 | 936 | } |
1da177e4 | 937 | |
f6dc1e5e | 938 | static int pci_pm_resume_noirq(struct device *dev) |
bbb44d9f RW |
939 | { |
940 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
6da2f2cc | 941 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
ad9001f2 MW |
942 | pci_power_t prev_state = pci_dev->current_state; |
943 | bool skip_bus_pm = pci_dev->skip_bus_pm; | |
bbb44d9f | 944 | |
76c70cb5 | 945 | if (dev_pm_skip_resume(dev)) |
bd755d77 RW |
946 | return 0; |
947 | ||
3e26c5fe RW |
948 | /* |
949 | * In the suspend-to-idle case, devices left in D0 during suspend will | |
950 | * stay in D0, so it is not necessary to restore or update their | |
471a739a RW |
951 | * configuration here and attempting to put them into D0 again is |
952 | * pointless, so avoid doing that. | |
3e26c5fe | 953 | */ |
ad9001f2 | 954 | if (!(skip_bus_pm && pm_suspend_no_platform())) |
3e26c5fe RW |
955 | pci_pm_default_resume_early(pci_dev); |
956 | ||
957 | pci_fixup_device(pci_fixup_resume_early, pci_dev); | |
ec6a75ef | 958 | pcie_pme_root_status_cleanup(pci_dev); |
aa8c6c93 | 959 | |
ad9001f2 | 960 | if (!skip_bus_pm && prev_state == PCI_D3cold) |
730643d3 | 961 | pci_pm_bridge_power_up_actions(pci_dev); |
ad9001f2 | 962 | |
ad8cfa1d | 963 | if (pci_has_legacy_pm_support(pci_dev)) |
89cdbc35 | 964 | return 0; |
bb808945 | 965 | |
6da2f2cc BH |
966 | if (pm && pm->resume_noirq) |
967 | return pm->resume_noirq(dev); | |
bbb44d9f | 968 | |
6da2f2cc | 969 | return 0; |
bbb44d9f RW |
970 | } |
971 | ||
6e176bf8 RW |
972 | static int pci_pm_resume_early(struct device *dev) |
973 | { | |
76c70cb5 | 974 | if (dev_pm_skip_resume(dev)) |
6e176bf8 RW |
975 | return 0; |
976 | ||
977 | return pm_generic_resume_early(dev); | |
978 | } | |
979 | ||
f6dc1e5e | 980 | static int pci_pm_resume(struct device *dev) |
bbb44d9f | 981 | { |
355a72d7 | 982 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 983 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 984 | |
418e4da3 RW |
985 | /* |
986 | * This is necessary for the suspend error path in which resume is | |
987 | * called without restoring the standard config registers of the device. | |
988 | */ | |
989 | if (pci_dev->state_saved) | |
990 | pci_restore_standard_config(pci_dev); | |
991 | ||
c01163db BH |
992 | pci_resume_ptm(pci_dev); |
993 | ||
ad8cfa1d | 994 | if (pci_has_legacy_pm_support(pci_dev)) |
f6dc1e5e | 995 | return pci_legacy_resume(dev); |
bb808945 | 996 | |
5294e256 | 997 | pci_pm_default_resume(pci_dev); |
73410429 | 998 | |
5294e256 RW |
999 | if (pm) { |
1000 | if (pm->resume) | |
6da2f2cc | 1001 | return pm->resume(dev); |
5294e256 RW |
1002 | } else { |
1003 | pci_pm_reenable_device(pci_dev); | |
1004 | } | |
bbb44d9f | 1005 | |
6da2f2cc | 1006 | return 0; |
bbb44d9f RW |
1007 | } |
1008 | ||
1009 | #else /* !CONFIG_SUSPEND */ | |
1010 | ||
1011 | #define pci_pm_suspend NULL | |
c4b65157 | 1012 | #define pci_pm_suspend_late NULL |
bbb44d9f RW |
1013 | #define pci_pm_suspend_noirq NULL |
1014 | #define pci_pm_resume NULL | |
6e176bf8 | 1015 | #define pci_pm_resume_early NULL |
bbb44d9f RW |
1016 | #define pci_pm_resume_noirq NULL |
1017 | ||
1018 | #endif /* !CONFIG_SUSPEND */ | |
1019 | ||
1f112cee | 1020 | #ifdef CONFIG_HIBERNATE_CALLBACKS |
bbb44d9f RW |
1021 | |
1022 | static int pci_pm_freeze(struct device *dev) | |
1023 | { | |
1024 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
8150f32b | 1025 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1026 | |
ad8cfa1d RW |
1027 | if (pci_has_legacy_pm_support(pci_dev)) |
1028 | return pci_legacy_suspend(dev, PMSG_FREEZE); | |
bb808945 | 1029 | |
5294e256 RW |
1030 | if (!pm) { |
1031 | pci_pm_default_suspend(pci_dev); | |
1032 | return 0; | |
bbb44d9f RW |
1033 | } |
1034 | ||
7cd0602d | 1035 | /* |
501debd4 RW |
1036 | * Resume all runtime-suspended devices before creating a snapshot |
1037 | * image of system memory, because the restore kernel generally cannot | |
1038 | * be expected to always handle them consistently and they need to be | |
1039 | * put into the runtime-active metastate during system resume anyway, | |
1040 | * so it is better to ensure that the state saved in the image will be | |
1041 | * always consistent with that. | |
7cd0602d | 1042 | */ |
501debd4 RW |
1043 | pm_runtime_resume(dev); |
1044 | pci_dev->state_saved = false; | |
7cd0602d | 1045 | |
5294e256 RW |
1046 | if (pm->freeze) { |
1047 | int error; | |
1048 | ||
1049 | error = pm->freeze(dev); | |
a759de69 | 1050 | suspend_report_result(dev, pm->freeze, error); |
5294e256 RW |
1051 | if (error) |
1052 | return error; | |
1053 | } | |
1054 | ||
5294e256 | 1055 | return 0; |
bbb44d9f RW |
1056 | } |
1057 | ||
1058 | static int pci_pm_freeze_noirq(struct device *dev) | |
1059 | { | |
355a72d7 | 1060 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 1061 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1062 | |
bb808945 | 1063 | if (pci_has_legacy_pm_support(pci_dev)) |
5984de0b | 1064 | return pci_legacy_suspend_late(dev); |
bb808945 | 1065 | |
6da2f2cc | 1066 | if (pm && pm->freeze_noirq) { |
46939f8b RW |
1067 | int error; |
1068 | ||
6da2f2cc | 1069 | error = pm->freeze_noirq(dev); |
a759de69 | 1070 | suspend_report_result(dev, pm->freeze_noirq, error); |
46939f8b RW |
1071 | if (error) |
1072 | return error; | |
bbb44d9f RW |
1073 | } |
1074 | ||
46939f8b RW |
1075 | if (!pci_dev->state_saved) |
1076 | pci_save_state(pci_dev); | |
d67e37d7 | 1077 | |
46939f8b RW |
1078 | pci_pm_set_unknown_state(pci_dev); |
1079 | ||
1080 | return 0; | |
bbb44d9f RW |
1081 | } |
1082 | ||
f6dc1e5e | 1083 | static int pci_pm_thaw_noirq(struct device *dev) |
bbb44d9f | 1084 | { |
355a72d7 | 1085 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 1086 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
699c1985 | 1087 | |
5839ee73 | 1088 | /* |
89cdbc35 BH |
1089 | * The pm->thaw_noirq() callback assumes the device has been |
1090 | * returned to D0 and its config state has been restored. | |
f2c33cca DC |
1091 | * |
1092 | * In addition, pci_restore_state() restores MSI-X state in MMIO | |
1093 | * space, which requires the device to be in D0, so return it to D0 | |
1094 | * in case the driver's "freeze" callbacks put it into a low-power | |
1095 | * state. | |
5839ee73 | 1096 | */ |
0f40ac35 | 1097 | pci_pm_power_up_and_verify_state(pci_dev); |
e60514bd | 1098 | pci_restore_state(pci_dev); |
d67e37d7 | 1099 | |
f2c33cca | 1100 | if (pci_has_legacy_pm_support(pci_dev)) |
89cdbc35 | 1101 | return 0; |
f2c33cca | 1102 | |
6da2f2cc BH |
1103 | if (pm && pm->thaw_noirq) |
1104 | return pm->thaw_noirq(dev); | |
bbb44d9f | 1105 | |
6da2f2cc | 1106 | return 0; |
bbb44d9f RW |
1107 | } |
1108 | ||
f6dc1e5e | 1109 | static int pci_pm_thaw(struct device *dev) |
bbb44d9f | 1110 | { |
355a72d7 | 1111 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 1112 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f RW |
1113 | int error = 0; |
1114 | ||
ad8cfa1d | 1115 | if (pci_has_legacy_pm_support(pci_dev)) |
f6dc1e5e | 1116 | return pci_legacy_resume(dev); |
bb808945 | 1117 | |
5294e256 RW |
1118 | if (pm) { |
1119 | if (pm->thaw) | |
1120 | error = pm->thaw(dev); | |
1121 | } else { | |
1122 | pci_pm_reenable_device(pci_dev); | |
1123 | } | |
bbb44d9f | 1124 | |
4b77b0a2 RW |
1125 | pci_dev->state_saved = false; |
1126 | ||
bbb44d9f RW |
1127 | return error; |
1128 | } | |
1129 | ||
1130 | static int pci_pm_poweroff(struct device *dev) | |
1131 | { | |
355a72d7 | 1132 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 1133 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1134 | |
ad8cfa1d RW |
1135 | if (pci_has_legacy_pm_support(pci_dev)) |
1136 | return pci_legacy_suspend(dev, PMSG_HIBERNATE); | |
bb808945 | 1137 | |
5294e256 RW |
1138 | if (!pm) { |
1139 | pci_pm_default_suspend(pci_dev); | |
c4b65157 | 1140 | return 0; |
5294e256 RW |
1141 | } |
1142 | ||
7cd0602d | 1143 | /* The reason to do that is the same as in pci_pm_suspend(). */ |
c4b65157 | 1144 | if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || |
0c7376ad | 1145 | pci_dev_need_resume(pci_dev)) { |
c4b65157 | 1146 | pm_runtime_resume(dev); |
0c7376ad RW |
1147 | pci_dev->state_saved = false; |
1148 | } else { | |
1149 | pci_dev_adjust_pme(pci_dev); | |
1150 | } | |
7cd0602d | 1151 | |
5294e256 | 1152 | if (pm->poweroff) { |
46939f8b RW |
1153 | int error; |
1154 | ||
ddb7c9d2 | 1155 | error = pm->poweroff(dev); |
a759de69 | 1156 | suspend_report_result(dev, pm->poweroff, error); |
46939f8b RW |
1157 | if (error) |
1158 | return error; | |
bbb44d9f RW |
1159 | } |
1160 | ||
46939f8b | 1161 | return 0; |
bbb44d9f | 1162 | } |
c9b9972b | 1163 | |
c4b65157 RW |
1164 | static int pci_pm_poweroff_late(struct device *dev) |
1165 | { | |
fa2bfead | 1166 | if (dev_pm_skip_suspend(dev)) |
c4b65157 | 1167 | return 0; |
699c1985 | 1168 | |
c4b65157 RW |
1169 | pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); |
1170 | ||
1171 | return pm_generic_poweroff_late(dev); | |
bbb44d9f RW |
1172 | } |
1173 | ||
1174 | static int pci_pm_poweroff_noirq(struct device *dev) | |
1175 | { | |
46939f8b | 1176 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 1177 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1178 | |
fa2bfead | 1179 | if (dev_pm_skip_suspend(dev)) |
c4b65157 RW |
1180 | return 0; |
1181 | ||
6da2f2cc | 1182 | if (pci_has_legacy_pm_support(pci_dev)) |
5984de0b | 1183 | return pci_legacy_suspend_late(dev); |
bb808945 | 1184 | |
6da2f2cc | 1185 | if (!pm) { |
7d2a01b8 | 1186 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); |
46939f8b | 1187 | return 0; |
7d2a01b8 | 1188 | } |
46939f8b | 1189 | |
6da2f2cc | 1190 | if (pm->poweroff_noirq) { |
46939f8b RW |
1191 | int error; |
1192 | ||
6da2f2cc | 1193 | error = pm->poweroff_noirq(dev); |
a759de69 | 1194 | suspend_report_result(dev, pm->poweroff_noirq, error); |
46939f8b RW |
1195 | if (error) |
1196 | return error; | |
bbb44d9f RW |
1197 | } |
1198 | ||
326c1cda | 1199 | if (!pci_dev->state_saved && !pci_has_subordinate(pci_dev)) |
46939f8b RW |
1200 | pci_prepare_to_sleep(pci_dev); |
1201 | ||
0b68c8e2 RW |
1202 | /* |
1203 | * The reason for doing this here is the same as for the analogous code | |
1204 | * in pci_pm_suspend_noirq(). | |
1205 | */ | |
1206 | if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) | |
1207 | pci_write_config_word(pci_dev, PCI_COMMAND, 0); | |
1208 | ||
7d2a01b8 AN |
1209 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); |
1210 | ||
46939f8b | 1211 | return 0; |
bbb44d9f RW |
1212 | } |
1213 | ||
f6dc1e5e | 1214 | static int pci_pm_restore_noirq(struct device *dev) |
bbb44d9f RW |
1215 | { |
1216 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
6da2f2cc | 1217 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
699c1985 | 1218 | |
6cbf8214 | 1219 | pci_pm_default_resume_early(pci_dev); |
3e26c5fe | 1220 | pci_fixup_device(pci_fixup_resume_early, pci_dev); |
aa8c6c93 | 1221 | |
ad8cfa1d | 1222 | if (pci_has_legacy_pm_support(pci_dev)) |
89cdbc35 | 1223 | return 0; |
bb808945 | 1224 | |
6da2f2cc BH |
1225 | if (pm && pm->restore_noirq) |
1226 | return pm->restore_noirq(dev); | |
bbb44d9f | 1227 | |
6da2f2cc | 1228 | return 0; |
bbb44d9f RW |
1229 | } |
1230 | ||
f6dc1e5e | 1231 | static int pci_pm_restore(struct device *dev) |
bbb44d9f RW |
1232 | { |
1233 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
8150f32b | 1234 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1235 | |
418e4da3 RW |
1236 | /* |
1237 | * This is necessary for the hibernation error path in which restore is | |
1238 | * called without restoring the standard config registers of the device. | |
1239 | */ | |
1240 | if (pci_dev->state_saved) | |
1241 | pci_restore_standard_config(pci_dev); | |
1242 | ||
ad8cfa1d | 1243 | if (pci_has_legacy_pm_support(pci_dev)) |
f6dc1e5e | 1244 | return pci_legacy_resume(dev); |
bb808945 | 1245 | |
5294e256 | 1246 | pci_pm_default_resume(pci_dev); |
73410429 | 1247 | |
5294e256 RW |
1248 | if (pm) { |
1249 | if (pm->restore) | |
6da2f2cc | 1250 | return pm->restore(dev); |
5294e256 RW |
1251 | } else { |
1252 | pci_pm_reenable_device(pci_dev); | |
1253 | } | |
bbb44d9f | 1254 | |
6da2f2cc | 1255 | return 0; |
c8958177 | 1256 | } |
1da177e4 | 1257 | |
1f112cee | 1258 | #else /* !CONFIG_HIBERNATE_CALLBACKS */ |
bbb44d9f RW |
1259 | |
1260 | #define pci_pm_freeze NULL | |
1261 | #define pci_pm_freeze_noirq NULL | |
1262 | #define pci_pm_thaw NULL | |
1263 | #define pci_pm_thaw_noirq NULL | |
1264 | #define pci_pm_poweroff NULL | |
c4b65157 | 1265 | #define pci_pm_poweroff_late NULL |
bbb44d9f RW |
1266 | #define pci_pm_poweroff_noirq NULL |
1267 | #define pci_pm_restore NULL | |
1268 | #define pci_pm_restore_noirq NULL | |
1269 | ||
1f112cee | 1270 | #endif /* !CONFIG_HIBERNATE_CALLBACKS */ |
bbb44d9f | 1271 | |
fbb988be | 1272 | #ifdef CONFIG_PM |
6cbf8214 RW |
1273 | |
1274 | static int pci_pm_runtime_suspend(struct device *dev) | |
1275 | { | |
1276 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
1277 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; | |
1278 | pci_power_t prev = pci_dev->current_state; | |
1279 | int error; | |
1280 | ||
c01163db BH |
1281 | pci_suspend_ptm(pci_dev); |
1282 | ||
967577b0 | 1283 | /* |
e0217c5b BH |
1284 | * If pci_dev->driver is not set (unbound), we leave the device in D0, |
1285 | * but it may go to D3cold when the bridge above it runtime suspends. | |
1286 | * Save its config space in case that happens. | |
967577b0 | 1287 | */ |
e0217c5b | 1288 | if (!pci_dev->driver) { |
5775b843 | 1289 | pci_save_state(pci_dev); |
967577b0 | 1290 | return 0; |
5775b843 | 1291 | } |
967577b0 | 1292 | |
82fee4d6 | 1293 | pci_dev->state_saved = false; |
c5eb1190 JN |
1294 | if (pm && pm->runtime_suspend) { |
1295 | error = pm->runtime_suspend(dev); | |
06bf403d ID |
1296 | /* |
1297 | * -EBUSY and -EAGAIN is used to request the runtime PM core | |
1298 | * to schedule a new suspend, so log the event only with debug | |
1299 | * log level. | |
1300 | */ | |
c5eb1190 | 1301 | if (error == -EBUSY || error == -EAGAIN) { |
6941a0c2 | 1302 | pci_dbg(pci_dev, "can't suspend now (%ps returned %d)\n", |
06bf403d | 1303 | pm->runtime_suspend, error); |
c5eb1190 JN |
1304 | return error; |
1305 | } else if (error) { | |
6941a0c2 | 1306 | pci_err(pci_dev, "can't suspend (%ps returned %d)\n", |
06bf403d | 1307 | pm->runtime_suspend, error); |
c5eb1190 JN |
1308 | return error; |
1309 | } | |
06bf403d | 1310 | } |
6cbf8214 RW |
1311 | |
1312 | pci_fixup_device(pci_fixup_suspend, pci_dev); | |
1313 | ||
c5eb1190 JN |
1314 | if (pm && pm->runtime_suspend |
1315 | && !pci_dev->state_saved && pci_dev->current_state != PCI_D0 | |
6cbf8214 | 1316 | && pci_dev->current_state != PCI_UNKNOWN) { |
12bcae44 BH |
1317 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
1318 | "PCI PM: State of device not saved by %pS\n", | |
1319 | pm->runtime_suspend); | |
6cbf8214 RW |
1320 | return 0; |
1321 | } | |
1322 | ||
42eca230 | 1323 | if (!pci_dev->state_saved) { |
6cbf8214 | 1324 | pci_save_state(pci_dev); |
42eca230 DA |
1325 | pci_finish_runtime_suspend(pci_dev); |
1326 | } | |
6cbf8214 RW |
1327 | |
1328 | return 0; | |
1329 | } | |
1330 | ||
1331 | static int pci_pm_runtime_resume(struct device *dev) | |
1332 | { | |
1333 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
1334 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; | |
ad9001f2 | 1335 | pci_power_t prev_state = pci_dev->current_state; |
6da2f2cc | 1336 | int error = 0; |
6cbf8214 | 1337 | |
967577b0 | 1338 | /* |
5775b843 RW |
1339 | * Restoring config space is necessary even if the device is not bound |
1340 | * to a driver because although we left it in D0, it may have gone to | |
1341 | * D3cold when the bridge above it runtime suspended. | |
967577b0 | 1342 | */ |
9a605831 | 1343 | pci_pm_default_resume_early(pci_dev); |
c01163db | 1344 | pci_resume_ptm(pci_dev); |
5775b843 | 1345 | |
e0217c5b | 1346 | if (!pci_dev->driver) |
967577b0 HY |
1347 | return 0; |
1348 | ||
db288c9c | 1349 | pci_fixup_device(pci_fixup_resume_early, pci_dev); |
f7b32a86 | 1350 | pci_pm_default_resume(pci_dev); |
6cbf8214 | 1351 | |
ad9001f2 | 1352 | if (prev_state == PCI_D3cold) |
730643d3 | 1353 | pci_pm_bridge_power_up_actions(pci_dev); |
ad9001f2 | 1354 | |
c5eb1190 | 1355 | if (pm && pm->runtime_resume) |
6da2f2cc | 1356 | error = pm->runtime_resume(dev); |
448bd857 | 1357 | |
6da2f2cc | 1358 | return error; |
6cbf8214 RW |
1359 | } |
1360 | ||
1361 | static int pci_pm_runtime_idle(struct device *dev) | |
1362 | { | |
e0217c5b | 1363 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6cbf8214 RW |
1364 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
1365 | ||
967577b0 | 1366 | /* |
e0217c5b BH |
1367 | * If pci_dev->driver is not set (unbound), the device should |
1368 | * always remain in D0 regardless of the runtime PM status | |
967577b0 | 1369 | */ |
e0217c5b | 1370 | if (!pci_dev->driver) |
45f0a85c | 1371 | return 0; |
967577b0 | 1372 | |
6cbf8214 RW |
1373 | if (!pm) |
1374 | return -ENOSYS; | |
1375 | ||
45f0a85c | 1376 | if (pm->runtime_idle) |
6da2f2cc | 1377 | return pm->runtime_idle(dev); |
6cbf8214 | 1378 | |
6da2f2cc | 1379 | return 0; |
6cbf8214 RW |
1380 | } |
1381 | ||
f91da04d | 1382 | static const struct dev_pm_ops pci_dev_pm_ops = { |
adf09493 | 1383 | .prepare = pci_pm_prepare, |
2cef548a | 1384 | .complete = pci_pm_complete, |
adf09493 | 1385 | .suspend = pci_pm_suspend, |
c4b65157 | 1386 | .suspend_late = pci_pm_suspend_late, |
adf09493 | 1387 | .resume = pci_pm_resume, |
6e176bf8 | 1388 | .resume_early = pci_pm_resume_early, |
adf09493 RW |
1389 | .freeze = pci_pm_freeze, |
1390 | .thaw = pci_pm_thaw, | |
1391 | .poweroff = pci_pm_poweroff, | |
c4b65157 | 1392 | .poweroff_late = pci_pm_poweroff_late, |
adf09493 | 1393 | .restore = pci_pm_restore, |
bbb44d9f RW |
1394 | .suspend_noirq = pci_pm_suspend_noirq, |
1395 | .resume_noirq = pci_pm_resume_noirq, | |
1396 | .freeze_noirq = pci_pm_freeze_noirq, | |
1397 | .thaw_noirq = pci_pm_thaw_noirq, | |
1398 | .poweroff_noirq = pci_pm_poweroff_noirq, | |
1399 | .restore_noirq = pci_pm_restore_noirq, | |
6cbf8214 RW |
1400 | .runtime_suspend = pci_pm_runtime_suspend, |
1401 | .runtime_resume = pci_pm_runtime_resume, | |
1402 | .runtime_idle = pci_pm_runtime_idle, | |
bbb44d9f RW |
1403 | }; |
1404 | ||
adf09493 | 1405 | #define PCI_PM_OPS_PTR (&pci_dev_pm_ops) |
bbb44d9f | 1406 | |
fbb988be RW |
1407 | #else /* !CONFIG_PM */ |
1408 | ||
1409 | #define pci_pm_runtime_suspend NULL | |
1410 | #define pci_pm_runtime_resume NULL | |
1411 | #define pci_pm_runtime_idle NULL | |
bbb44d9f RW |
1412 | |
1413 | #define PCI_PM_OPS_PTR NULL | |
1414 | ||
fbb988be | 1415 | #endif /* !CONFIG_PM */ |
bbb44d9f | 1416 | |
1da177e4 | 1417 | /** |
863b18f4 | 1418 | * __pci_register_driver - register a new pci driver |
1da177e4 | 1419 | * @drv: the driver structure to register |
863b18f4 | 1420 | * @owner: owner module of drv |
f95d882d | 1421 | * @mod_name: module name string |
f7625980 | 1422 | * |
1da177e4 | 1423 | * Adds the driver structure to the list of registered drivers. |
f7625980 BH |
1424 | * Returns a negative value on error, otherwise 0. |
1425 | * If no error occurred, the driver remains registered even if | |
1da177e4 LT |
1426 | * no device was claimed during registration. |
1427 | */ | |
725522b5 GKH |
1428 | int __pci_register_driver(struct pci_driver *drv, struct module *owner, |
1429 | const char *mod_name) | |
1da177e4 | 1430 | { |
1da177e4 LT |
1431 | /* initialize common driver fields */ |
1432 | drv->driver.name = drv->name; | |
1433 | drv->driver.bus = &pci_bus_type; | |
863b18f4 | 1434 | drv->driver.owner = owner; |
725522b5 | 1435 | drv->driver.mod_name = mod_name; |
92d50fc1 | 1436 | drv->driver.groups = drv->groups; |
ded13b9c | 1437 | drv->driver.dev_groups = drv->dev_groups; |
50b00755 | 1438 | |
75865858 GKH |
1439 | spin_lock_init(&drv->dynids.lock); |
1440 | INIT_LIST_HEAD(&drv->dynids.list); | |
1da177e4 LT |
1441 | |
1442 | /* register with core */ | |
bfb09a86 | 1443 | return driver_register(&drv->driver); |
1da177e4 | 1444 | } |
b7fe9434 | 1445 | EXPORT_SYMBOL(__pci_register_driver); |
1da177e4 LT |
1446 | |
1447 | /** | |
1448 | * pci_unregister_driver - unregister a pci driver | |
1449 | * @drv: the driver structure to unregister | |
f7625980 | 1450 | * |
1da177e4 LT |
1451 | * Deletes the driver structure from the list of registered PCI drivers, |
1452 | * gives it a chance to clean up by calling its remove() function for | |
1453 | * each device it was responsible for, and marks those devices as | |
1454 | * driverless. | |
1455 | */ | |
1456 | ||
3c78bc61 | 1457 | void pci_unregister_driver(struct pci_driver *drv) |
1da177e4 LT |
1458 | { |
1459 | driver_unregister(&drv->driver); | |
1460 | pci_free_dynids(drv); | |
1461 | } | |
b7fe9434 | 1462 | EXPORT_SYMBOL(pci_unregister_driver); |
1da177e4 LT |
1463 | |
1464 | static struct pci_driver pci_compat_driver = { | |
1465 | .name = "compat" | |
1466 | }; | |
1467 | ||
1468 | /** | |
1469 | * pci_dev_driver - get the pci_driver of a device | |
1470 | * @dev: the device to query | |
1471 | * | |
f7625980 | 1472 | * Returns the appropriate pci_driver structure or %NULL if there is no |
1da177e4 LT |
1473 | * registered driver for the device. |
1474 | */ | |
3c78bc61 | 1475 | struct pci_driver *pci_dev_driver(const struct pci_dev *dev) |
1da177e4 | 1476 | { |
1ec09529 BH |
1477 | int i; |
1478 | ||
e0217c5b BH |
1479 | if (dev->driver) |
1480 | return dev->driver; | |
1ec09529 BH |
1481 | |
1482 | for (i = 0; i <= PCI_ROM_RESOURCE; i++) | |
1483 | if (dev->resource[i].flags & IORESOURCE_BUSY) | |
1484 | return &pci_compat_driver; | |
1485 | ||
1da177e4 LT |
1486 | return NULL; |
1487 | } | |
b7fe9434 | 1488 | EXPORT_SYMBOL(pci_dev_driver); |
1da177e4 LT |
1489 | |
1490 | /** | |
1491 | * pci_bus_match - Tell if a PCI device structure has a matching PCI device id structure | |
1da177e4 | 1492 | * @dev: the PCI device structure to match against |
8f7020d3 | 1493 | * @drv: the device driver to search for matching PCI device id structures |
f7625980 | 1494 | * |
1da177e4 | 1495 | * Used by a driver to check whether a PCI device present in the |
8f7020d3 | 1496 | * system is in its list of supported devices. Returns the matching |
1da177e4 LT |
1497 | * pci_device_id structure or %NULL if there is no match. |
1498 | */ | |
75865858 | 1499 | static int pci_bus_match(struct device *dev, struct device_driver *drv) |
1da177e4 | 1500 | { |
75865858 | 1501 | struct pci_dev *pci_dev = to_pci_dev(dev); |
58d9a38f | 1502 | struct pci_driver *pci_drv; |
1da177e4 LT |
1503 | const struct pci_device_id *found_id; |
1504 | ||
58d9a38f YL |
1505 | if (!pci_dev->match_driver) |
1506 | return 0; | |
1507 | ||
1508 | pci_drv = to_pci_driver(drv); | |
75865858 | 1509 | found_id = pci_match_device(pci_drv, pci_dev); |
1da177e4 LT |
1510 | if (found_id) |
1511 | return 1; | |
1512 | ||
75865858 | 1513 | return 0; |
1da177e4 LT |
1514 | } |
1515 | ||
1516 | /** | |
1517 | * pci_dev_get - increments the reference count of the pci device structure | |
1518 | * @dev: the device being referenced | |
1519 | * | |
1520 | * Each live reference to a device should be refcounted. | |
1521 | * | |
1522 | * Drivers for PCI devices should normally record such references in | |
1523 | * their probe() methods, when they bind to a device, and release | |
1524 | * them by calling pci_dev_put(), in their disconnect() methods. | |
1525 | * | |
1526 | * A pointer to the device with the incremented reference counter is returned. | |
1527 | */ | |
1528 | struct pci_dev *pci_dev_get(struct pci_dev *dev) | |
1529 | { | |
1530 | if (dev) | |
1531 | get_device(&dev->dev); | |
1532 | return dev; | |
1533 | } | |
b7fe9434 | 1534 | EXPORT_SYMBOL(pci_dev_get); |
1da177e4 LT |
1535 | |
1536 | /** | |
1537 | * pci_dev_put - release a use of the pci device structure | |
1538 | * @dev: device that's been disconnected | |
1539 | * | |
1540 | * Must be called when a user of a device is finished with it. When the last | |
1541 | * user of the device calls this function, the memory of the device is freed. | |
1542 | */ | |
1543 | void pci_dev_put(struct pci_dev *dev) | |
1544 | { | |
1545 | if (dev) | |
1546 | put_device(&dev->dev); | |
1547 | } | |
b7fe9434 | 1548 | EXPORT_SYMBOL(pci_dev_put); |
1da177e4 | 1549 | |
2a81ada3 | 1550 | static int pci_uevent(const struct device *dev, struct kobj_uevent_env *env) |
8ccc9aa1 | 1551 | { |
2a81ada3 | 1552 | const struct pci_dev *pdev; |
8ccc9aa1 BP |
1553 | |
1554 | if (!dev) | |
1555 | return -ENODEV; | |
1556 | ||
1557 | pdev = to_pci_dev(dev); | |
8ccc9aa1 BP |
1558 | |
1559 | if (add_uevent_var(env, "PCI_CLASS=%04X", pdev->class)) | |
1560 | return -ENOMEM; | |
1561 | ||
1562 | if (add_uevent_var(env, "PCI_ID=%04X:%04X", pdev->vendor, pdev->device)) | |
1563 | return -ENOMEM; | |
1564 | ||
1565 | if (add_uevent_var(env, "PCI_SUBSYS_ID=%04X:%04X", pdev->subsystem_vendor, | |
1566 | pdev->subsystem_device)) | |
1567 | return -ENOMEM; | |
1568 | ||
1569 | if (add_uevent_var(env, "PCI_SLOT_NAME=%s", pci_name(pdev))) | |
1570 | return -ENOMEM; | |
1571 | ||
145b3fe5 | 1572 | if (add_uevent_var(env, "MODALIAS=pci:v%08Xd%08Xsv%08Xsd%08Xbc%02Xsc%02Xi%02X", |
8ccc9aa1 BP |
1573 | pdev->vendor, pdev->device, |
1574 | pdev->subsystem_vendor, pdev->subsystem_device, | |
1575 | (u8)(pdev->class >> 16), (u8)(pdev->class >> 8), | |
1576 | (u8)(pdev->class))) | |
1577 | return -ENOMEM; | |
efdd4070 | 1578 | |
8ccc9aa1 BP |
1579 | return 0; |
1580 | } | |
1581 | ||
f9a6c8ad | 1582 | #if defined(CONFIG_PCIEAER) || defined(CONFIG_EEH) |
3ecac020 ME |
1583 | /** |
1584 | * pci_uevent_ers - emit a uevent during recovery path of PCI device | |
1585 | * @pdev: PCI device undergoing error recovery | |
1586 | * @err_type: type of error event | |
1587 | */ | |
1588 | void pci_uevent_ers(struct pci_dev *pdev, enum pci_ers_result err_type) | |
1589 | { | |
1590 | int idx = 0; | |
1591 | char *envp[3]; | |
1592 | ||
1593 | switch (err_type) { | |
1594 | case PCI_ERS_RESULT_NONE: | |
1595 | case PCI_ERS_RESULT_CAN_RECOVER: | |
1596 | envp[idx++] = "ERROR_EVENT=BEGIN_RECOVERY"; | |
1597 | envp[idx++] = "DEVICE_ONLINE=0"; | |
1598 | break; | |
1599 | case PCI_ERS_RESULT_RECOVERED: | |
1600 | envp[idx++] = "ERROR_EVENT=SUCCESSFUL_RECOVERY"; | |
1601 | envp[idx++] = "DEVICE_ONLINE=1"; | |
1602 | break; | |
1603 | case PCI_ERS_RESULT_DISCONNECT: | |
1604 | envp[idx++] = "ERROR_EVENT=FAILED_RECOVERY"; | |
1605 | envp[idx++] = "DEVICE_ONLINE=0"; | |
1606 | break; | |
1607 | default: | |
1608 | break; | |
1609 | } | |
1610 | ||
1611 | if (idx > 0) { | |
1612 | envp[idx++] = NULL; | |
1613 | kobject_uevent_env(&pdev->dev.kobj, KOBJ_CHANGE, envp); | |
1614 | } | |
1615 | } | |
1616 | #endif | |
1617 | ||
02e0bea6 PS |
1618 | static int pci_bus_num_vf(struct device *dev) |
1619 | { | |
1620 | return pci_num_vf(to_pci_dev(dev)); | |
1621 | } | |
1622 | ||
07397df2 NG |
1623 | /** |
1624 | * pci_dma_configure - Setup DMA configuration | |
1625 | * @dev: ptr to dev structure | |
1626 | * | |
1627 | * Function to update PCI devices's DMA configuration using the same | |
1628 | * info from the OF node or ACPI node of host bridge's parent (if any). | |
1629 | */ | |
1630 | static int pci_dma_configure(struct device *dev) | |
1631 | { | |
512881ea | 1632 | struct pci_driver *driver = to_pci_driver(dev->driver); |
07397df2 NG |
1633 | struct device *bridge; |
1634 | int ret = 0; | |
1635 | ||
1636 | bridge = pci_get_host_bridge_device(to_pci_dev(dev)); | |
1637 | ||
1638 | if (IS_ENABLED(CONFIG_OF) && bridge->parent && | |
1639 | bridge->parent->of_node) { | |
3d6ce86e | 1640 | ret = of_dma_configure(dev, bridge->parent->of_node, true); |
07397df2 NG |
1641 | } else if (has_acpi_companion(bridge)) { |
1642 | struct acpi_device *adev = to_acpi_device_node(bridge->fwnode); | |
07397df2 | 1643 | |
e5361ca2 | 1644 | ret = acpi_dma_configure(dev, acpi_get_dma_attr(adev)); |
07397df2 NG |
1645 | } |
1646 | ||
1647 | pci_put_host_bridge_device(bridge); | |
512881ea LB |
1648 | |
1649 | if (!ret && !driver->driver_managed_dma) { | |
1650 | ret = iommu_device_use_default_domain(dev); | |
1651 | if (ret) | |
1652 | arch_teardown_dma_ops(dev); | |
1653 | } | |
1654 | ||
07397df2 NG |
1655 | return ret; |
1656 | } | |
1657 | ||
512881ea LB |
1658 | static void pci_dma_cleanup(struct device *dev) |
1659 | { | |
1660 | struct pci_driver *driver = to_pci_driver(dev->driver); | |
1661 | ||
1662 | if (!driver->driver_managed_dma) | |
1663 | iommu_device_unuse_default_domain(dev); | |
1664 | } | |
1665 | ||
1da177e4 LT |
1666 | struct bus_type pci_bus_type = { |
1667 | .name = "pci", | |
1668 | .match = pci_bus_match, | |
312c004d | 1669 | .uevent = pci_uevent, |
b15d686a RK |
1670 | .probe = pci_device_probe, |
1671 | .remove = pci_device_remove, | |
cbd69dbb | 1672 | .shutdown = pci_device_shutdown, |
5136b2da | 1673 | .dev_groups = pci_dev_groups, |
0f49ba55 | 1674 | .bus_groups = pci_bus_groups, |
2229c1fb | 1675 | .drv_groups = pci_drv_groups, |
bbb44d9f | 1676 | .pm = PCI_PM_OPS_PTR, |
02e0bea6 | 1677 | .num_vf = pci_bus_num_vf, |
07397df2 | 1678 | .dma_configure = pci_dma_configure, |
512881ea | 1679 | .dma_cleanup = pci_dma_cleanup, |
1da177e4 | 1680 | }; |
b7fe9434 | 1681 | EXPORT_SYMBOL(pci_bus_type); |
1da177e4 | 1682 | |
c6c889d9 BH |
1683 | #ifdef CONFIG_PCIEPORTBUS |
1684 | static int pcie_port_bus_match(struct device *dev, struct device_driver *drv) | |
1685 | { | |
1686 | struct pcie_device *pciedev; | |
1687 | struct pcie_port_service_driver *driver; | |
1688 | ||
1689 | if (drv->bus != &pcie_port_bus_type || dev->bus != &pcie_port_bus_type) | |
1690 | return 0; | |
1691 | ||
1692 | pciedev = to_pcie_device(dev); | |
1693 | driver = to_service_driver(drv); | |
1694 | ||
1695 | if (driver->service != pciedev->service) | |
1696 | return 0; | |
1697 | ||
1698 | if (driver->port_type != PCIE_ANY_PORT && | |
1699 | driver->port_type != pci_pcie_type(pciedev->port)) | |
1700 | return 0; | |
1701 | ||
1702 | return 1; | |
1703 | } | |
1704 | ||
1705 | struct bus_type pcie_port_bus_type = { | |
1706 | .name = "pci_express", | |
1707 | .match = pcie_port_bus_match, | |
1708 | }; | |
c6c889d9 BH |
1709 | #endif |
1710 | ||
1da177e4 LT |
1711 | static int __init pci_driver_init(void) |
1712 | { | |
c6c889d9 BH |
1713 | int ret; |
1714 | ||
1715 | ret = bus_register(&pci_bus_type); | |
1716 | if (ret) | |
1717 | return ret; | |
1718 | ||
1719 | #ifdef CONFIG_PCIEPORTBUS | |
1720 | ret = bus_register(&pcie_port_bus_type); | |
1721 | if (ret) | |
1722 | return ret; | |
1723 | #endif | |
a8651194 | 1724 | dma_debug_add_bus(&pci_bus_type); |
c6c889d9 | 1725 | return 0; |
1da177e4 | 1726 | } |
1da177e4 | 1727 | postcore_initcall(pci_driver_init); |