Commit | Line | Data |
---|---|---|
8cfab3cf | 1 | // SPDX-License-Identifier: GPL-2.0 |
1da177e4 | 2 | /* |
2b937303 GKH |
3 | * (C) Copyright 2002-2004, 2007 Greg Kroah-Hartman <greg@kroah.com> |
4 | * (C) Copyright 2007 Novell Inc. | |
1da177e4 LT |
5 | */ |
6 | ||
7 | #include <linux/pci.h> | |
8 | #include <linux/module.h> | |
9 | #include <linux/init.h> | |
10 | #include <linux/device.h> | |
d42c6997 | 11 | #include <linux/mempolicy.h> |
4e57b681 TS |
12 | #include <linux/string.h> |
13 | #include <linux/slab.h> | |
8c65b4a6 | 14 | #include <linux/sched.h> |
69a18b18 | 15 | #include <linux/sched/isolation.h> |
873392ca | 16 | #include <linux/cpu.h> |
6cbf8214 | 17 | #include <linux/pm_runtime.h> |
eea3fc03 | 18 | #include <linux/suspend.h> |
4fc9bbf9 | 19 | #include <linux/kexec.h> |
07397df2 NG |
20 | #include <linux/of_device.h> |
21 | #include <linux/acpi.h> | |
a1fd09e8 | 22 | #include <linux/dma-map-ops.h> |
512881ea | 23 | #include <linux/iommu.h> |
1da177e4 | 24 | #include "pci.h" |
c6c889d9 | 25 | #include "pcie/portdrv.h" |
1da177e4 | 26 | |
75865858 GKH |
27 | struct pci_dynid { |
28 | struct list_head node; | |
29 | struct pci_device_id id; | |
30 | }; | |
1da177e4 | 31 | |
9dba910e TH |
32 | /** |
33 | * pci_add_dynid - add a new PCI device ID to this driver and re-probe devices | |
34 | * @drv: target pci driver | |
35 | * @vendor: PCI vendor ID | |
36 | * @device: PCI device ID | |
37 | * @subvendor: PCI subvendor ID | |
38 | * @subdevice: PCI subdevice ID | |
39 | * @class: PCI class | |
40 | * @class_mask: PCI class mask | |
41 | * @driver_data: private driver data | |
42 | * | |
43 | * Adds a new dynamic pci device ID to this driver and causes the | |
44 | * driver to probe for all devices again. @drv must have been | |
45 | * registered prior to calling this function. | |
46 | * | |
47 | * CONTEXT: | |
48 | * Does GFP_KERNEL allocation. | |
49 | * | |
50 | * RETURNS: | |
51 | * 0 on success, -errno on failure. | |
52 | */ | |
53 | int pci_add_dynid(struct pci_driver *drv, | |
54 | unsigned int vendor, unsigned int device, | |
55 | unsigned int subvendor, unsigned int subdevice, | |
56 | unsigned int class, unsigned int class_mask, | |
57 | unsigned long driver_data) | |
58 | { | |
59 | struct pci_dynid *dynid; | |
9dba910e TH |
60 | |
61 | dynid = kzalloc(sizeof(*dynid), GFP_KERNEL); | |
62 | if (!dynid) | |
63 | return -ENOMEM; | |
64 | ||
65 | dynid->id.vendor = vendor; | |
66 | dynid->id.device = device; | |
67 | dynid->id.subvendor = subvendor; | |
68 | dynid->id.subdevice = subdevice; | |
69 | dynid->id.class = class; | |
70 | dynid->id.class_mask = class_mask; | |
71 | dynid->id.driver_data = driver_data; | |
3d3c2ae1 | 72 | |
9dba910e TH |
73 | spin_lock(&drv->dynids.lock); |
74 | list_add_tail(&dynid->node, &drv->dynids.list); | |
75 | spin_unlock(&drv->dynids.lock); | |
76 | ||
3b7f1016 | 77 | return driver_attach(&drv->driver); |
9dba910e | 78 | } |
b7fe9434 | 79 | EXPORT_SYMBOL_GPL(pci_add_dynid); |
9dba910e TH |
80 | |
81 | static void pci_free_dynids(struct pci_driver *drv) | |
82 | { | |
83 | struct pci_dynid *dynid, *n; | |
3d3c2ae1 | 84 | |
9dba910e TH |
85 | spin_lock(&drv->dynids.lock); |
86 | list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { | |
87 | list_del(&dynid->node); | |
88 | kfree(dynid); | |
89 | } | |
90 | spin_unlock(&drv->dynids.lock); | |
91 | } | |
92 | ||
1f40704b ZD |
93 | /** |
94 | * pci_match_id - See if a PCI device matches a given pci_id table | |
95 | * @ids: array of PCI device ID structures to search in | |
96 | * @dev: the PCI device structure to match against. | |
97 | * | |
98 | * Used by a driver to check whether a PCI device is in its list of | |
99 | * supported devices. Returns the matching pci_device_id structure or | |
100 | * %NULL if there is no match. | |
101 | * | |
102 | * Deprecated; don't use this as it will not catch any dynamic IDs | |
103 | * that a driver might want to check for. | |
104 | */ | |
105 | const struct pci_device_id *pci_match_id(const struct pci_device_id *ids, | |
106 | struct pci_dev *dev) | |
107 | { | |
108 | if (ids) { | |
109 | while (ids->vendor || ids->subvendor || ids->class_mask) { | |
110 | if (pci_match_one_device(ids, dev)) | |
111 | return ids; | |
112 | ids++; | |
113 | } | |
114 | } | |
115 | return NULL; | |
116 | } | |
117 | EXPORT_SYMBOL(pci_match_id); | |
118 | ||
119 | static const struct pci_device_id pci_device_id_any = { | |
120 | .vendor = PCI_ANY_ID, | |
121 | .device = PCI_ANY_ID, | |
122 | .subvendor = PCI_ANY_ID, | |
123 | .subdevice = PCI_ANY_ID, | |
124 | }; | |
125 | ||
126 | /** | |
127 | * pci_match_device - See if a device matches a driver's list of IDs | |
128 | * @drv: the PCI driver to match against | |
129 | * @dev: the PCI device structure to match against | |
130 | * | |
131 | * Used by a driver to check whether a PCI device is in its list of | |
132 | * supported devices or in the dynids list, which may have been augmented | |
133 | * via the sysfs "new_id" file. Returns the matching pci_device_id | |
134 | * structure or %NULL if there is no match. | |
135 | */ | |
136 | static const struct pci_device_id *pci_match_device(struct pci_driver *drv, | |
137 | struct pci_dev *dev) | |
138 | { | |
139 | struct pci_dynid *dynid; | |
343b7258 | 140 | const struct pci_device_id *found_id = NULL, *ids; |
1f40704b ZD |
141 | |
142 | /* When driver_override is set, only bind to the matching driver */ | |
143 | if (dev->driver_override && strcmp(dev->driver_override, drv->name)) | |
144 | return NULL; | |
145 | ||
146 | /* Look at the dynamic ids first, before the static ones */ | |
147 | spin_lock(&drv->dynids.lock); | |
148 | list_for_each_entry(dynid, &drv->dynids.list, node) { | |
149 | if (pci_match_one_device(&dynid->id, dev)) { | |
150 | found_id = &dynid->id; | |
151 | break; | |
152 | } | |
153 | } | |
154 | spin_unlock(&drv->dynids.lock); | |
155 | ||
343b7258 MG |
156 | if (found_id) |
157 | return found_id; | |
158 | ||
159 | for (ids = drv->id_table; (found_id = pci_match_id(ids, dev)); | |
160 | ids = found_id + 1) { | |
161 | /* | |
162 | * The match table is split based on driver_override. | |
163 | * In case override_only was set, enforce driver_override | |
164 | * matching. | |
165 | */ | |
166 | if (found_id->override_only) { | |
167 | if (dev->driver_override) | |
168 | return found_id; | |
169 | } else { | |
170 | return found_id; | |
171 | } | |
172 | } | |
1f40704b ZD |
173 | |
174 | /* driver_override will always match, send a dummy id */ | |
343b7258 MG |
175 | if (dev->driver_override) |
176 | return &pci_device_id_any; | |
177 | return NULL; | |
1f40704b ZD |
178 | } |
179 | ||
1da177e4 | 180 | /** |
2f0cd59c | 181 | * new_id_store - sysfs frontend to pci_add_dynid() |
8f7020d3 RD |
182 | * @driver: target device driver |
183 | * @buf: buffer for scanning device ID data | |
184 | * @count: input size | |
1da177e4 | 185 | * |
9dba910e | 186 | * Allow PCI IDs to be added to an existing driver via sysfs. |
1da177e4 | 187 | */ |
a9427741 | 188 | static ssize_t new_id_store(struct device_driver *driver, const char *buf, |
3c78bc61 | 189 | size_t count) |
1da177e4 | 190 | { |
1da177e4 | 191 | struct pci_driver *pdrv = to_pci_driver(driver); |
b41d6cf3 | 192 | const struct pci_device_id *ids = pdrv->id_table; |
20a796a9 | 193 | u32 vendor, device, subvendor = PCI_ANY_ID, |
3c78bc61 RD |
194 | subdevice = PCI_ANY_ID, class = 0, class_mask = 0; |
195 | unsigned long driver_data = 0; | |
196 | int fields = 0; | |
8895d3bc | 197 | int retval = 0; |
1da177e4 | 198 | |
b41d6cf3 | 199 | fields = sscanf(buf, "%x %x %x %x %x %x %lx", |
1da177e4 LT |
200 | &vendor, &device, &subvendor, &subdevice, |
201 | &class, &class_mask, &driver_data); | |
6ba18636 | 202 | if (fields < 2) |
1da177e4 LT |
203 | return -EINVAL; |
204 | ||
8895d3bc BD |
205 | if (fields != 7) { |
206 | struct pci_dev *pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); | |
207 | if (!pdev) | |
208 | return -ENOMEM; | |
209 | ||
210 | pdev->vendor = vendor; | |
211 | pdev->device = device; | |
212 | pdev->subsystem_vendor = subvendor; | |
213 | pdev->subsystem_device = subdevice; | |
214 | pdev->class = class; | |
215 | ||
3853f912 | 216 | if (pci_match_device(pdrv, pdev)) |
8895d3bc BD |
217 | retval = -EEXIST; |
218 | ||
219 | kfree(pdev); | |
220 | ||
221 | if (retval) | |
222 | return retval; | |
223 | } | |
224 | ||
b41d6cf3 JD |
225 | /* Only accept driver_data values that match an existing id_table |
226 | entry */ | |
2debb4d2 CW |
227 | if (ids) { |
228 | retval = -EINVAL; | |
229 | while (ids->vendor || ids->subvendor || ids->class_mask) { | |
230 | if (driver_data == ids->driver_data) { | |
231 | retval = 0; | |
232 | break; | |
233 | } | |
234 | ids++; | |
b41d6cf3 | 235 | } |
2debb4d2 CW |
236 | if (retval) /* No match */ |
237 | return retval; | |
b41d6cf3 | 238 | } |
b41d6cf3 | 239 | |
9dba910e TH |
240 | retval = pci_add_dynid(pdrv, vendor, device, subvendor, subdevice, |
241 | class, class_mask, driver_data); | |
b19441af GKH |
242 | if (retval) |
243 | return retval; | |
1da177e4 LT |
244 | return count; |
245 | } | |
a9427741 | 246 | static DRIVER_ATTR_WO(new_id); |
1da177e4 | 247 | |
0994375e | 248 | /** |
2f0cd59c | 249 | * remove_id_store - remove a PCI device ID from this driver |
0994375e CW |
250 | * @driver: target device driver |
251 | * @buf: buffer for scanning device ID data | |
252 | * @count: input size | |
253 | * | |
254 | * Removes a dynamic pci device ID to this driver. | |
255 | */ | |
a9427741 | 256 | static ssize_t remove_id_store(struct device_driver *driver, const char *buf, |
3c78bc61 | 257 | size_t count) |
0994375e CW |
258 | { |
259 | struct pci_dynid *dynid, *n; | |
260 | struct pci_driver *pdrv = to_pci_driver(driver); | |
20a796a9 | 261 | u32 vendor, device, subvendor = PCI_ANY_ID, |
0994375e CW |
262 | subdevice = PCI_ANY_ID, class = 0, class_mask = 0; |
263 | int fields = 0; | |
9222097f | 264 | size_t retval = -ENODEV; |
0994375e CW |
265 | |
266 | fields = sscanf(buf, "%x %x %x %x %x %x", | |
267 | &vendor, &device, &subvendor, &subdevice, | |
268 | &class, &class_mask); | |
269 | if (fields < 2) | |
270 | return -EINVAL; | |
271 | ||
272 | spin_lock(&pdrv->dynids.lock); | |
273 | list_for_each_entry_safe(dynid, n, &pdrv->dynids.list, node) { | |
274 | struct pci_device_id *id = &dynid->id; | |
275 | if ((id->vendor == vendor) && | |
276 | (id->device == device) && | |
277 | (subvendor == PCI_ANY_ID || id->subvendor == subvendor) && | |
278 | (subdevice == PCI_ANY_ID || id->subdevice == subdevice) && | |
279 | !((id->class ^ class) & class_mask)) { | |
280 | list_del(&dynid->node); | |
281 | kfree(dynid); | |
9222097f | 282 | retval = count; |
0994375e CW |
283 | break; |
284 | } | |
285 | } | |
286 | spin_unlock(&pdrv->dynids.lock); | |
287 | ||
9222097f | 288 | return retval; |
0994375e | 289 | } |
a9427741 | 290 | static DRIVER_ATTR_WO(remove_id); |
0994375e | 291 | |
2229c1fb GKH |
292 | static struct attribute *pci_drv_attrs[] = { |
293 | &driver_attr_new_id.attr, | |
294 | &driver_attr_remove_id.attr, | |
295 | NULL, | |
bfb09a86 | 296 | }; |
2229c1fb | 297 | ATTRIBUTE_GROUPS(pci_drv); |
0994375e | 298 | |
873392ca RR |
299 | struct drv_dev_and_id { |
300 | struct pci_driver *drv; | |
301 | struct pci_dev *dev; | |
302 | const struct pci_device_id *id; | |
303 | }; | |
304 | ||
305 | static long local_pci_probe(void *_ddi) | |
306 | { | |
307 | struct drv_dev_and_id *ddi = _ddi; | |
967577b0 HY |
308 | struct pci_dev *pci_dev = ddi->dev; |
309 | struct pci_driver *pci_drv = ddi->drv; | |
310 | struct device *dev = &pci_dev->dev; | |
f3ec4f87 AS |
311 | int rc; |
312 | ||
967577b0 HY |
313 | /* |
314 | * Unbound PCI devices are always put in D0, regardless of | |
315 | * runtime PM status. During probe, the device is set to | |
316 | * active and the usage count is incremented. If the driver | |
a8360062 RW |
317 | * supports runtime PM, it should call pm_runtime_put_noidle(), |
318 | * or any other runtime PM helper function decrementing the usage | |
319 | * count, in its probe routine and pm_runtime_get_noresume() in | |
320 | * its remove routine. | |
f3ec4f87 | 321 | */ |
967577b0 | 322 | pm_runtime_get_sync(dev); |
68da4e0e | 323 | pci_dev->driver = pci_drv; |
967577b0 | 324 | rc = pci_drv->probe(pci_dev, ddi->id); |
f92d74c1 SC |
325 | if (!rc) |
326 | return rc; | |
327 | if (rc < 0) { | |
68da4e0e | 328 | pci_dev->driver = NULL; |
967577b0 | 329 | pm_runtime_put_sync(dev); |
f92d74c1 | 330 | return rc; |
f3ec4f87 | 331 | } |
f92d74c1 SC |
332 | /* |
333 | * Probe function should return < 0 for failure, 0 for success | |
334 | * Treat values > 0 as success, but warn. | |
335 | */ | |
6941a0c2 BH |
336 | pci_warn(pci_dev, "Driver probe function unexpectedly returned %d\n", |
337 | rc); | |
f92d74c1 | 338 | return 0; |
873392ca RR |
339 | } |
340 | ||
0b2c2a71 TG |
341 | static bool pci_physfn_is_probed(struct pci_dev *dev) |
342 | { | |
343 | #ifdef CONFIG_PCI_IOV | |
344 | return dev->is_virtfn && dev->physfn->is_probed; | |
345 | #else | |
346 | return false; | |
347 | #endif | |
348 | } | |
349 | ||
d42c6997 AK |
350 | static int pci_call_probe(struct pci_driver *drv, struct pci_dev *dev, |
351 | const struct pci_device_id *id) | |
352 | { | |
0b2c2a71 | 353 | int error, node, cpu; |
873392ca RR |
354 | struct drv_dev_and_id ddi = { drv, dev, id }; |
355 | ||
12c3156f AD |
356 | /* |
357 | * Execute driver initialization on node where the device is | |
358 | * attached. This way the driver likely allocates its local memory | |
359 | * on the right node. | |
360 | */ | |
873392ca | 361 | node = dev_to_node(&dev->dev); |
0b2c2a71 TG |
362 | dev->is_probed = 1; |
363 | ||
364 | cpu_hotplug_disable(); | |
12c3156f AD |
365 | |
366 | /* | |
0b2c2a71 TG |
367 | * Prevent nesting work_on_cpu() for the case where a Virtual Function |
368 | * device is probed from work_on_cpu() of the Physical device. | |
12c3156f | 369 | */ |
0b2c2a71 | 370 | if (node < 0 || node >= MAX_NUMNODES || !node_online(node) || |
9d42ea0d | 371 | pci_physfn_is_probed(dev)) { |
0b2c2a71 | 372 | cpu = nr_cpu_ids; |
9d42ea0d FW |
373 | } else { |
374 | cpumask_var_t wq_domain_mask; | |
375 | ||
376 | if (!zalloc_cpumask_var(&wq_domain_mask, GFP_KERNEL)) { | |
377 | error = -ENOMEM; | |
378 | goto out; | |
379 | } | |
380 | cpumask_and(wq_domain_mask, | |
04d4e665 FW |
381 | housekeeping_cpumask(HK_TYPE_WQ), |
382 | housekeeping_cpumask(HK_TYPE_DOMAIN)); | |
9d42ea0d | 383 | |
69a18b18 | 384 | cpu = cpumask_any_and(cpumask_of_node(node), |
9d42ea0d FW |
385 | wq_domain_mask); |
386 | free_cpumask_var(wq_domain_mask); | |
387 | } | |
0b2c2a71 TG |
388 | |
389 | if (cpu < nr_cpu_ids) | |
390 | error = work_on_cpu(cpu, local_pci_probe, &ddi); | |
391 | else | |
873392ca | 392 | error = local_pci_probe(&ddi); |
9d42ea0d | 393 | out: |
0b2c2a71 TG |
394 | dev->is_probed = 0; |
395 | cpu_hotplug_enable(); | |
d42c6997 AK |
396 | return error; |
397 | } | |
398 | ||
1da177e4 | 399 | /** |
23ea3793 | 400 | * __pci_device_probe - check if a driver wants to claim a specific PCI device |
8f7020d3 RD |
401 | * @drv: driver to call to check if it wants the PCI device |
402 | * @pci_dev: PCI device being probed | |
f7625980 | 403 | * |
8f7020d3 | 404 | * returns 0 on success, else error. |
68da4e0e | 405 | * side-effect: pci_dev->driver is set to drv when drv claims pci_dev. |
1da177e4 | 406 | */ |
3c78bc61 | 407 | static int __pci_device_probe(struct pci_driver *drv, struct pci_dev *pci_dev) |
75865858 GKH |
408 | { |
409 | const struct pci_device_id *id; | |
1da177e4 LT |
410 | int error = 0; |
411 | ||
ae232f09 | 412 | if (drv->probe) { |
75865858 GKH |
413 | error = -ENODEV; |
414 | ||
415 | id = pci_match_device(drv, pci_dev); | |
416 | if (id) | |
d42c6997 | 417 | error = pci_call_probe(drv, pci_dev, id); |
1da177e4 LT |
418 | } |
419 | return error; | |
420 | } | |
421 | ||
890e4847 JL |
422 | int __weak pcibios_alloc_irq(struct pci_dev *dev) |
423 | { | |
424 | return 0; | |
425 | } | |
426 | ||
427 | void __weak pcibios_free_irq(struct pci_dev *dev) | |
428 | { | |
429 | } | |
430 | ||
0e7df224 BW |
431 | #ifdef CONFIG_PCI_IOV |
432 | static inline bool pci_device_can_probe(struct pci_dev *pdev) | |
433 | { | |
2d2f4273 AW |
434 | return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe || |
435 | pdev->driver_override); | |
0e7df224 BW |
436 | } |
437 | #else | |
438 | static inline bool pci_device_can_probe(struct pci_dev *pdev) | |
439 | { | |
440 | return true; | |
441 | } | |
442 | #endif | |
443 | ||
3c78bc61 | 444 | static int pci_device_probe(struct device *dev) |
1da177e4 | 445 | { |
890e4847 JL |
446 | int error; |
447 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
448 | struct pci_driver *drv = to_pci_driver(dev->driver); | |
449 | ||
76002d8b AW |
450 | if (!pci_device_can_probe(pci_dev)) |
451 | return -ENODEV; | |
452 | ||
30fdfb92 MM |
453 | pci_assign_irq(pci_dev); |
454 | ||
890e4847 JL |
455 | error = pcibios_alloc_irq(pci_dev); |
456 | if (error < 0) | |
457 | return error; | |
1da177e4 | 458 | |
1da177e4 | 459 | pci_dev_get(pci_dev); |
76002d8b AW |
460 | error = __pci_device_probe(drv, pci_dev); |
461 | if (error) { | |
462 | pcibios_free_irq(pci_dev); | |
463 | pci_dev_put(pci_dev); | |
890e4847 | 464 | } |
1da177e4 LT |
465 | |
466 | return error; | |
467 | } | |
468 | ||
fc7a6209 | 469 | static void pci_device_remove(struct device *dev) |
1da177e4 | 470 | { |
3c78bc61 | 471 | struct pci_dev *pci_dev = to_pci_dev(dev); |
e0217c5b | 472 | struct pci_driver *drv = pci_dev->driver; |
1da177e4 | 473 | |
097d9d41 UKK |
474 | if (drv->remove) { |
475 | pm_runtime_get_sync(dev); | |
476 | drv->remove(pci_dev); | |
477 | pm_runtime_put_noidle(dev); | |
1da177e4 | 478 | } |
097d9d41 | 479 | pcibios_free_irq(pci_dev); |
68da4e0e | 480 | pci_dev->driver = NULL; |
097d9d41 | 481 | pci_iov_remove(pci_dev); |
1da177e4 | 482 | |
f3ec4f87 | 483 | /* Undo the runtime PM settings in local_pci_probe() */ |
967577b0 | 484 | pm_runtime_put_sync(dev); |
f3ec4f87 | 485 | |
2449e06a SL |
486 | /* |
487 | * If the device is still on, set the power state as "unknown", | |
488 | * since it might change by the next time we load the driver. | |
489 | */ | |
490 | if (pci_dev->current_state == PCI_D0) | |
491 | pci_dev->current_state = PCI_UNKNOWN; | |
492 | ||
1da177e4 LT |
493 | /* |
494 | * We would love to complain here if pci_dev->is_enabled is set, that | |
495 | * the driver should have called pci_disable_device(), but the | |
496 | * unfortunate fact is there are too many odd BIOS and bridge setups | |
f7625980 | 497 | * that don't like drivers doing that all of the time. |
1da177e4 LT |
498 | * Oh well, we can dream of sane hardware when we sleep, no matter how |
499 | * horrible the crap we have to deal with is when we are awake... | |
500 | */ | |
501 | ||
502 | pci_dev_put(pci_dev); | |
1da177e4 LT |
503 | } |
504 | ||
bbb44d9f RW |
505 | static void pci_device_shutdown(struct device *dev) |
506 | { | |
507 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
e0217c5b | 508 | struct pci_driver *drv = pci_dev->driver; |
bbb44d9f | 509 | |
3ff2de9b HY |
510 | pm_runtime_resume(dev); |
511 | ||
bbb44d9f RW |
512 | if (drv && drv->shutdown) |
513 | drv->shutdown(pci_dev); | |
5b415f1e | 514 | |
b566a22c | 515 | /* |
4fc9bbf9 KA |
516 | * If this is a kexec reboot, turn off Bus Master bit on the |
517 | * device to tell it to not continue to do DMA. Don't touch | |
518 | * devices in D3cold or unknown states. | |
519 | * If it is not a kexec reboot, firmware will hit the PCI | |
520 | * devices with big hammer and stop their DMA any way. | |
b566a22c | 521 | */ |
4fc9bbf9 | 522 | if (kexec_in_progress && (pci_dev->current_state <= PCI_D3hot)) |
6e0eda3c | 523 | pci_clear_master(pci_dev); |
bbb44d9f RW |
524 | } |
525 | ||
18a94192 | 526 | #ifdef CONFIG_PM_SLEEP |
6cbf8214 | 527 | |
18a94192 | 528 | /* Auxiliary functions used for system resume */ |
6cbf8214 RW |
529 | |
530 | /** | |
531 | * pci_restore_standard_config - restore standard config registers of PCI device | |
532 | * @pci_dev: PCI device to handle | |
533 | */ | |
534 | static int pci_restore_standard_config(struct pci_dev *pci_dev) | |
535 | { | |
536 | pci_update_current_state(pci_dev, PCI_UNKNOWN); | |
537 | ||
538 | if (pci_dev->current_state != PCI_D0) { | |
539 | int error = pci_set_power_state(pci_dev, PCI_D0); | |
540 | if (error) | |
541 | return error; | |
542 | } | |
543 | ||
1d3c16a8 | 544 | pci_restore_state(pci_dev); |
0ce3fcaf | 545 | pci_pme_restore(pci_dev); |
1d3c16a8 | 546 | return 0; |
6cbf8214 | 547 | } |
18a94192 KK |
548 | #endif /* CONFIG_PM_SLEEP */ |
549 | ||
550 | #ifdef CONFIG_PM | |
551 | ||
552 | /* Auxiliary functions used for system resume and run-time resume */ | |
6cbf8214 | 553 | |
f7b32a86 BH |
554 | static void pci_pm_default_resume(struct pci_dev *pci_dev) |
555 | { | |
556 | pci_fixup_device(pci_fixup_resume, pci_dev); | |
557 | pci_enable_wake(pci_dev, PCI_D0, false); | |
558 | } | |
559 | ||
0f40ac35 | 560 | static void pci_pm_power_up_and_verify_state(struct pci_dev *pci_dev) |
6cbf8214 | 561 | { |
db288c9c | 562 | pci_power_up(pci_dev); |
81cfa590 | 563 | pci_update_current_state(pci_dev, PCI_D0); |
0f40ac35 RW |
564 | } |
565 | ||
566 | static void pci_pm_default_resume_early(struct pci_dev *pci_dev) | |
567 | { | |
568 | pci_pm_power_up_and_verify_state(pci_dev); | |
db288c9c | 569 | pci_restore_state(pci_dev); |
0ce3fcaf | 570 | pci_pme_restore(pci_dev); |
6cbf8214 RW |
571 | } |
572 | ||
730643d3 RW |
573 | static void pci_pm_bridge_power_up_actions(struct pci_dev *pci_dev) |
574 | { | |
575 | pci_bridge_wait_for_secondary_bus(pci_dev); | |
576 | /* | |
577 | * When powering on a bridge from D3cold, the whole hierarchy may be | |
578 | * powered on into D0uninitialized state, resume them to give them a | |
579 | * chance to suspend again | |
580 | */ | |
581 | pci_resume_bus(pci_dev->subordinate); | |
582 | } | |
583 | ||
18a94192 KK |
584 | #endif /* CONFIG_PM */ |
585 | ||
586 | #ifdef CONFIG_PM_SLEEP | |
587 | ||
fa58d305 RW |
588 | /* |
589 | * Default "suspend" method for devices that have no driver provided suspend, | |
590 | * or not even a driver at all (second part). | |
bbb44d9f | 591 | */ |
bb808945 | 592 | static void pci_pm_set_unknown_state(struct pci_dev *pci_dev) |
bbb44d9f | 593 | { |
bbb44d9f RW |
594 | /* |
595 | * mark its power state as "unknown", since we don't know if | |
596 | * e.g. the BIOS will change its device state when we suspend. | |
597 | */ | |
598 | if (pci_dev->current_state == PCI_D0) | |
599 | pci_dev->current_state = PCI_UNKNOWN; | |
600 | } | |
601 | ||
355a72d7 RW |
602 | /* |
603 | * Default "resume" method for devices that have no driver provided resume, | |
604 | * or not even a driver at all (second part). | |
605 | */ | |
bb808945 | 606 | static int pci_pm_reenable_device(struct pci_dev *pci_dev) |
355a72d7 RW |
607 | { |
608 | int retval; | |
609 | ||
b2105b9f | 610 | /* if the device was enabled before suspend, re-enable */ |
bbb44d9f RW |
611 | retval = pci_reenable_device(pci_dev); |
612 | /* | |
613 | * if the device was busmaster before the suspend, make it busmaster | |
614 | * again | |
615 | */ | |
616 | if (pci_dev->is_busmaster) | |
617 | pci_set_master(pci_dev); | |
618 | ||
619 | return retval; | |
620 | } | |
621 | ||
622 | static int pci_legacy_suspend(struct device *dev, pm_message_t state) | |
1da177e4 | 623 | { |
3c78bc61 | 624 | struct pci_dev *pci_dev = to_pci_dev(dev); |
e0217c5b | 625 | struct pci_driver *drv = pci_dev->driver; |
46939f8b | 626 | |
02669492 | 627 | if (drv && drv->suspend) { |
99dadce8 | 628 | pci_power_t prev = pci_dev->current_state; |
46939f8b | 629 | int error; |
aa8c6c93 | 630 | |
57ef8026 | 631 | error = drv->suspend(pci_dev, state); |
a759de69 | 632 | suspend_report_result(dev, drv->suspend, error); |
57ef8026 FP |
633 | if (error) |
634 | return error; | |
aa8c6c93 | 635 | |
46939f8b | 636 | if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 |
99dadce8 | 637 | && pci_dev->current_state != PCI_UNKNOWN) { |
12bcae44 BH |
638 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
639 | "PCI PM: Device state not saved by %pS\n", | |
640 | drv->suspend); | |
99dadce8 | 641 | } |
02669492 | 642 | } |
ad8cfa1d RW |
643 | |
644 | pci_fixup_device(pci_fixup_suspend, pci_dev); | |
645 | ||
46939f8b | 646 | return 0; |
1da177e4 LT |
647 | } |
648 | ||
bbb44d9f | 649 | static int pci_legacy_suspend_late(struct device *dev, pm_message_t state) |
cbd69dbb | 650 | { |
3c78bc61 | 651 | struct pci_dev *pci_dev = to_pci_dev(dev); |
46939f8b RW |
652 | |
653 | if (!pci_dev->state_saved) | |
654 | pci_save_state(pci_dev); | |
655 | ||
656 | pci_pm_set_unknown_state(pci_dev); | |
657 | ||
7d2a01b8 AN |
658 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); |
659 | ||
46939f8b | 660 | return 0; |
cbd69dbb | 661 | } |
1da177e4 | 662 | |
bbb44d9f | 663 | static int pci_legacy_resume(struct device *dev) |
1da177e4 | 664 | { |
3c78bc61 | 665 | struct pci_dev *pci_dev = to_pci_dev(dev); |
e0217c5b | 666 | struct pci_driver *drv = pci_dev->driver; |
1da177e4 | 667 | |
ad8cfa1d RW |
668 | pci_fixup_device(pci_fixup_resume, pci_dev); |
669 | ||
aa8c6c93 RW |
670 | return drv && drv->resume ? |
671 | drv->resume(pci_dev) : pci_pm_reenable_device(pci_dev); | |
1da177e4 LT |
672 | } |
673 | ||
571ff758 RW |
674 | /* Auxiliary functions used by the new power management framework */ |
675 | ||
5294e256 | 676 | static void pci_pm_default_suspend(struct pci_dev *pci_dev) |
73410429 | 677 | { |
5294e256 | 678 | /* Disable non-bridge devices without PM support */ |
326c1cda | 679 | if (!pci_has_subordinate(pci_dev)) |
cbbc2f6b | 680 | pci_disable_enabled_device(pci_dev); |
73410429 RW |
681 | } |
682 | ||
07e836e8 RW |
683 | static bool pci_has_legacy_pm_support(struct pci_dev *pci_dev) |
684 | { | |
e0217c5b | 685 | struct pci_driver *drv = pci_dev->driver; |
1a1daf09 | 686 | bool ret = drv && (drv->suspend || drv->resume); |
bb808945 RW |
687 | |
688 | /* | |
689 | * Legacy PM support is used by default, so warn if the new framework is | |
690 | * supported as well. Drivers are supposed to support either the | |
691 | * former, or the latter, but not both at the same time. | |
692 | */ | |
12bcae44 BH |
693 | pci_WARN(pci_dev, ret && drv->driver.pm, "device %04x:%04x\n", |
694 | pci_dev->vendor, pci_dev->device); | |
bb808945 RW |
695 | |
696 | return ret; | |
07e836e8 RW |
697 | } |
698 | ||
571ff758 RW |
699 | /* New power management framework */ |
700 | ||
bbb44d9f RW |
701 | static int pci_pm_prepare(struct device *dev) |
702 | { | |
0c7376ad | 703 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 704 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 705 | |
6da2f2cc BH |
706 | if (pm && pm->prepare) { |
707 | int error = pm->prepare(dev); | |
08810a41 | 708 | if (error < 0) |
bac2a909 | 709 | return error; |
08810a41 RW |
710 | |
711 | if (!error && dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_PREPARE)) | |
712 | return 0; | |
bac2a909 | 713 | } |
0c7376ad RW |
714 | if (pci_dev_need_resume(pci_dev)) |
715 | return 0; | |
716 | ||
717 | /* | |
718 | * The PME setting needs to be adjusted here in case the direct-complete | |
719 | * optimization is used with respect to this device. | |
720 | */ | |
721 | pci_dev_adjust_pme(pci_dev); | |
722 | return 1; | |
bbb44d9f RW |
723 | } |
724 | ||
2cef548a RW |
725 | static void pci_pm_complete(struct device *dev) |
726 | { | |
a0d2a959 LW |
727 | struct pci_dev *pci_dev = to_pci_dev(dev); |
728 | ||
729 | pci_dev_complete_resume(pci_dev); | |
730 | pm_generic_complete(dev); | |
731 | ||
732 | /* Resume device if platform firmware has put it in reset-power-on */ | |
bd755d77 | 733 | if (pm_runtime_suspended(dev) && pm_resume_via_firmware()) { |
a0d2a959 LW |
734 | pci_power_t pre_sleep_state = pci_dev->current_state; |
735 | ||
b51033e0 RW |
736 | pci_refresh_power_state(pci_dev); |
737 | /* | |
738 | * On platforms with ACPI this check may also trigger for | |
739 | * devices sharing power resources if one of those power | |
740 | * resources has been activated as a result of a change of the | |
741 | * power state of another device sharing it. However, in that | |
742 | * case it is also better to resume the device, in general. | |
743 | */ | |
a0d2a959 LW |
744 | if (pci_dev->current_state < pre_sleep_state) |
745 | pm_request_resume(dev); | |
746 | } | |
2cef548a | 747 | } |
bbb44d9f | 748 | |
6cbf8214 RW |
749 | #else /* !CONFIG_PM_SLEEP */ |
750 | ||
751 | #define pci_pm_prepare NULL | |
2cef548a | 752 | #define pci_pm_complete NULL |
6cbf8214 RW |
753 | |
754 | #endif /* !CONFIG_PM_SLEEP */ | |
755 | ||
bbb44d9f | 756 | #ifdef CONFIG_SUSPEND |
a39bd851 BH |
757 | static void pcie_pme_root_status_cleanup(struct pci_dev *pci_dev) |
758 | { | |
759 | /* | |
760 | * Some BIOSes forget to clear Root PME Status bits after system | |
761 | * wakeup, which breaks ACPI-based runtime wakeup on PCI Express. | |
762 | * Clear those bits now just in case (shouldn't hurt). | |
763 | */ | |
764 | if (pci_is_pcie(pci_dev) && | |
3620c714 BH |
765 | (pci_pcie_type(pci_dev) == PCI_EXP_TYPE_ROOT_PORT || |
766 | pci_pcie_type(pci_dev) == PCI_EXP_TYPE_RC_EC)) | |
a39bd851 BH |
767 | pcie_clear_root_pme_status(pci_dev); |
768 | } | |
bbb44d9f RW |
769 | |
770 | static int pci_pm_suspend(struct device *dev) | |
771 | { | |
772 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
8150f32b | 773 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 774 | |
d491f2b7 RW |
775 | pci_dev->skip_bus_pm = false; |
776 | ||
c01163db BH |
777 | /* |
778 | * Disabling PTM allows some systems, e.g., Intel mobile chips | |
779 | * since Coffee Lake, to enter a lower-power PM state. | |
780 | */ | |
781 | pci_suspend_ptm(pci_dev); | |
782 | ||
ad8cfa1d RW |
783 | if (pci_has_legacy_pm_support(pci_dev)) |
784 | return pci_legacy_suspend(dev, PMSG_SUSPEND); | |
bb808945 | 785 | |
5294e256 RW |
786 | if (!pm) { |
787 | pci_pm_default_suspend(pci_dev); | |
c4b65157 | 788 | return 0; |
5294e256 RW |
789 | } |
790 | ||
7cd0602d | 791 | /* |
c4b65157 RW |
792 | * PCI devices suspended at run time may need to be resumed at this |
793 | * point, because in general it may be necessary to reconfigure them for | |
794 | * system suspend. Namely, if the device is expected to wake up the | |
795 | * system from the sleep state, it may have to be reconfigured for this | |
796 | * purpose, or if the device is not expected to wake up the system from | |
797 | * the sleep state, it should be prevented from signaling wakeup events | |
798 | * going forward. | |
799 | * | |
800 | * Also if the driver of the device does not indicate that its system | |
801 | * suspend callbacks can cope with runtime-suspended devices, it is | |
802 | * better to resume the device from runtime suspend here. | |
7cd0602d | 803 | */ |
c4b65157 | 804 | if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || |
0c7376ad | 805 | pci_dev_need_resume(pci_dev)) { |
c4b65157 | 806 | pm_runtime_resume(dev); |
656088aa | 807 | pci_dev->state_saved = false; |
0c7376ad RW |
808 | } else { |
809 | pci_dev_adjust_pme(pci_dev); | |
656088aa | 810 | } |
7cd0602d | 811 | |
5294e256 RW |
812 | if (pm->suspend) { |
813 | pci_power_t prev = pci_dev->current_state; | |
814 | int error; | |
815 | ||
ddb7c9d2 | 816 | error = pm->suspend(dev); |
a759de69 | 817 | suspend_report_result(dev, pm->suspend, error); |
5294e256 RW |
818 | if (error) |
819 | return error; | |
820 | ||
46939f8b | 821 | if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 |
5294e256 | 822 | && pci_dev->current_state != PCI_UNKNOWN) { |
12bcae44 BH |
823 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
824 | "PCI PM: State of device not saved by %pS\n", | |
825 | pm->suspend); | |
5294e256 | 826 | } |
bbb44d9f | 827 | } |
fa58d305 | 828 | |
5294e256 | 829 | return 0; |
bbb44d9f RW |
830 | } |
831 | ||
c4b65157 RW |
832 | static int pci_pm_suspend_late(struct device *dev) |
833 | { | |
fa2bfead | 834 | if (dev_pm_skip_suspend(dev)) |
c4b65157 RW |
835 | return 0; |
836 | ||
837 | pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); | |
838 | ||
839 | return pm_generic_suspend_late(dev); | |
840 | } | |
841 | ||
bbb44d9f | 842 | static int pci_pm_suspend_noirq(struct device *dev) |
c8958177 | 843 | { |
355a72d7 | 844 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 845 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
c8958177 | 846 | |
fa2bfead | 847 | if (dev_pm_skip_suspend(dev)) |
c4b65157 RW |
848 | return 0; |
849 | ||
bb808945 RW |
850 | if (pci_has_legacy_pm_support(pci_dev)) |
851 | return pci_legacy_suspend_late(dev, PMSG_SUSPEND); | |
852 | ||
931ff68a RW |
853 | if (!pm) { |
854 | pci_save_state(pci_dev); | |
7d2a01b8 | 855 | goto Fixup; |
931ff68a | 856 | } |
46939f8b RW |
857 | |
858 | if (pm->suspend_noirq) { | |
859 | pci_power_t prev = pci_dev->current_state; | |
860 | int error; | |
861 | ||
862 | error = pm->suspend_noirq(dev); | |
a759de69 | 863 | suspend_report_result(dev, pm->suspend_noirq, error); |
46939f8b RW |
864 | if (error) |
865 | return error; | |
866 | ||
867 | if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 | |
868 | && pci_dev->current_state != PCI_UNKNOWN) { | |
12bcae44 BH |
869 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
870 | "PCI PM: State of device not saved by %pS\n", | |
871 | pm->suspend_noirq); | |
7d2a01b8 | 872 | goto Fixup; |
46939f8b | 873 | } |
bbb44d9f RW |
874 | } |
875 | ||
4c00cba1 RJ |
876 | if (!pci_dev->state_saved) { |
877 | pci_save_state(pci_dev); | |
878 | ||
d491f2b7 | 879 | /* |
4c00cba1 RJ |
880 | * If the device is a bridge with a child in D0 below it, |
881 | * it needs to stay in D0, so check skip_bus_pm to avoid | |
882 | * putting it into a low-power state in that case. | |
d491f2b7 | 883 | */ |
4c00cba1 | 884 | if (!pci_dev->skip_bus_pm && pci_power_manageable(pci_dev)) |
46939f8b RW |
885 | pci_prepare_to_sleep(pci_dev); |
886 | } | |
d67e37d7 | 887 | |
6941a0c2 | 888 | pci_dbg(pci_dev, "PCI PM: Suspend power state: %s\n", |
ca67ab5c RW |
889 | pci_power_name(pci_dev->current_state)); |
890 | ||
3e26c5fe RW |
891 | if (pci_dev->current_state == PCI_D0) { |
892 | pci_dev->skip_bus_pm = true; | |
893 | /* | |
894 | * Per PCI PM r1.2, table 6-1, a bridge must be in D0 if any | |
895 | * downstream device is in D0, so avoid changing the power state | |
896 | * of the parent bridge by setting the skip_bus_pm flag for it. | |
897 | */ | |
898 | if (pci_dev->bus->self) | |
899 | pci_dev->bus->self->skip_bus_pm = true; | |
900 | } | |
901 | ||
471a739a | 902 | if (pci_dev->skip_bus_pm && pm_suspend_no_platform()) { |
6941a0c2 | 903 | pci_dbg(pci_dev, "PCI PM: Skipped\n"); |
3e26c5fe RW |
904 | goto Fixup; |
905 | } | |
906 | ||
46939f8b RW |
907 | pci_pm_set_unknown_state(pci_dev); |
908 | ||
dbf0e4c7 AS |
909 | /* |
910 | * Some BIOSes from ASUS have a bug: If a USB EHCI host controller's | |
911 | * PCI COMMAND register isn't 0, the BIOS assumes that the controller | |
912 | * hasn't been quiesced and tries to turn it off. If the controller | |
913 | * is already in D3, this can hang or cause memory corruption. | |
914 | * | |
915 | * Since the value of the COMMAND register doesn't matter once the | |
916 | * device has been suspended, we can safely set it to 0 here. | |
917 | */ | |
918 | if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) | |
919 | pci_write_config_word(pci_dev, PCI_COMMAND, 0); | |
920 | ||
7d2a01b8 AN |
921 | Fixup: |
922 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); | |
923 | ||
bd755d77 RW |
924 | /* |
925 | * If the target system sleep state is suspend-to-idle, it is sufficient | |
926 | * to check whether or not the device's wakeup settings are good for | |
927 | * runtime PM. Otherwise, the pm_resume_via_firmware() check will cause | |
928 | * pci_pm_complete() to take care of fixing up the device's state | |
929 | * anyway, if need be. | |
930 | */ | |
0fe8a1be RW |
931 | if (device_can_wakeup(dev) && !device_may_wakeup(dev)) |
932 | dev->power.may_skip_resume = false; | |
bd755d77 | 933 | |
46939f8b | 934 | return 0; |
c8958177 | 935 | } |
1da177e4 | 936 | |
f6dc1e5e | 937 | static int pci_pm_resume_noirq(struct device *dev) |
bbb44d9f RW |
938 | { |
939 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
6da2f2cc | 940 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
ad9001f2 MW |
941 | pci_power_t prev_state = pci_dev->current_state; |
942 | bool skip_bus_pm = pci_dev->skip_bus_pm; | |
bbb44d9f | 943 | |
76c70cb5 | 944 | if (dev_pm_skip_resume(dev)) |
bd755d77 RW |
945 | return 0; |
946 | ||
3e26c5fe RW |
947 | /* |
948 | * In the suspend-to-idle case, devices left in D0 during suspend will | |
949 | * stay in D0, so it is not necessary to restore or update their | |
471a739a RW |
950 | * configuration here and attempting to put them into D0 again is |
951 | * pointless, so avoid doing that. | |
3e26c5fe | 952 | */ |
ad9001f2 | 953 | if (!(skip_bus_pm && pm_suspend_no_platform())) |
3e26c5fe RW |
954 | pci_pm_default_resume_early(pci_dev); |
955 | ||
956 | pci_fixup_device(pci_fixup_resume_early, pci_dev); | |
ec6a75ef | 957 | pcie_pme_root_status_cleanup(pci_dev); |
aa8c6c93 | 958 | |
ad9001f2 | 959 | if (!skip_bus_pm && prev_state == PCI_D3cold) |
730643d3 | 960 | pci_pm_bridge_power_up_actions(pci_dev); |
ad9001f2 | 961 | |
ad8cfa1d | 962 | if (pci_has_legacy_pm_support(pci_dev)) |
89cdbc35 | 963 | return 0; |
bb808945 | 964 | |
6da2f2cc BH |
965 | if (pm && pm->resume_noirq) |
966 | return pm->resume_noirq(dev); | |
bbb44d9f | 967 | |
6da2f2cc | 968 | return 0; |
bbb44d9f RW |
969 | } |
970 | ||
6e176bf8 RW |
971 | static int pci_pm_resume_early(struct device *dev) |
972 | { | |
76c70cb5 | 973 | if (dev_pm_skip_resume(dev)) |
6e176bf8 RW |
974 | return 0; |
975 | ||
976 | return pm_generic_resume_early(dev); | |
977 | } | |
978 | ||
f6dc1e5e | 979 | static int pci_pm_resume(struct device *dev) |
bbb44d9f | 980 | { |
355a72d7 | 981 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 982 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 983 | |
418e4da3 RW |
984 | /* |
985 | * This is necessary for the suspend error path in which resume is | |
986 | * called without restoring the standard config registers of the device. | |
987 | */ | |
988 | if (pci_dev->state_saved) | |
989 | pci_restore_standard_config(pci_dev); | |
990 | ||
c01163db BH |
991 | pci_resume_ptm(pci_dev); |
992 | ||
ad8cfa1d | 993 | if (pci_has_legacy_pm_support(pci_dev)) |
f6dc1e5e | 994 | return pci_legacy_resume(dev); |
bb808945 | 995 | |
5294e256 | 996 | pci_pm_default_resume(pci_dev); |
73410429 | 997 | |
5294e256 RW |
998 | if (pm) { |
999 | if (pm->resume) | |
6da2f2cc | 1000 | return pm->resume(dev); |
5294e256 RW |
1001 | } else { |
1002 | pci_pm_reenable_device(pci_dev); | |
1003 | } | |
bbb44d9f | 1004 | |
6da2f2cc | 1005 | return 0; |
bbb44d9f RW |
1006 | } |
1007 | ||
1008 | #else /* !CONFIG_SUSPEND */ | |
1009 | ||
1010 | #define pci_pm_suspend NULL | |
c4b65157 | 1011 | #define pci_pm_suspend_late NULL |
bbb44d9f RW |
1012 | #define pci_pm_suspend_noirq NULL |
1013 | #define pci_pm_resume NULL | |
6e176bf8 | 1014 | #define pci_pm_resume_early NULL |
bbb44d9f RW |
1015 | #define pci_pm_resume_noirq NULL |
1016 | ||
1017 | #endif /* !CONFIG_SUSPEND */ | |
1018 | ||
1f112cee | 1019 | #ifdef CONFIG_HIBERNATE_CALLBACKS |
bbb44d9f RW |
1020 | |
1021 | static int pci_pm_freeze(struct device *dev) | |
1022 | { | |
1023 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
8150f32b | 1024 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1025 | |
ad8cfa1d RW |
1026 | if (pci_has_legacy_pm_support(pci_dev)) |
1027 | return pci_legacy_suspend(dev, PMSG_FREEZE); | |
bb808945 | 1028 | |
5294e256 RW |
1029 | if (!pm) { |
1030 | pci_pm_default_suspend(pci_dev); | |
1031 | return 0; | |
bbb44d9f RW |
1032 | } |
1033 | ||
7cd0602d | 1034 | /* |
501debd4 RW |
1035 | * Resume all runtime-suspended devices before creating a snapshot |
1036 | * image of system memory, because the restore kernel generally cannot | |
1037 | * be expected to always handle them consistently and they need to be | |
1038 | * put into the runtime-active metastate during system resume anyway, | |
1039 | * so it is better to ensure that the state saved in the image will be | |
1040 | * always consistent with that. | |
7cd0602d | 1041 | */ |
501debd4 RW |
1042 | pm_runtime_resume(dev); |
1043 | pci_dev->state_saved = false; | |
7cd0602d | 1044 | |
5294e256 RW |
1045 | if (pm->freeze) { |
1046 | int error; | |
1047 | ||
1048 | error = pm->freeze(dev); | |
a759de69 | 1049 | suspend_report_result(dev, pm->freeze, error); |
5294e256 RW |
1050 | if (error) |
1051 | return error; | |
1052 | } | |
1053 | ||
5294e256 | 1054 | return 0; |
bbb44d9f RW |
1055 | } |
1056 | ||
1057 | static int pci_pm_freeze_noirq(struct device *dev) | |
1058 | { | |
355a72d7 | 1059 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 1060 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1061 | |
bb808945 RW |
1062 | if (pci_has_legacy_pm_support(pci_dev)) |
1063 | return pci_legacy_suspend_late(dev, PMSG_FREEZE); | |
1064 | ||
6da2f2cc | 1065 | if (pm && pm->freeze_noirq) { |
46939f8b RW |
1066 | int error; |
1067 | ||
6da2f2cc | 1068 | error = pm->freeze_noirq(dev); |
a759de69 | 1069 | suspend_report_result(dev, pm->freeze_noirq, error); |
46939f8b RW |
1070 | if (error) |
1071 | return error; | |
bbb44d9f RW |
1072 | } |
1073 | ||
46939f8b RW |
1074 | if (!pci_dev->state_saved) |
1075 | pci_save_state(pci_dev); | |
d67e37d7 | 1076 | |
46939f8b RW |
1077 | pci_pm_set_unknown_state(pci_dev); |
1078 | ||
1079 | return 0; | |
bbb44d9f RW |
1080 | } |
1081 | ||
f6dc1e5e | 1082 | static int pci_pm_thaw_noirq(struct device *dev) |
bbb44d9f | 1083 | { |
355a72d7 | 1084 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 1085 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
699c1985 | 1086 | |
5839ee73 | 1087 | /* |
89cdbc35 BH |
1088 | * The pm->thaw_noirq() callback assumes the device has been |
1089 | * returned to D0 and its config state has been restored. | |
f2c33cca DC |
1090 | * |
1091 | * In addition, pci_restore_state() restores MSI-X state in MMIO | |
1092 | * space, which requires the device to be in D0, so return it to D0 | |
1093 | * in case the driver's "freeze" callbacks put it into a low-power | |
1094 | * state. | |
5839ee73 | 1095 | */ |
0f40ac35 | 1096 | pci_pm_power_up_and_verify_state(pci_dev); |
e60514bd | 1097 | pci_restore_state(pci_dev); |
d67e37d7 | 1098 | |
f2c33cca | 1099 | if (pci_has_legacy_pm_support(pci_dev)) |
89cdbc35 | 1100 | return 0; |
f2c33cca | 1101 | |
6da2f2cc BH |
1102 | if (pm && pm->thaw_noirq) |
1103 | return pm->thaw_noirq(dev); | |
bbb44d9f | 1104 | |
6da2f2cc | 1105 | return 0; |
bbb44d9f RW |
1106 | } |
1107 | ||
f6dc1e5e | 1108 | static int pci_pm_thaw(struct device *dev) |
bbb44d9f | 1109 | { |
355a72d7 | 1110 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 1111 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f RW |
1112 | int error = 0; |
1113 | ||
ad8cfa1d | 1114 | if (pci_has_legacy_pm_support(pci_dev)) |
f6dc1e5e | 1115 | return pci_legacy_resume(dev); |
bb808945 | 1116 | |
5294e256 RW |
1117 | if (pm) { |
1118 | if (pm->thaw) | |
1119 | error = pm->thaw(dev); | |
1120 | } else { | |
1121 | pci_pm_reenable_device(pci_dev); | |
1122 | } | |
bbb44d9f | 1123 | |
4b77b0a2 RW |
1124 | pci_dev->state_saved = false; |
1125 | ||
bbb44d9f RW |
1126 | return error; |
1127 | } | |
1128 | ||
1129 | static int pci_pm_poweroff(struct device *dev) | |
1130 | { | |
355a72d7 | 1131 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 1132 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1133 | |
ad8cfa1d RW |
1134 | if (pci_has_legacy_pm_support(pci_dev)) |
1135 | return pci_legacy_suspend(dev, PMSG_HIBERNATE); | |
bb808945 | 1136 | |
5294e256 RW |
1137 | if (!pm) { |
1138 | pci_pm_default_suspend(pci_dev); | |
c4b65157 | 1139 | return 0; |
5294e256 RW |
1140 | } |
1141 | ||
7cd0602d | 1142 | /* The reason to do that is the same as in pci_pm_suspend(). */ |
c4b65157 | 1143 | if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || |
0c7376ad | 1144 | pci_dev_need_resume(pci_dev)) { |
c4b65157 | 1145 | pm_runtime_resume(dev); |
0c7376ad RW |
1146 | pci_dev->state_saved = false; |
1147 | } else { | |
1148 | pci_dev_adjust_pme(pci_dev); | |
1149 | } | |
7cd0602d | 1150 | |
5294e256 | 1151 | if (pm->poweroff) { |
46939f8b RW |
1152 | int error; |
1153 | ||
ddb7c9d2 | 1154 | error = pm->poweroff(dev); |
a759de69 | 1155 | suspend_report_result(dev, pm->poweroff, error); |
46939f8b RW |
1156 | if (error) |
1157 | return error; | |
bbb44d9f RW |
1158 | } |
1159 | ||
46939f8b | 1160 | return 0; |
bbb44d9f | 1161 | } |
c9b9972b | 1162 | |
c4b65157 RW |
1163 | static int pci_pm_poweroff_late(struct device *dev) |
1164 | { | |
fa2bfead | 1165 | if (dev_pm_skip_suspend(dev)) |
c4b65157 | 1166 | return 0; |
699c1985 | 1167 | |
c4b65157 RW |
1168 | pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); |
1169 | ||
1170 | return pm_generic_poweroff_late(dev); | |
bbb44d9f RW |
1171 | } |
1172 | ||
1173 | static int pci_pm_poweroff_noirq(struct device *dev) | |
1174 | { | |
46939f8b | 1175 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 1176 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1177 | |
fa2bfead | 1178 | if (dev_pm_skip_suspend(dev)) |
c4b65157 RW |
1179 | return 0; |
1180 | ||
6da2f2cc | 1181 | if (pci_has_legacy_pm_support(pci_dev)) |
bb808945 RW |
1182 | return pci_legacy_suspend_late(dev, PMSG_HIBERNATE); |
1183 | ||
6da2f2cc | 1184 | if (!pm) { |
7d2a01b8 | 1185 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); |
46939f8b | 1186 | return 0; |
7d2a01b8 | 1187 | } |
46939f8b | 1188 | |
6da2f2cc | 1189 | if (pm->poweroff_noirq) { |
46939f8b RW |
1190 | int error; |
1191 | ||
6da2f2cc | 1192 | error = pm->poweroff_noirq(dev); |
a759de69 | 1193 | suspend_report_result(dev, pm->poweroff_noirq, error); |
46939f8b RW |
1194 | if (error) |
1195 | return error; | |
bbb44d9f RW |
1196 | } |
1197 | ||
326c1cda | 1198 | if (!pci_dev->state_saved && !pci_has_subordinate(pci_dev)) |
46939f8b RW |
1199 | pci_prepare_to_sleep(pci_dev); |
1200 | ||
0b68c8e2 RW |
1201 | /* |
1202 | * The reason for doing this here is the same as for the analogous code | |
1203 | * in pci_pm_suspend_noirq(). | |
1204 | */ | |
1205 | if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) | |
1206 | pci_write_config_word(pci_dev, PCI_COMMAND, 0); | |
1207 | ||
7d2a01b8 AN |
1208 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); |
1209 | ||
46939f8b | 1210 | return 0; |
bbb44d9f RW |
1211 | } |
1212 | ||
f6dc1e5e | 1213 | static int pci_pm_restore_noirq(struct device *dev) |
bbb44d9f RW |
1214 | { |
1215 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
6da2f2cc | 1216 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
699c1985 | 1217 | |
6cbf8214 | 1218 | pci_pm_default_resume_early(pci_dev); |
3e26c5fe | 1219 | pci_fixup_device(pci_fixup_resume_early, pci_dev); |
aa8c6c93 | 1220 | |
ad8cfa1d | 1221 | if (pci_has_legacy_pm_support(pci_dev)) |
89cdbc35 | 1222 | return 0; |
bb808945 | 1223 | |
6da2f2cc BH |
1224 | if (pm && pm->restore_noirq) |
1225 | return pm->restore_noirq(dev); | |
bbb44d9f | 1226 | |
6da2f2cc | 1227 | return 0; |
bbb44d9f RW |
1228 | } |
1229 | ||
f6dc1e5e | 1230 | static int pci_pm_restore(struct device *dev) |
bbb44d9f RW |
1231 | { |
1232 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
8150f32b | 1233 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1234 | |
418e4da3 RW |
1235 | /* |
1236 | * This is necessary for the hibernation error path in which restore is | |
1237 | * called without restoring the standard config registers of the device. | |
1238 | */ | |
1239 | if (pci_dev->state_saved) | |
1240 | pci_restore_standard_config(pci_dev); | |
1241 | ||
ad8cfa1d | 1242 | if (pci_has_legacy_pm_support(pci_dev)) |
f6dc1e5e | 1243 | return pci_legacy_resume(dev); |
bb808945 | 1244 | |
5294e256 | 1245 | pci_pm_default_resume(pci_dev); |
73410429 | 1246 | |
5294e256 RW |
1247 | if (pm) { |
1248 | if (pm->restore) | |
6da2f2cc | 1249 | return pm->restore(dev); |
5294e256 RW |
1250 | } else { |
1251 | pci_pm_reenable_device(pci_dev); | |
1252 | } | |
bbb44d9f | 1253 | |
6da2f2cc | 1254 | return 0; |
c8958177 | 1255 | } |
1da177e4 | 1256 | |
1f112cee | 1257 | #else /* !CONFIG_HIBERNATE_CALLBACKS */ |
bbb44d9f RW |
1258 | |
1259 | #define pci_pm_freeze NULL | |
1260 | #define pci_pm_freeze_noirq NULL | |
1261 | #define pci_pm_thaw NULL | |
1262 | #define pci_pm_thaw_noirq NULL | |
1263 | #define pci_pm_poweroff NULL | |
c4b65157 | 1264 | #define pci_pm_poweroff_late NULL |
bbb44d9f RW |
1265 | #define pci_pm_poweroff_noirq NULL |
1266 | #define pci_pm_restore NULL | |
1267 | #define pci_pm_restore_noirq NULL | |
1268 | ||
1f112cee | 1269 | #endif /* !CONFIG_HIBERNATE_CALLBACKS */ |
bbb44d9f | 1270 | |
fbb988be | 1271 | #ifdef CONFIG_PM |
6cbf8214 RW |
1272 | |
1273 | static int pci_pm_runtime_suspend(struct device *dev) | |
1274 | { | |
1275 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
1276 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; | |
1277 | pci_power_t prev = pci_dev->current_state; | |
1278 | int error; | |
1279 | ||
c01163db BH |
1280 | pci_suspend_ptm(pci_dev); |
1281 | ||
967577b0 | 1282 | /* |
e0217c5b BH |
1283 | * If pci_dev->driver is not set (unbound), we leave the device in D0, |
1284 | * but it may go to D3cold when the bridge above it runtime suspends. | |
1285 | * Save its config space in case that happens. | |
967577b0 | 1286 | */ |
e0217c5b | 1287 | if (!pci_dev->driver) { |
5775b843 | 1288 | pci_save_state(pci_dev); |
967577b0 | 1289 | return 0; |
5775b843 | 1290 | } |
967577b0 | 1291 | |
82fee4d6 | 1292 | pci_dev->state_saved = false; |
c5eb1190 JN |
1293 | if (pm && pm->runtime_suspend) { |
1294 | error = pm->runtime_suspend(dev); | |
06bf403d ID |
1295 | /* |
1296 | * -EBUSY and -EAGAIN is used to request the runtime PM core | |
1297 | * to schedule a new suspend, so log the event only with debug | |
1298 | * log level. | |
1299 | */ | |
c5eb1190 | 1300 | if (error == -EBUSY || error == -EAGAIN) { |
6941a0c2 | 1301 | pci_dbg(pci_dev, "can't suspend now (%ps returned %d)\n", |
06bf403d | 1302 | pm->runtime_suspend, error); |
c5eb1190 JN |
1303 | return error; |
1304 | } else if (error) { | |
6941a0c2 | 1305 | pci_err(pci_dev, "can't suspend (%ps returned %d)\n", |
06bf403d | 1306 | pm->runtime_suspend, error); |
c5eb1190 JN |
1307 | return error; |
1308 | } | |
06bf403d | 1309 | } |
6cbf8214 RW |
1310 | |
1311 | pci_fixup_device(pci_fixup_suspend, pci_dev); | |
1312 | ||
c5eb1190 JN |
1313 | if (pm && pm->runtime_suspend |
1314 | && !pci_dev->state_saved && pci_dev->current_state != PCI_D0 | |
6cbf8214 | 1315 | && pci_dev->current_state != PCI_UNKNOWN) { |
12bcae44 BH |
1316 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
1317 | "PCI PM: State of device not saved by %pS\n", | |
1318 | pm->runtime_suspend); | |
6cbf8214 RW |
1319 | return 0; |
1320 | } | |
1321 | ||
42eca230 | 1322 | if (!pci_dev->state_saved) { |
6cbf8214 | 1323 | pci_save_state(pci_dev); |
42eca230 DA |
1324 | pci_finish_runtime_suspend(pci_dev); |
1325 | } | |
6cbf8214 RW |
1326 | |
1327 | return 0; | |
1328 | } | |
1329 | ||
1330 | static int pci_pm_runtime_resume(struct device *dev) | |
1331 | { | |
1332 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
1333 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; | |
ad9001f2 | 1334 | pci_power_t prev_state = pci_dev->current_state; |
6da2f2cc | 1335 | int error = 0; |
6cbf8214 | 1336 | |
967577b0 | 1337 | /* |
5775b843 RW |
1338 | * Restoring config space is necessary even if the device is not bound |
1339 | * to a driver because although we left it in D0, it may have gone to | |
1340 | * D3cold when the bridge above it runtime suspended. | |
967577b0 | 1341 | */ |
9a605831 | 1342 | pci_pm_default_resume_early(pci_dev); |
c01163db | 1343 | pci_resume_ptm(pci_dev); |
5775b843 | 1344 | |
e0217c5b | 1345 | if (!pci_dev->driver) |
967577b0 HY |
1346 | return 0; |
1347 | ||
db288c9c | 1348 | pci_fixup_device(pci_fixup_resume_early, pci_dev); |
f7b32a86 | 1349 | pci_pm_default_resume(pci_dev); |
6cbf8214 | 1350 | |
ad9001f2 | 1351 | if (prev_state == PCI_D3cold) |
730643d3 | 1352 | pci_pm_bridge_power_up_actions(pci_dev); |
ad9001f2 | 1353 | |
c5eb1190 | 1354 | if (pm && pm->runtime_resume) |
6da2f2cc | 1355 | error = pm->runtime_resume(dev); |
448bd857 | 1356 | |
6da2f2cc | 1357 | return error; |
6cbf8214 RW |
1358 | } |
1359 | ||
1360 | static int pci_pm_runtime_idle(struct device *dev) | |
1361 | { | |
e0217c5b | 1362 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6cbf8214 RW |
1363 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
1364 | ||
967577b0 | 1365 | /* |
e0217c5b BH |
1366 | * If pci_dev->driver is not set (unbound), the device should |
1367 | * always remain in D0 regardless of the runtime PM status | |
967577b0 | 1368 | */ |
e0217c5b | 1369 | if (!pci_dev->driver) |
45f0a85c | 1370 | return 0; |
967577b0 | 1371 | |
6cbf8214 RW |
1372 | if (!pm) |
1373 | return -ENOSYS; | |
1374 | ||
45f0a85c | 1375 | if (pm->runtime_idle) |
6da2f2cc | 1376 | return pm->runtime_idle(dev); |
6cbf8214 | 1377 | |
6da2f2cc | 1378 | return 0; |
6cbf8214 RW |
1379 | } |
1380 | ||
f91da04d | 1381 | static const struct dev_pm_ops pci_dev_pm_ops = { |
adf09493 | 1382 | .prepare = pci_pm_prepare, |
2cef548a | 1383 | .complete = pci_pm_complete, |
adf09493 | 1384 | .suspend = pci_pm_suspend, |
c4b65157 | 1385 | .suspend_late = pci_pm_suspend_late, |
adf09493 | 1386 | .resume = pci_pm_resume, |
6e176bf8 | 1387 | .resume_early = pci_pm_resume_early, |
adf09493 RW |
1388 | .freeze = pci_pm_freeze, |
1389 | .thaw = pci_pm_thaw, | |
1390 | .poweroff = pci_pm_poweroff, | |
c4b65157 | 1391 | .poweroff_late = pci_pm_poweroff_late, |
adf09493 | 1392 | .restore = pci_pm_restore, |
bbb44d9f RW |
1393 | .suspend_noirq = pci_pm_suspend_noirq, |
1394 | .resume_noirq = pci_pm_resume_noirq, | |
1395 | .freeze_noirq = pci_pm_freeze_noirq, | |
1396 | .thaw_noirq = pci_pm_thaw_noirq, | |
1397 | .poweroff_noirq = pci_pm_poweroff_noirq, | |
1398 | .restore_noirq = pci_pm_restore_noirq, | |
6cbf8214 RW |
1399 | .runtime_suspend = pci_pm_runtime_suspend, |
1400 | .runtime_resume = pci_pm_runtime_resume, | |
1401 | .runtime_idle = pci_pm_runtime_idle, | |
bbb44d9f RW |
1402 | }; |
1403 | ||
adf09493 | 1404 | #define PCI_PM_OPS_PTR (&pci_dev_pm_ops) |
bbb44d9f | 1405 | |
fbb988be RW |
1406 | #else /* !CONFIG_PM */ |
1407 | ||
1408 | #define pci_pm_runtime_suspend NULL | |
1409 | #define pci_pm_runtime_resume NULL | |
1410 | #define pci_pm_runtime_idle NULL | |
bbb44d9f RW |
1411 | |
1412 | #define PCI_PM_OPS_PTR NULL | |
1413 | ||
fbb988be | 1414 | #endif /* !CONFIG_PM */ |
bbb44d9f | 1415 | |
1da177e4 | 1416 | /** |
863b18f4 | 1417 | * __pci_register_driver - register a new pci driver |
1da177e4 | 1418 | * @drv: the driver structure to register |
863b18f4 | 1419 | * @owner: owner module of drv |
f95d882d | 1420 | * @mod_name: module name string |
f7625980 | 1421 | * |
1da177e4 | 1422 | * Adds the driver structure to the list of registered drivers. |
f7625980 BH |
1423 | * Returns a negative value on error, otherwise 0. |
1424 | * If no error occurred, the driver remains registered even if | |
1da177e4 LT |
1425 | * no device was claimed during registration. |
1426 | */ | |
725522b5 GKH |
1427 | int __pci_register_driver(struct pci_driver *drv, struct module *owner, |
1428 | const char *mod_name) | |
1da177e4 | 1429 | { |
1da177e4 LT |
1430 | /* initialize common driver fields */ |
1431 | drv->driver.name = drv->name; | |
1432 | drv->driver.bus = &pci_bus_type; | |
863b18f4 | 1433 | drv->driver.owner = owner; |
725522b5 | 1434 | drv->driver.mod_name = mod_name; |
92d50fc1 | 1435 | drv->driver.groups = drv->groups; |
ded13b9c | 1436 | drv->driver.dev_groups = drv->dev_groups; |
50b00755 | 1437 | |
75865858 GKH |
1438 | spin_lock_init(&drv->dynids.lock); |
1439 | INIT_LIST_HEAD(&drv->dynids.list); | |
1da177e4 LT |
1440 | |
1441 | /* register with core */ | |
bfb09a86 | 1442 | return driver_register(&drv->driver); |
1da177e4 | 1443 | } |
b7fe9434 | 1444 | EXPORT_SYMBOL(__pci_register_driver); |
1da177e4 LT |
1445 | |
1446 | /** | |
1447 | * pci_unregister_driver - unregister a pci driver | |
1448 | * @drv: the driver structure to unregister | |
f7625980 | 1449 | * |
1da177e4 LT |
1450 | * Deletes the driver structure from the list of registered PCI drivers, |
1451 | * gives it a chance to clean up by calling its remove() function for | |
1452 | * each device it was responsible for, and marks those devices as | |
1453 | * driverless. | |
1454 | */ | |
1455 | ||
3c78bc61 | 1456 | void pci_unregister_driver(struct pci_driver *drv) |
1da177e4 LT |
1457 | { |
1458 | driver_unregister(&drv->driver); | |
1459 | pci_free_dynids(drv); | |
1460 | } | |
b7fe9434 | 1461 | EXPORT_SYMBOL(pci_unregister_driver); |
1da177e4 LT |
1462 | |
1463 | static struct pci_driver pci_compat_driver = { | |
1464 | .name = "compat" | |
1465 | }; | |
1466 | ||
1467 | /** | |
1468 | * pci_dev_driver - get the pci_driver of a device | |
1469 | * @dev: the device to query | |
1470 | * | |
f7625980 | 1471 | * Returns the appropriate pci_driver structure or %NULL if there is no |
1da177e4 LT |
1472 | * registered driver for the device. |
1473 | */ | |
3c78bc61 | 1474 | struct pci_driver *pci_dev_driver(const struct pci_dev *dev) |
1da177e4 | 1475 | { |
e0217c5b BH |
1476 | if (dev->driver) |
1477 | return dev->driver; | |
1da177e4 LT |
1478 | else { |
1479 | int i; | |
3c78bc61 | 1480 | for (i = 0; i <= PCI_ROM_RESOURCE; i++) |
1da177e4 LT |
1481 | if (dev->resource[i].flags & IORESOURCE_BUSY) |
1482 | return &pci_compat_driver; | |
1483 | } | |
1484 | return NULL; | |
1485 | } | |
b7fe9434 | 1486 | EXPORT_SYMBOL(pci_dev_driver); |
1da177e4 LT |
1487 | |
1488 | /** | |
1489 | * pci_bus_match - Tell if a PCI device structure has a matching PCI device id structure | |
1da177e4 | 1490 | * @dev: the PCI device structure to match against |
8f7020d3 | 1491 | * @drv: the device driver to search for matching PCI device id structures |
f7625980 | 1492 | * |
1da177e4 | 1493 | * Used by a driver to check whether a PCI device present in the |
8f7020d3 | 1494 | * system is in its list of supported devices. Returns the matching |
1da177e4 LT |
1495 | * pci_device_id structure or %NULL if there is no match. |
1496 | */ | |
75865858 | 1497 | static int pci_bus_match(struct device *dev, struct device_driver *drv) |
1da177e4 | 1498 | { |
75865858 | 1499 | struct pci_dev *pci_dev = to_pci_dev(dev); |
58d9a38f | 1500 | struct pci_driver *pci_drv; |
1da177e4 LT |
1501 | const struct pci_device_id *found_id; |
1502 | ||
58d9a38f YL |
1503 | if (!pci_dev->match_driver) |
1504 | return 0; | |
1505 | ||
1506 | pci_drv = to_pci_driver(drv); | |
75865858 | 1507 | found_id = pci_match_device(pci_drv, pci_dev); |
1da177e4 LT |
1508 | if (found_id) |
1509 | return 1; | |
1510 | ||
75865858 | 1511 | return 0; |
1da177e4 LT |
1512 | } |
1513 | ||
1514 | /** | |
1515 | * pci_dev_get - increments the reference count of the pci device structure | |
1516 | * @dev: the device being referenced | |
1517 | * | |
1518 | * Each live reference to a device should be refcounted. | |
1519 | * | |
1520 | * Drivers for PCI devices should normally record such references in | |
1521 | * their probe() methods, when they bind to a device, and release | |
1522 | * them by calling pci_dev_put(), in their disconnect() methods. | |
1523 | * | |
1524 | * A pointer to the device with the incremented reference counter is returned. | |
1525 | */ | |
1526 | struct pci_dev *pci_dev_get(struct pci_dev *dev) | |
1527 | { | |
1528 | if (dev) | |
1529 | get_device(&dev->dev); | |
1530 | return dev; | |
1531 | } | |
b7fe9434 | 1532 | EXPORT_SYMBOL(pci_dev_get); |
1da177e4 LT |
1533 | |
1534 | /** | |
1535 | * pci_dev_put - release a use of the pci device structure | |
1536 | * @dev: device that's been disconnected | |
1537 | * | |
1538 | * Must be called when a user of a device is finished with it. When the last | |
1539 | * user of the device calls this function, the memory of the device is freed. | |
1540 | */ | |
1541 | void pci_dev_put(struct pci_dev *dev) | |
1542 | { | |
1543 | if (dev) | |
1544 | put_device(&dev->dev); | |
1545 | } | |
b7fe9434 | 1546 | EXPORT_SYMBOL(pci_dev_put); |
1da177e4 | 1547 | |
8ccc9aa1 BP |
1548 | static int pci_uevent(struct device *dev, struct kobj_uevent_env *env) |
1549 | { | |
1550 | struct pci_dev *pdev; | |
1551 | ||
1552 | if (!dev) | |
1553 | return -ENODEV; | |
1554 | ||
1555 | pdev = to_pci_dev(dev); | |
8ccc9aa1 BP |
1556 | |
1557 | if (add_uevent_var(env, "PCI_CLASS=%04X", pdev->class)) | |
1558 | return -ENOMEM; | |
1559 | ||
1560 | if (add_uevent_var(env, "PCI_ID=%04X:%04X", pdev->vendor, pdev->device)) | |
1561 | return -ENOMEM; | |
1562 | ||
1563 | if (add_uevent_var(env, "PCI_SUBSYS_ID=%04X:%04X", pdev->subsystem_vendor, | |
1564 | pdev->subsystem_device)) | |
1565 | return -ENOMEM; | |
1566 | ||
1567 | if (add_uevent_var(env, "PCI_SLOT_NAME=%s", pci_name(pdev))) | |
1568 | return -ENOMEM; | |
1569 | ||
145b3fe5 | 1570 | if (add_uevent_var(env, "MODALIAS=pci:v%08Xd%08Xsv%08Xsd%08Xbc%02Xsc%02Xi%02X", |
8ccc9aa1 BP |
1571 | pdev->vendor, pdev->device, |
1572 | pdev->subsystem_vendor, pdev->subsystem_device, | |
1573 | (u8)(pdev->class >> 16), (u8)(pdev->class >> 8), | |
1574 | (u8)(pdev->class))) | |
1575 | return -ENOMEM; | |
efdd4070 | 1576 | |
8ccc9aa1 BP |
1577 | return 0; |
1578 | } | |
1579 | ||
f9a6c8ad | 1580 | #if defined(CONFIG_PCIEAER) || defined(CONFIG_EEH) |
3ecac020 ME |
1581 | /** |
1582 | * pci_uevent_ers - emit a uevent during recovery path of PCI device | |
1583 | * @pdev: PCI device undergoing error recovery | |
1584 | * @err_type: type of error event | |
1585 | */ | |
1586 | void pci_uevent_ers(struct pci_dev *pdev, enum pci_ers_result err_type) | |
1587 | { | |
1588 | int idx = 0; | |
1589 | char *envp[3]; | |
1590 | ||
1591 | switch (err_type) { | |
1592 | case PCI_ERS_RESULT_NONE: | |
1593 | case PCI_ERS_RESULT_CAN_RECOVER: | |
1594 | envp[idx++] = "ERROR_EVENT=BEGIN_RECOVERY"; | |
1595 | envp[idx++] = "DEVICE_ONLINE=0"; | |
1596 | break; | |
1597 | case PCI_ERS_RESULT_RECOVERED: | |
1598 | envp[idx++] = "ERROR_EVENT=SUCCESSFUL_RECOVERY"; | |
1599 | envp[idx++] = "DEVICE_ONLINE=1"; | |
1600 | break; | |
1601 | case PCI_ERS_RESULT_DISCONNECT: | |
1602 | envp[idx++] = "ERROR_EVENT=FAILED_RECOVERY"; | |
1603 | envp[idx++] = "DEVICE_ONLINE=0"; | |
1604 | break; | |
1605 | default: | |
1606 | break; | |
1607 | } | |
1608 | ||
1609 | if (idx > 0) { | |
1610 | envp[idx++] = NULL; | |
1611 | kobject_uevent_env(&pdev->dev.kobj, KOBJ_CHANGE, envp); | |
1612 | } | |
1613 | } | |
1614 | #endif | |
1615 | ||
02e0bea6 PS |
1616 | static int pci_bus_num_vf(struct device *dev) |
1617 | { | |
1618 | return pci_num_vf(to_pci_dev(dev)); | |
1619 | } | |
1620 | ||
07397df2 NG |
1621 | /** |
1622 | * pci_dma_configure - Setup DMA configuration | |
1623 | * @dev: ptr to dev structure | |
1624 | * | |
1625 | * Function to update PCI devices's DMA configuration using the same | |
1626 | * info from the OF node or ACPI node of host bridge's parent (if any). | |
1627 | */ | |
1628 | static int pci_dma_configure(struct device *dev) | |
1629 | { | |
512881ea | 1630 | struct pci_driver *driver = to_pci_driver(dev->driver); |
07397df2 NG |
1631 | struct device *bridge; |
1632 | int ret = 0; | |
1633 | ||
1634 | bridge = pci_get_host_bridge_device(to_pci_dev(dev)); | |
1635 | ||
1636 | if (IS_ENABLED(CONFIG_OF) && bridge->parent && | |
1637 | bridge->parent->of_node) { | |
3d6ce86e | 1638 | ret = of_dma_configure(dev, bridge->parent->of_node, true); |
07397df2 NG |
1639 | } else if (has_acpi_companion(bridge)) { |
1640 | struct acpi_device *adev = to_acpi_device_node(bridge->fwnode); | |
07397df2 | 1641 | |
e5361ca2 | 1642 | ret = acpi_dma_configure(dev, acpi_get_dma_attr(adev)); |
07397df2 NG |
1643 | } |
1644 | ||
1645 | pci_put_host_bridge_device(bridge); | |
512881ea LB |
1646 | |
1647 | if (!ret && !driver->driver_managed_dma) { | |
1648 | ret = iommu_device_use_default_domain(dev); | |
1649 | if (ret) | |
1650 | arch_teardown_dma_ops(dev); | |
1651 | } | |
1652 | ||
07397df2 NG |
1653 | return ret; |
1654 | } | |
1655 | ||
512881ea LB |
1656 | static void pci_dma_cleanup(struct device *dev) |
1657 | { | |
1658 | struct pci_driver *driver = to_pci_driver(dev->driver); | |
1659 | ||
1660 | if (!driver->driver_managed_dma) | |
1661 | iommu_device_unuse_default_domain(dev); | |
1662 | } | |
1663 | ||
1da177e4 LT |
1664 | struct bus_type pci_bus_type = { |
1665 | .name = "pci", | |
1666 | .match = pci_bus_match, | |
312c004d | 1667 | .uevent = pci_uevent, |
b15d686a RK |
1668 | .probe = pci_device_probe, |
1669 | .remove = pci_device_remove, | |
cbd69dbb | 1670 | .shutdown = pci_device_shutdown, |
5136b2da | 1671 | .dev_groups = pci_dev_groups, |
0f49ba55 | 1672 | .bus_groups = pci_bus_groups, |
2229c1fb | 1673 | .drv_groups = pci_drv_groups, |
bbb44d9f | 1674 | .pm = PCI_PM_OPS_PTR, |
02e0bea6 | 1675 | .num_vf = pci_bus_num_vf, |
07397df2 | 1676 | .dma_configure = pci_dma_configure, |
512881ea | 1677 | .dma_cleanup = pci_dma_cleanup, |
1da177e4 | 1678 | }; |
b7fe9434 | 1679 | EXPORT_SYMBOL(pci_bus_type); |
1da177e4 | 1680 | |
c6c889d9 BH |
1681 | #ifdef CONFIG_PCIEPORTBUS |
1682 | static int pcie_port_bus_match(struct device *dev, struct device_driver *drv) | |
1683 | { | |
1684 | struct pcie_device *pciedev; | |
1685 | struct pcie_port_service_driver *driver; | |
1686 | ||
1687 | if (drv->bus != &pcie_port_bus_type || dev->bus != &pcie_port_bus_type) | |
1688 | return 0; | |
1689 | ||
1690 | pciedev = to_pcie_device(dev); | |
1691 | driver = to_service_driver(drv); | |
1692 | ||
1693 | if (driver->service != pciedev->service) | |
1694 | return 0; | |
1695 | ||
1696 | if (driver->port_type != PCIE_ANY_PORT && | |
1697 | driver->port_type != pci_pcie_type(pciedev->port)) | |
1698 | return 0; | |
1699 | ||
1700 | return 1; | |
1701 | } | |
1702 | ||
1703 | struct bus_type pcie_port_bus_type = { | |
1704 | .name = "pci_express", | |
1705 | .match = pcie_port_bus_match, | |
1706 | }; | |
1707 | EXPORT_SYMBOL_GPL(pcie_port_bus_type); | |
1708 | #endif | |
1709 | ||
1da177e4 LT |
1710 | static int __init pci_driver_init(void) |
1711 | { | |
c6c889d9 BH |
1712 | int ret; |
1713 | ||
1714 | ret = bus_register(&pci_bus_type); | |
1715 | if (ret) | |
1716 | return ret; | |
1717 | ||
1718 | #ifdef CONFIG_PCIEPORTBUS | |
1719 | ret = bus_register(&pcie_port_bus_type); | |
1720 | if (ret) | |
1721 | return ret; | |
1722 | #endif | |
a8651194 | 1723 | dma_debug_add_bus(&pci_bus_type); |
c6c889d9 | 1724 | return 0; |
1da177e4 | 1725 | } |
1da177e4 | 1726 | postcore_initcall(pci_driver_init); |