Merge tag 'scsi-fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/jejb/scsi
[linux-2.6-block.git] / drivers / base / firmware_loader / fallback.c
CommitLineData
d73f821c
LR
1// SPDX-License-Identifier: GPL-2.0
2
3#include <linux/types.h>
4#include <linux/kconfig.h>
5#include <linux/list.h>
6#include <linux/slab.h>
7#include <linux/security.h>
8#include <linux/highmem.h>
9#include <linux/umh.h>
ceb18132 10#include <linux/sysctl.h>
ccce305b 11#include <linux/vmalloc.h>
d73f821c 12
5d6d1ddd
LR
13#include "fallback.h"
14#include "firmware.h"
d73f821c
LR
15
16/*
17 * firmware fallback mechanism
18 */
19
20extern struct firmware_fallback_config fw_fallback_config;
21
22/* These getters are vetted to use int properly */
23static inline int __firmware_loading_timeout(void)
24{
25 return fw_fallback_config.loading_timeout;
26}
27
28/* These setters are vetted to use int properly */
29static void __fw_fallback_set_timeout(int timeout)
30{
31 fw_fallback_config.loading_timeout = timeout;
32}
33
34/*
35 * use small loading timeout for caching devices' firmware because all these
36 * firmware images have been loaded successfully at lease once, also system is
37 * ready for completing firmware loading now. The maximum size of firmware in
38 * current distributions is about 2M bytes, so 10 secs should be enough.
39 */
40void fw_fallback_set_cache_timeout(void)
41{
42 fw_fallback_config.old_timeout = __firmware_loading_timeout();
43 __fw_fallback_set_timeout(10);
44}
45
46/* Restores the timeout to the value last configured during normal operation */
47void fw_fallback_set_default_timeout(void)
48{
49 __fw_fallback_set_timeout(fw_fallback_config.old_timeout);
50}
51
52static long firmware_loading_timeout(void)
53{
54 return __firmware_loading_timeout() > 0 ?
55 __firmware_loading_timeout() * HZ : MAX_JIFFY_OFFSET;
56}
57
58static inline bool fw_sysfs_done(struct fw_priv *fw_priv)
59{
60 return __fw_state_check(fw_priv, FW_STATUS_DONE);
61}
62
63static inline bool fw_sysfs_loading(struct fw_priv *fw_priv)
64{
65 return __fw_state_check(fw_priv, FW_STATUS_LOADING);
66}
67
68static inline int fw_sysfs_wait_timeout(struct fw_priv *fw_priv, long timeout)
69{
70 return __fw_state_wait_common(fw_priv, timeout);
71}
72
73struct fw_sysfs {
74 bool nowait;
75 struct device dev;
76 struct fw_priv *fw_priv;
77 struct firmware *fw;
78};
79
80static struct fw_sysfs *to_fw_sysfs(struct device *dev)
81{
82 return container_of(dev, struct fw_sysfs, dev);
83}
84
85static void __fw_load_abort(struct fw_priv *fw_priv)
86{
87 /*
88 * There is a small window in which user can write to 'loading'
89 * between loading done and disappearance of 'loading'
90 */
91 if (fw_sysfs_done(fw_priv))
92 return;
93
94 list_del_init(&fw_priv->pending_list);
95 fw_state_aborted(fw_priv);
96}
97
98static void fw_load_abort(struct fw_sysfs *fw_sysfs)
99{
100 struct fw_priv *fw_priv = fw_sysfs->fw_priv;
101
102 __fw_load_abort(fw_priv);
103}
104
105static LIST_HEAD(pending_fw_head);
106
107void kill_pending_fw_fallback_reqs(bool only_kill_custom)
108{
109 struct fw_priv *fw_priv;
110 struct fw_priv *next;
111
112 mutex_lock(&fw_lock);
113 list_for_each_entry_safe(fw_priv, next, &pending_fw_head,
114 pending_list) {
115 if (!fw_priv->need_uevent || !only_kill_custom)
116 __fw_load_abort(fw_priv);
117 }
118 mutex_unlock(&fw_lock);
119}
120
121static ssize_t timeout_show(struct class *class, struct class_attribute *attr,
122 char *buf)
123{
124 return sprintf(buf, "%d\n", __firmware_loading_timeout());
125}
126
127/**
128 * firmware_timeout_store - set number of seconds to wait for firmware
129 * @class: device class pointer
130 * @attr: device attribute pointer
131 * @buf: buffer to scan for timeout value
132 * @count: number of bytes in @buf
133 *
134 * Sets the number of seconds to wait for the firmware. Once
135 * this expires an error will be returned to the driver and no
136 * firmware will be provided.
137 *
138 * Note: zero means 'wait forever'.
139 **/
140static ssize_t timeout_store(struct class *class, struct class_attribute *attr,
141 const char *buf, size_t count)
142{
143 int tmp_loading_timeout = simple_strtol(buf, NULL, 10);
144
145 if (tmp_loading_timeout < 0)
146 tmp_loading_timeout = 0;
147
148 __fw_fallback_set_timeout(tmp_loading_timeout);
149
150 return count;
151}
152static CLASS_ATTR_RW(timeout);
153
154static struct attribute *firmware_class_attrs[] = {
155 &class_attr_timeout.attr,
156 NULL,
157};
158ATTRIBUTE_GROUPS(firmware_class);
159
160static void fw_dev_release(struct device *dev)
161{
162 struct fw_sysfs *fw_sysfs = to_fw_sysfs(dev);
163
164 kfree(fw_sysfs);
165}
166
167static int do_firmware_uevent(struct fw_sysfs *fw_sysfs, struct kobj_uevent_env *env)
168{
169 if (add_uevent_var(env, "FIRMWARE=%s", fw_sysfs->fw_priv->fw_name))
170 return -ENOMEM;
171 if (add_uevent_var(env, "TIMEOUT=%i", __firmware_loading_timeout()))
172 return -ENOMEM;
173 if (add_uevent_var(env, "ASYNC=%d", fw_sysfs->nowait))
174 return -ENOMEM;
175
176 return 0;
177}
178
179static int firmware_uevent(struct device *dev, struct kobj_uevent_env *env)
180{
181 struct fw_sysfs *fw_sysfs = to_fw_sysfs(dev);
182 int err = 0;
183
184 mutex_lock(&fw_lock);
185 if (fw_sysfs->fw_priv)
186 err = do_firmware_uevent(fw_sysfs, env);
187 mutex_unlock(&fw_lock);
188 return err;
189}
190
191static struct class firmware_class = {
192 .name = "firmware",
193 .class_groups = firmware_class_groups,
194 .dev_uevent = firmware_uevent,
195 .dev_release = fw_dev_release,
196};
197
198int register_sysfs_loader(void)
199{
200 return class_register(&firmware_class);
201}
202
203void unregister_sysfs_loader(void)
204{
205 class_unregister(&firmware_class);
206}
207
208static ssize_t firmware_loading_show(struct device *dev,
209 struct device_attribute *attr, char *buf)
210{
211 struct fw_sysfs *fw_sysfs = to_fw_sysfs(dev);
212 int loading = 0;
213
214 mutex_lock(&fw_lock);
215 if (fw_sysfs->fw_priv)
216 loading = fw_sysfs_loading(fw_sysfs->fw_priv);
217 mutex_unlock(&fw_lock);
218
219 return sprintf(buf, "%d\n", loading);
220}
221
222/* Some architectures don't have PAGE_KERNEL_RO */
223#ifndef PAGE_KERNEL_RO
224#define PAGE_KERNEL_RO PAGE_KERNEL
225#endif
226
227/* one pages buffer should be mapped/unmapped only once */
228static int map_fw_priv_pages(struct fw_priv *fw_priv)
229{
230 if (!fw_priv->is_paged_buf)
231 return 0;
232
233 vunmap(fw_priv->data);
234 fw_priv->data = vmap(fw_priv->pages, fw_priv->nr_pages, 0,
235 PAGE_KERNEL_RO);
236 if (!fw_priv->data)
237 return -ENOMEM;
238 return 0;
239}
240
241/**
242 * firmware_loading_store - set value in the 'loading' control file
243 * @dev: device pointer
244 * @attr: device attribute pointer
245 * @buf: buffer to scan for loading control value
246 * @count: number of bytes in @buf
247 *
248 * The relevant values are:
249 *
250 * 1: Start a load, discarding any previous partial load.
251 * 0: Conclude the load and hand the data to the driver code.
252 * -1: Conclude the load with an error and discard any written data.
253 **/
254static ssize_t firmware_loading_store(struct device *dev,
255 struct device_attribute *attr,
256 const char *buf, size_t count)
257{
258 struct fw_sysfs *fw_sysfs = to_fw_sysfs(dev);
259 struct fw_priv *fw_priv;
260 ssize_t written = count;
261 int loading = simple_strtol(buf, NULL, 10);
262 int i;
263
264 mutex_lock(&fw_lock);
265 fw_priv = fw_sysfs->fw_priv;
266 if (fw_state_is_aborted(fw_priv))
267 goto out;
268
269 switch (loading) {
270 case 1:
271 /* discarding any previous partial load */
272 if (!fw_sysfs_done(fw_priv)) {
273 for (i = 0; i < fw_priv->nr_pages; i++)
274 __free_page(fw_priv->pages[i]);
275 vfree(fw_priv->pages);
276 fw_priv->pages = NULL;
277 fw_priv->page_array_size = 0;
278 fw_priv->nr_pages = 0;
279 fw_state_start(fw_priv);
280 }
281 break;
282 case 0:
283 if (fw_sysfs_loading(fw_priv)) {
284 int rc;
285
286 /*
287 * Several loading requests may be pending on
288 * one same firmware buf, so let all requests
289 * see the mapped 'buf->data' once the loading
290 * is completed.
291 * */
292 rc = map_fw_priv_pages(fw_priv);
293 if (rc)
294 dev_err(dev, "%s: map pages failed\n",
295 __func__);
296 else
297 rc = security_kernel_post_read_file(NULL,
298 fw_priv->data, fw_priv->size,
299 READING_FIRMWARE);
300
301 /*
302 * Same logic as fw_load_abort, only the DONE bit
303 * is ignored and we set ABORT only on failure.
304 */
305 list_del_init(&fw_priv->pending_list);
306 if (rc) {
307 fw_state_aborted(fw_priv);
308 written = rc;
309 } else {
310 fw_state_done(fw_priv);
311 }
312 break;
313 }
314 /* fallthrough */
315 default:
316 dev_err(dev, "%s: unexpected value (%d)\n", __func__, loading);
317 /* fallthrough */
318 case -1:
319 fw_load_abort(fw_sysfs);
320 break;
321 }
322out:
323 mutex_unlock(&fw_lock);
324 return written;
325}
326
327static DEVICE_ATTR(loading, 0644, firmware_loading_show, firmware_loading_store);
328
329static void firmware_rw_data(struct fw_priv *fw_priv, char *buffer,
330 loff_t offset, size_t count, bool read)
331{
332 if (read)
333 memcpy(buffer, fw_priv->data + offset, count);
334 else
335 memcpy(fw_priv->data + offset, buffer, count);
336}
337
338static void firmware_rw(struct fw_priv *fw_priv, char *buffer,
339 loff_t offset, size_t count, bool read)
340{
341 while (count) {
342 void *page_data;
343 int page_nr = offset >> PAGE_SHIFT;
344 int page_ofs = offset & (PAGE_SIZE-1);
345 int page_cnt = min_t(size_t, PAGE_SIZE - page_ofs, count);
346
347 page_data = kmap(fw_priv->pages[page_nr]);
348
349 if (read)
350 memcpy(buffer, page_data + page_ofs, page_cnt);
351 else
352 memcpy(page_data + page_ofs, buffer, page_cnt);
353
354 kunmap(fw_priv->pages[page_nr]);
355 buffer += page_cnt;
356 offset += page_cnt;
357 count -= page_cnt;
358 }
359}
360
361static ssize_t firmware_data_read(struct file *filp, struct kobject *kobj,
362 struct bin_attribute *bin_attr,
363 char *buffer, loff_t offset, size_t count)
364{
365 struct device *dev = kobj_to_dev(kobj);
366 struct fw_sysfs *fw_sysfs = to_fw_sysfs(dev);
367 struct fw_priv *fw_priv;
368 ssize_t ret_count;
369
370 mutex_lock(&fw_lock);
371 fw_priv = fw_sysfs->fw_priv;
372 if (!fw_priv || fw_sysfs_done(fw_priv)) {
373 ret_count = -ENODEV;
374 goto out;
375 }
376 if (offset > fw_priv->size) {
377 ret_count = 0;
378 goto out;
379 }
380 if (count > fw_priv->size - offset)
381 count = fw_priv->size - offset;
382
383 ret_count = count;
384
385 if (fw_priv->data)
386 firmware_rw_data(fw_priv, buffer, offset, count, true);
387 else
388 firmware_rw(fw_priv, buffer, offset, count, true);
389
390out:
391 mutex_unlock(&fw_lock);
392 return ret_count;
393}
394
395static int fw_realloc_pages(struct fw_sysfs *fw_sysfs, int min_size)
396{
397 struct fw_priv *fw_priv= fw_sysfs->fw_priv;
398 int pages_needed = PAGE_ALIGN(min_size) >> PAGE_SHIFT;
399
400 /* If the array of pages is too small, grow it... */
401 if (fw_priv->page_array_size < pages_needed) {
402 int new_array_size = max(pages_needed,
403 fw_priv->page_array_size * 2);
404 struct page **new_pages;
405
406 new_pages = vmalloc(new_array_size * sizeof(void *));
407 if (!new_pages) {
408 fw_load_abort(fw_sysfs);
409 return -ENOMEM;
410 }
411 memcpy(new_pages, fw_priv->pages,
412 fw_priv->page_array_size * sizeof(void *));
413 memset(&new_pages[fw_priv->page_array_size], 0, sizeof(void *) *
414 (new_array_size - fw_priv->page_array_size));
415 vfree(fw_priv->pages);
416 fw_priv->pages = new_pages;
417 fw_priv->page_array_size = new_array_size;
418 }
419
420 while (fw_priv->nr_pages < pages_needed) {
421 fw_priv->pages[fw_priv->nr_pages] =
422 alloc_page(GFP_KERNEL | __GFP_HIGHMEM);
423
424 if (!fw_priv->pages[fw_priv->nr_pages]) {
425 fw_load_abort(fw_sysfs);
426 return -ENOMEM;
427 }
428 fw_priv->nr_pages++;
429 }
430 return 0;
431}
432
433/**
434 * firmware_data_write - write method for firmware
435 * @filp: open sysfs file
436 * @kobj: kobject for the device
437 * @bin_attr: bin_attr structure
438 * @buffer: buffer being written
439 * @offset: buffer offset for write in total data store area
440 * @count: buffer size
441 *
442 * Data written to the 'data' attribute will be later handed to
443 * the driver as a firmware image.
444 **/
445static ssize_t firmware_data_write(struct file *filp, struct kobject *kobj,
446 struct bin_attribute *bin_attr,
447 char *buffer, loff_t offset, size_t count)
448{
449 struct device *dev = kobj_to_dev(kobj);
450 struct fw_sysfs *fw_sysfs = to_fw_sysfs(dev);
451 struct fw_priv *fw_priv;
452 ssize_t retval;
453
454 if (!capable(CAP_SYS_RAWIO))
455 return -EPERM;
456
457 mutex_lock(&fw_lock);
458 fw_priv = fw_sysfs->fw_priv;
459 if (!fw_priv || fw_sysfs_done(fw_priv)) {
460 retval = -ENODEV;
461 goto out;
462 }
463
464 if (fw_priv->data) {
465 if (offset + count > fw_priv->allocated_size) {
466 retval = -ENOMEM;
467 goto out;
468 }
469 firmware_rw_data(fw_priv, buffer, offset, count, false);
470 retval = count;
471 } else {
472 retval = fw_realloc_pages(fw_sysfs, offset + count);
473 if (retval)
474 goto out;
475
476 retval = count;
477 firmware_rw(fw_priv, buffer, offset, count, false);
478 }
479
480 fw_priv->size = max_t(size_t, offset + count, fw_priv->size);
481out:
482 mutex_unlock(&fw_lock);
483 return retval;
484}
485
486static struct bin_attribute firmware_attr_data = {
487 .attr = { .name = "data", .mode = 0644 },
488 .size = 0,
489 .read = firmware_data_read,
490 .write = firmware_data_write,
491};
492
493static struct attribute *fw_dev_attrs[] = {
494 &dev_attr_loading.attr,
495 NULL
496};
497
498static struct bin_attribute *fw_dev_bin_attrs[] = {
499 &firmware_attr_data,
500 NULL
501};
502
503static const struct attribute_group fw_dev_attr_group = {
504 .attrs = fw_dev_attrs,
505 .bin_attrs = fw_dev_bin_attrs,
506};
507
508static const struct attribute_group *fw_dev_attr_groups[] = {
509 &fw_dev_attr_group,
510 NULL
511};
512
513static struct fw_sysfs *
514fw_create_instance(struct firmware *firmware, const char *fw_name,
515 struct device *device, unsigned int opt_flags)
516{
517 struct fw_sysfs *fw_sysfs;
518 struct device *f_dev;
519
520 fw_sysfs = kzalloc(sizeof(*fw_sysfs), GFP_KERNEL);
521 if (!fw_sysfs) {
522 fw_sysfs = ERR_PTR(-ENOMEM);
523 goto exit;
524 }
525
526 fw_sysfs->nowait = !!(opt_flags & FW_OPT_NOWAIT);
527 fw_sysfs->fw = firmware;
528 f_dev = &fw_sysfs->dev;
529
530 device_initialize(f_dev);
531 dev_set_name(f_dev, "%s", fw_name);
532 f_dev->parent = device;
533 f_dev->class = &firmware_class;
534 f_dev->groups = fw_dev_attr_groups;
535exit:
536 return fw_sysfs;
537}
538
60fa7426 539/**
b93815d0
AR
540 * fw_load_sysfs_fallback - load a firmware via the sysfs fallback mechanism
541 * @fw_sysfs: firmware sysfs information for the firmware to load
60fa7426
LR
542 * @opt_flags: flags of options, FW_OPT_*
543 * @timeout: timeout to wait for the load
544 *
545 * In charge of constructing a sysfs fallback interface for firmware loading.
546 **/
547static int fw_load_sysfs_fallback(struct fw_sysfs *fw_sysfs,
d73f821c
LR
548 unsigned int opt_flags, long timeout)
549{
550 int retval = 0;
551 struct device *f_dev = &fw_sysfs->dev;
552 struct fw_priv *fw_priv = fw_sysfs->fw_priv;
553
554 /* fall back on userspace loading */
555 if (!fw_priv->data)
556 fw_priv->is_paged_buf = true;
557
558 dev_set_uevent_suppress(f_dev, true);
559
560 retval = device_add(f_dev);
561 if (retval) {
562 dev_err(f_dev, "%s: device_register failed\n", __func__);
563 goto err_put_dev;
564 }
565
566 mutex_lock(&fw_lock);
567 list_add(&fw_priv->pending_list, &pending_fw_head);
568 mutex_unlock(&fw_lock);
569
570 if (opt_flags & FW_OPT_UEVENT) {
571 fw_priv->need_uevent = true;
572 dev_set_uevent_suppress(f_dev, false);
573 dev_dbg(f_dev, "firmware: requesting %s\n", fw_priv->fw_name);
574 kobject_uevent(&fw_sysfs->dev.kobj, KOBJ_ADD);
575 } else {
576 timeout = MAX_JIFFY_OFFSET;
577 }
578
579 retval = fw_sysfs_wait_timeout(fw_priv, timeout);
580 if (retval < 0) {
581 mutex_lock(&fw_lock);
582 fw_load_abort(fw_sysfs);
583 mutex_unlock(&fw_lock);
584 }
585
586 if (fw_state_is_aborted(fw_priv)) {
587 if (retval == -ERESTARTSYS)
588 retval = -EINTR;
589 else
590 retval = -EAGAIN;
591 } else if (fw_priv->is_paged_buf && !fw_priv->data)
592 retval = -ENOMEM;
593
594 device_del(f_dev);
595err_put_dev:
596 put_device(f_dev);
597 return retval;
598}
599
600static int fw_load_from_user_helper(struct firmware *firmware,
601 const char *name, struct device *device,
602 unsigned int opt_flags)
603{
604 struct fw_sysfs *fw_sysfs;
605 long timeout;
606 int ret;
607
608 timeout = firmware_loading_timeout();
609 if (opt_flags & FW_OPT_NOWAIT) {
610 timeout = usermodehelper_read_lock_wait(timeout);
611 if (!timeout) {
612 dev_dbg(device, "firmware: %s loading timed out\n",
613 name);
614 return -EBUSY;
615 }
616 } else {
617 ret = usermodehelper_read_trylock();
618 if (WARN_ON(ret)) {
619 dev_err(device, "firmware: %s will not be loaded\n",
620 name);
621 return ret;
622 }
623 }
624
625 fw_sysfs = fw_create_instance(firmware, name, device, opt_flags);
626 if (IS_ERR(fw_sysfs)) {
627 ret = PTR_ERR(fw_sysfs);
628 goto out_unlock;
629 }
630
631 fw_sysfs->fw_priv = firmware->priv;
60fa7426 632 ret = fw_load_sysfs_fallback(fw_sysfs, opt_flags, timeout);
d73f821c
LR
633
634 if (!ret)
635 ret = assign_fw(firmware, device, opt_flags);
636
637out_unlock:
638 usermodehelper_read_unlock();
639
640 return ret;
641}
642
643static bool fw_force_sysfs_fallback(unsigned int opt_flags)
644{
645 if (fw_fallback_config.force_sysfs_fallback)
646 return true;
647 if (!(opt_flags & FW_OPT_USERHELPER))
648 return false;
649 return true;
650}
651
652static bool fw_run_sysfs_fallback(unsigned int opt_flags)
653{
2cd7a1c6 654 if (fw_fallback_config.ignore_sysfs_fallback) {
c6263a48 655 pr_info_once("Ignoring firmware sysfs fallback due to sysctl knob\n");
2cd7a1c6
LR
656 return false;
657 }
658
d73f821c
LR
659 if ((opt_flags & FW_OPT_NOFALLBACK))
660 return false;
661
662 return fw_force_sysfs_fallback(opt_flags);
663}
664
665int fw_sysfs_fallback(struct firmware *fw, const char *name,
666 struct device *device,
667 unsigned int opt_flags,
668 int ret)
669{
670 if (!fw_run_sysfs_fallback(opt_flags))
671 return ret;
672
673 dev_warn(device, "Falling back to user helper\n");
674 return fw_load_from_user_helper(fw, name, device, opt_flags);
675}