Commit | Line | Data |
---|---|---|
bcb877e4 | 1 | /* |
1da177e4 LT |
2 | * \author Rickard E. (Rik) Faith <faith@valinux.com> |
3 | * \author Daryll Strauss <daryll@valinux.com> | |
4 | * \author Gareth Hughes <gareth@valinux.com> | |
5 | */ | |
6 | ||
7 | /* | |
8 | * Created: Mon Jan 4 08:58:31 1999 by faith@valinux.com | |
9 | * | |
10 | * Copyright 1999 Precision Insight, Inc., Cedar Park, Texas. | |
11 | * Copyright 2000 VA Linux Systems, Inc., Sunnyvale, California. | |
12 | * All Rights Reserved. | |
13 | * | |
14 | * Permission is hereby granted, free of charge, to any person obtaining a | |
15 | * copy of this software and associated documentation files (the "Software"), | |
16 | * to deal in the Software without restriction, including without limitation | |
17 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, | |
18 | * and/or sell copies of the Software, and to permit persons to whom the | |
19 | * Software is furnished to do so, subject to the following conditions: | |
20 | * | |
21 | * The above copyright notice and this permission notice (including the next | |
22 | * paragraph) shall be included in all copies or substantial portions of the | |
23 | * Software. | |
24 | * | |
25 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
26 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
27 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |
28 | * VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR | |
29 | * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, | |
30 | * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR | |
31 | * OTHER DEALINGS IN THE SOFTWARE. | |
32 | */ | |
33 | ||
4748aa16 | 34 | #include <linux/anon_inodes.h> |
0500c04e | 35 | #include <linux/dma-fence.h> |
4748aa16 | 36 | #include <linux/file.h> |
0500c04e SR |
37 | #include <linux/module.h> |
38 | #include <linux/pci.h> | |
1da177e4 | 39 | #include <linux/poll.h> |
5a0e3ad6 | 40 | #include <linux/slab.h> |
a8f8b1d9 | 41 | |
c76f0f7c | 42 | #include <drm/drm_client.h> |
0500c04e | 43 | #include <drm/drm_drv.h> |
a8f8b1d9 | 44 | #include <drm/drm_file.h> |
0500c04e | 45 | #include <drm/drm_print.h> |
a8f8b1d9 | 46 | |
81065548 | 47 | #include "drm_crtc_internal.h" |
0500c04e SR |
48 | #include "drm_internal.h" |
49 | #include "drm_legacy.h" | |
1da177e4 | 50 | |
b1823416 THV |
51 | #if defined(CONFIG_MMU) && defined(CONFIG_TRANSPARENT_HUGEPAGE) |
52 | #include <uapi/asm/mman.h> | |
53 | #include <drm/drm_vma_manager.h> | |
54 | #endif | |
55 | ||
0d639883 | 56 | /* from BKL pushdown */ |
58374713 AB |
57 | DEFINE_MUTEX(drm_global_mutex); |
58 | ||
4017ad7b DV |
59 | bool drm_dev_needs_global_mutex(struct drm_device *dev) |
60 | { | |
61 | /* | |
62 | * Legacy drivers rely on all kinds of BKL locking semantics, don't | |
63 | * bother. They also still need BKL locking for their ioctls, so better | |
64 | * safe than sorry. | |
65 | */ | |
66 | if (drm_core_check_feature(dev, DRIVER_LEGACY)) | |
67 | return true; | |
68 | ||
69 | /* | |
70 | * The deprecated ->load callback must be called after the driver is | |
71 | * already registered. This means such drivers rely on the BKL to make | |
72 | * sure an open can't proceed until the driver is actually fully set up. | |
73 | * Similar hilarity holds for the unload callback. | |
74 | */ | |
75 | if (dev->driver->load || dev->driver->unload) | |
76 | return true; | |
77 | ||
78 | /* | |
79 | * Drivers with the lastclose callback assume that it's synchronized | |
80 | * against concurrent opens, which again needs the BKL. The proper fix | |
81 | * is to use the drm_client infrastructure with proper locking for each | |
82 | * client. | |
83 | */ | |
84 | if (dev->driver->lastclose) | |
85 | return true; | |
86 | ||
87 | return false; | |
88 | } | |
89 | ||
bcb877e4 DV |
90 | /** |
91 | * DOC: file operations | |
92 | * | |
93 | * Drivers must define the file operations structure that forms the DRM | |
94 | * userspace API entry point, even though most of those operations are | |
b93658f8 DV |
95 | * implemented in the DRM core. The resulting &struct file_operations must be |
96 | * stored in the &drm_driver.fops field. The mandatory functions are drm_open(), | |
55edf41b | 97 | * drm_read(), drm_ioctl() and drm_compat_ioctl() if CONFIG_COMPAT is enabled |
b93658f8 DV |
98 | * Note that drm_compat_ioctl will be NULL if CONFIG_COMPAT=n, so there's no |
99 | * need to sprinkle #ifdef into the code. Drivers which implement private ioctls | |
100 | * that require 32/64 bit compatibility support must provide their own | |
101 | * &file_operations.compat_ioctl handler that processes private ioctls and calls | |
102 | * drm_compat_ioctl() for core ioctls. | |
bcb877e4 DV |
103 | * |
104 | * In addition drm_read() and drm_poll() provide support for DRM events. DRM | |
105 | * events are a generic and extensible means to send asynchronous events to | |
106 | * userspace through the file descriptor. They are used to send vblank event and | |
107 | * page flip completions by the KMS API. But drivers can also use it for their | |
108 | * own needs, e.g. to signal completion of rendering. | |
109 | * | |
b93658f8 DV |
110 | * For the driver-side event interface see drm_event_reserve_init() and |
111 | * drm_send_event() as the main starting points. | |
112 | * | |
bcb877e4 DV |
113 | * The memory mapping implementation will vary depending on how the driver |
114 | * manages memory. Legacy drivers will use the deprecated drm_legacy_mmap() | |
115 | * function, modern drivers should use one of the provided memory-manager | |
b93658f8 DV |
116 | * specific implementations. For GEM-based drivers this is drm_gem_mmap(), and |
117 | * for drivers which use the CMA GEM helpers it's drm_gem_cma_mmap(). | |
bcb877e4 DV |
118 | * |
119 | * No other file operations are supported by the DRM userspace API. Overall the | |
bb2eaba6 | 120 | * following is an example &file_operations structure:: |
bcb877e4 DV |
121 | * |
122 | * static const example_drm_fops = { | |
123 | * .owner = THIS_MODULE, | |
124 | * .open = drm_open, | |
125 | * .release = drm_release, | |
126 | * .unlocked_ioctl = drm_ioctl, | |
55edf41b | 127 | * .compat_ioctl = drm_compat_ioctl, // NULL if CONFIG_COMPAT=n |
bcb877e4 DV |
128 | * .poll = drm_poll, |
129 | * .read = drm_read, | |
130 | * .llseek = no_llseek, | |
131 | * .mmap = drm_gem_mmap, | |
132 | * }; | |
b93658f8 | 133 | * |
f42e1819 DV |
134 | * For plain GEM based drivers there is the DEFINE_DRM_GEM_FOPS() macro, and for |
135 | * CMA based drivers there is the DEFINE_DRM_GEM_CMA_FOPS() macro to make this | |
136 | * simpler. | |
bb2eaba6 DV |
137 | * |
138 | * The driver's &file_operations must be stored in &drm_driver.fops. | |
139 | * | |
140 | * For driver-private IOCTL handling see the more detailed discussion in | |
141 | * :ref:`IOCTL support in the userland interfaces chapter<drm_driver_ioctl>`. | |
bcb877e4 DV |
142 | */ |
143 | ||
1572042a DH |
144 | /** |
145 | * drm_file_alloc - allocate file context | |
146 | * @minor: minor to allocate on | |
147 | * | |
148 | * This allocates a new DRM file context. It is not linked into any context and | |
149 | * can be used by the caller freely. Note that the context keeps a pointer to | |
150 | * @minor, so it must be freed before @minor is. | |
151 | * | |
152 | * RETURNS: | |
153 | * Pointer to newly allocated context, ERR_PTR on failure. | |
154 | */ | |
155 | struct drm_file *drm_file_alloc(struct drm_minor *minor) | |
156 | { | |
157 | struct drm_device *dev = minor->dev; | |
158 | struct drm_file *file; | |
159 | int ret; | |
160 | ||
161 | file = kzalloc(sizeof(*file), GFP_KERNEL); | |
162 | if (!file) | |
163 | return ERR_PTR(-ENOMEM); | |
164 | ||
165 | file->pid = get_pid(task_pid(current)); | |
166 | file->minor = minor; | |
167 | ||
168 | /* for compatibility root is always authenticated */ | |
169 | file->authenticated = capable(CAP_SYS_ADMIN); | |
1572042a DH |
170 | |
171 | INIT_LIST_HEAD(&file->lhead); | |
172 | INIT_LIST_HEAD(&file->fbs); | |
173 | mutex_init(&file->fbs_lock); | |
174 | INIT_LIST_HEAD(&file->blobs); | |
175 | INIT_LIST_HEAD(&file->pending_event_list); | |
176 | INIT_LIST_HEAD(&file->event_list); | |
177 | init_waitqueue_head(&file->event_wait); | |
178 | file->event_space = 4096; /* set aside 4k for event buffer */ | |
179 | ||
180 | mutex_init(&file->event_read_lock); | |
181 | ||
182 | if (drm_core_check_feature(dev, DRIVER_GEM)) | |
183 | drm_gem_open(dev, file); | |
184 | ||
185 | if (drm_core_check_feature(dev, DRIVER_SYNCOBJ)) | |
186 | drm_syncobj_open(file); | |
187 | ||
ae75f836 | 188 | drm_prime_init_file_private(&file->prime); |
1572042a DH |
189 | |
190 | if (dev->driver->open) { | |
191 | ret = dev->driver->open(dev, file); | |
192 | if (ret < 0) | |
193 | goto out_prime_destroy; | |
194 | } | |
195 | ||
1572042a DH |
196 | return file; |
197 | ||
1572042a | 198 | out_prime_destroy: |
ae75f836 | 199 | drm_prime_destroy_file_private(&file->prime); |
1572042a DH |
200 | if (drm_core_check_feature(dev, DRIVER_SYNCOBJ)) |
201 | drm_syncobj_release(file); | |
202 | if (drm_core_check_feature(dev, DRIVER_GEM)) | |
203 | drm_gem_release(dev, file); | |
204 | put_pid(file->pid); | |
205 | kfree(file); | |
206 | ||
207 | return ERR_PTR(ret); | |
208 | } | |
209 | ||
210 | static void drm_events_release(struct drm_file *file_priv) | |
211 | { | |
212 | struct drm_device *dev = file_priv->minor->dev; | |
213 | struct drm_pending_event *e, *et; | |
214 | unsigned long flags; | |
215 | ||
216 | spin_lock_irqsave(&dev->event_lock, flags); | |
217 | ||
218 | /* Unlink pending events */ | |
219 | list_for_each_entry_safe(e, et, &file_priv->pending_event_list, | |
220 | pending_link) { | |
221 | list_del(&e->pending_link); | |
222 | e->file_priv = NULL; | |
223 | } | |
224 | ||
225 | /* Remove unconsumed events */ | |
226 | list_for_each_entry_safe(e, et, &file_priv->event_list, link) { | |
227 | list_del(&e->link); | |
228 | kfree(e); | |
229 | } | |
230 | ||
231 | spin_unlock_irqrestore(&dev->event_lock, flags); | |
232 | } | |
233 | ||
234 | /** | |
235 | * drm_file_free - free file context | |
236 | * @file: context to free, or NULL | |
237 | * | |
238 | * This destroys and deallocates a DRM file context previously allocated via | |
239 | * drm_file_alloc(). The caller must make sure to unlink it from any contexts | |
240 | * before calling this. | |
241 | * | |
242 | * If NULL is passed, this is a no-op. | |
243 | * | |
244 | * RETURNS: | |
245 | * 0 on success, or error code on failure. | |
246 | */ | |
247 | void drm_file_free(struct drm_file *file) | |
248 | { | |
249 | struct drm_device *dev; | |
250 | ||
251 | if (!file) | |
252 | return; | |
253 | ||
254 | dev = file->minor->dev; | |
255 | ||
5a2ba992 EV |
256 | DRM_DEBUG("comm=\"%s\", pid=%d, dev=0x%lx, open_count=%d\n", |
257 | current->comm, task_pid_nr(current), | |
1572042a | 258 | (long)old_encode_dev(file->minor->kdev->devt), |
7e13ad89 | 259 | atomic_read(&dev->open_count)); |
1572042a | 260 | |
57bb1ee6 | 261 | #ifdef CONFIG_DRM_LEGACY |
1572042a DH |
262 | if (drm_core_check_feature(dev, DRIVER_LEGACY) && |
263 | dev->driver->preclose) | |
264 | dev->driver->preclose(dev, file); | |
57bb1ee6 | 265 | #endif |
1572042a DH |
266 | |
267 | if (drm_core_check_feature(dev, DRIVER_LEGACY)) | |
268 | drm_legacy_lock_release(dev, file->filp); | |
269 | ||
270 | if (drm_core_check_feature(dev, DRIVER_HAVE_DMA)) | |
271 | drm_legacy_reclaim_buffers(dev, file); | |
272 | ||
273 | drm_events_release(file); | |
274 | ||
275 | if (drm_core_check_feature(dev, DRIVER_MODESET)) { | |
276 | drm_fb_release(file); | |
277 | drm_property_destroy_user_blobs(dev, file); | |
278 | } | |
279 | ||
280 | if (drm_core_check_feature(dev, DRIVER_SYNCOBJ)) | |
281 | drm_syncobj_release(file); | |
282 | ||
283 | if (drm_core_check_feature(dev, DRIVER_GEM)) | |
284 | drm_gem_release(dev, file); | |
285 | ||
286 | drm_legacy_ctxbitmap_flush(dev, file); | |
287 | ||
288 | if (drm_is_primary_client(file)) | |
289 | drm_master_release(file); | |
290 | ||
291 | if (dev->driver->postclose) | |
292 | dev->driver->postclose(dev, file); | |
293 | ||
ae75f836 | 294 | drm_prime_destroy_file_private(&file->prime); |
1572042a DH |
295 | |
296 | WARN_ON(!list_empty(&file->event_list)); | |
297 | ||
298 | put_pid(file->pid); | |
299 | kfree(file); | |
300 | } | |
301 | ||
e21710a8 EV |
302 | static void drm_close_helper(struct file *filp) |
303 | { | |
304 | struct drm_file *file_priv = filp->private_data; | |
305 | struct drm_device *dev = file_priv->minor->dev; | |
306 | ||
307 | mutex_lock(&dev->filelist_mutex); | |
308 | list_del(&file_priv->lhead); | |
309 | mutex_unlock(&dev->filelist_mutex); | |
310 | ||
311 | drm_file_free(file_priv); | |
312 | } | |
313 | ||
bcb877e4 | 314 | /* |
d985c108 DA |
315 | * Check whether DRI will run on this CPU. |
316 | * | |
317 | * \return non-zero if the DRI will run on this CPU, or zero otherwise. | |
318 | */ | |
319 | static int drm_cpu_valid(void) | |
320 | { | |
d985c108 DA |
321 | #if defined(__sparc__) && !defined(__sparc_v9__) |
322 | return 0; /* No cmpxchg before v9 sparc. */ | |
323 | #endif | |
324 | return 1; | |
325 | } | |
326 | ||
bcb877e4 | 327 | /* |
85dce7ff | 328 | * Called whenever a process opens a drm node |
d985c108 | 329 | * |
d985c108 | 330 | * \param filp file pointer. |
f4aede2e | 331 | * \param minor acquired minor-object. |
d985c108 DA |
332 | * \return zero on success or a negative number on failure. |
333 | * | |
334 | * Creates and initializes a drm_file structure for the file private data in \p | |
335 | * filp and add it into the double linked list in \p dev. | |
336 | */ | |
1dcc0ceb | 337 | static int drm_open_helper(struct file *filp, struct drm_minor *minor) |
d985c108 | 338 | { |
f4aede2e | 339 | struct drm_device *dev = minor->dev; |
84b1fd10 | 340 | struct drm_file *priv; |
7eeaeb90 | 341 | int ret; |
d985c108 DA |
342 | |
343 | if (filp->f_flags & O_EXCL) | |
344 | return -EBUSY; /* No exclusive opens */ | |
345 | if (!drm_cpu_valid()) | |
346 | return -EINVAL; | |
17ee1eb6 EV |
347 | if (dev->switch_power_state != DRM_SWITCH_POWER_ON && |
348 | dev->switch_power_state != DRM_SWITCH_POWER_DYNAMIC_OFF) | |
5bcf719b | 349 | return -EINVAL; |
d985c108 | 350 | |
5a2ba992 EV |
351 | DRM_DEBUG("comm=\"%s\", pid=%d, minor=%d\n", current->comm, |
352 | task_pid_nr(current), minor->index); | |
d985c108 | 353 | |
1572042a DH |
354 | priv = drm_file_alloc(minor); |
355 | if (IS_ERR(priv)) | |
356 | return PTR_ERR(priv); | |
d985c108 | 357 | |
7eeaeb90 NT |
358 | if (drm_is_primary_client(priv)) { |
359 | ret = drm_master_open(priv); | |
360 | if (ret) { | |
361 | drm_file_free(priv); | |
362 | return ret; | |
363 | } | |
364 | } | |
365 | ||
d985c108 | 366 | filp->private_data = priv; |
76ef6b28 | 367 | filp->f_mode |= FMODE_UNSIGNED_OFFSET; |
6c340eac | 368 | priv->filp = filp; |
bd1b331f | 369 | |
1d2ac403 | 370 | mutex_lock(&dev->filelist_mutex); |
bd1b331f | 371 | list_add(&priv->lhead, &dev->filelist); |
1d2ac403 | 372 | mutex_unlock(&dev->filelist_mutex); |
d985c108 DA |
373 | |
374 | #ifdef __alpha__ | |
375 | /* | |
376 | * Default the hose | |
377 | */ | |
378 | if (!dev->hose) { | |
379 | struct pci_dev *pci_dev; | |
948de842 | 380 | |
d985c108 DA |
381 | pci_dev = pci_get_class(PCI_CLASS_DISPLAY_VGA << 8, NULL); |
382 | if (pci_dev) { | |
383 | dev->hose = pci_dev->sysdata; | |
384 | pci_dev_put(pci_dev); | |
385 | } | |
386 | if (!dev->hose) { | |
59c1ad3b YW |
387 | struct pci_bus *b = list_entry(pci_root_buses.next, |
388 | struct pci_bus, node); | |
d985c108 DA |
389 | if (b) |
390 | dev->hose = b->sysdata; | |
391 | } | |
392 | } | |
393 | #endif | |
394 | ||
395 | return 0; | |
c9a9c5e0 KH |
396 | } |
397 | ||
094aa54f DV |
398 | /** |
399 | * drm_open - open method for DRM file | |
400 | * @inode: device inode | |
401 | * @filp: file pointer. | |
402 | * | |
403 | * This function must be used by drivers as their &file_operations.open method. | |
404 | * It looks up the correct DRM device and instantiates all the per-file | |
405 | * resources for it. It also calls the &drm_driver.open driver callback. | |
406 | * | |
407 | * RETURNS: | |
408 | * | |
409 | * 0 on success or negative errno value on falure. | |
410 | */ | |
411 | int drm_open(struct inode *inode, struct file *filp) | |
412 | { | |
413 | struct drm_device *dev; | |
414 | struct drm_minor *minor; | |
415 | int retcode; | |
416 | int need_setup = 0; | |
417 | ||
418 | minor = drm_minor_acquire(iminor(inode)); | |
419 | if (IS_ERR(minor)) | |
420 | return PTR_ERR(minor); | |
421 | ||
422 | dev = minor->dev; | |
4017ad7b DV |
423 | if (drm_dev_needs_global_mutex(dev)) |
424 | mutex_lock(&drm_global_mutex); | |
425 | ||
7e13ad89 | 426 | if (!atomic_fetch_inc(&dev->open_count)) |
094aa54f DV |
427 | need_setup = 1; |
428 | ||
429 | /* share address_space across all char-devs of a single device */ | |
430 | filp->f_mapping = dev->anon_inode->i_mapping; | |
431 | ||
432 | retcode = drm_open_helper(filp, minor); | |
433 | if (retcode) | |
434 | goto err_undo; | |
435 | if (need_setup) { | |
436 | retcode = drm_legacy_setup(dev); | |
437 | if (retcode) { | |
438 | drm_close_helper(filp); | |
439 | goto err_undo; | |
440 | } | |
441 | } | |
591a2abf | 442 | |
4017ad7b DV |
443 | if (drm_dev_needs_global_mutex(dev)) |
444 | mutex_unlock(&drm_global_mutex); | |
591a2abf | 445 | |
094aa54f DV |
446 | return 0; |
447 | ||
448 | err_undo: | |
7e13ad89 | 449 | atomic_dec(&dev->open_count); |
4017ad7b DV |
450 | if (drm_dev_needs_global_mutex(dev)) |
451 | mutex_unlock(&drm_global_mutex); | |
094aa54f DV |
452 | drm_minor_release(minor); |
453 | return retcode; | |
454 | } | |
455 | EXPORT_SYMBOL(drm_open); | |
456 | ||
68dfbeba | 457 | void drm_lastclose(struct drm_device * dev) |
1c8887dd | 458 | { |
1c8887dd DH |
459 | DRM_DEBUG("\n"); |
460 | ||
461 | if (dev->driver->lastclose) | |
462 | dev->driver->lastclose(dev); | |
463 | DRM_DEBUG("driver lastclose completed\n"); | |
464 | ||
fa538645 | 465 | if (drm_core_check_feature(dev, DRIVER_LEGACY)) |
68dfbeba | 466 | drm_legacy_dev_reinit(dev); |
c76f0f7c NT |
467 | |
468 | drm_client_dev_restore(dev); | |
1c8887dd DH |
469 | } |
470 | ||
1da177e4 | 471 | /** |
bcb877e4 DV |
472 | * drm_release - release method for DRM file |
473 | * @inode: device inode | |
474 | * @filp: file pointer. | |
1da177e4 | 475 | * |
b93658f8 DV |
476 | * This function must be used by drivers as their &file_operations.release |
477 | * method. It frees any resources associated with the open file, and calls the | |
45c3d213 DV |
478 | * &drm_driver.postclose driver callback. If this is the last open file for the |
479 | * DRM device also proceeds to call the &drm_driver.lastclose driver callback. | |
1da177e4 | 480 | * |
bcb877e4 DV |
481 | * RETURNS: |
482 | * | |
483 | * Always succeeds and returns 0. | |
1da177e4 | 484 | */ |
b5e89ed5 | 485 | int drm_release(struct inode *inode, struct file *filp) |
1da177e4 | 486 | { |
6c340eac | 487 | struct drm_file *file_priv = filp->private_data; |
1616c525 DH |
488 | struct drm_minor *minor = file_priv->minor; |
489 | struct drm_device *dev = minor->dev; | |
1da177e4 | 490 | |
4017ad7b DV |
491 | if (drm_dev_needs_global_mutex(dev)) |
492 | mutex_lock(&drm_global_mutex); | |
1da177e4 | 493 | |
7e13ad89 | 494 | DRM_DEBUG("open_count = %d\n", atomic_read(&dev->open_count)); |
1da177e4 | 495 | |
e21710a8 | 496 | drm_close_helper(filp); |
1da177e4 | 497 | |
7e13ad89 | 498 | if (atomic_dec_and_test(&dev->open_count)) |
68dfbeba | 499 | drm_lastclose(dev); |
1ee57d4d | 500 | |
4017ad7b DV |
501 | if (drm_dev_needs_global_mutex(dev)) |
502 | mutex_unlock(&drm_global_mutex); | |
1da177e4 | 503 | |
1616c525 DH |
504 | drm_minor_release(minor); |
505 | ||
68dfbeba | 506 | return 0; |
1da177e4 LT |
507 | } |
508 | EXPORT_SYMBOL(drm_release); | |
509 | ||
7a2c65dd CW |
510 | /** |
511 | * drm_release_noglobal - release method for DRM file | |
512 | * @inode: device inode | |
513 | * @filp: file pointer. | |
514 | * | |
515 | * This function may be used by drivers as their &file_operations.release | |
516 | * method. It frees any resources associated with the open file prior to taking | |
517 | * the drm_global_mutex, which then calls the &drm_driver.postclose driver | |
518 | * callback. If this is the last open file for the DRM device also proceeds to | |
519 | * call the &drm_driver.lastclose driver callback. | |
520 | * | |
521 | * RETURNS: | |
522 | * | |
523 | * Always succeeds and returns 0. | |
524 | */ | |
525 | int drm_release_noglobal(struct inode *inode, struct file *filp) | |
526 | { | |
527 | struct drm_file *file_priv = filp->private_data; | |
528 | struct drm_minor *minor = file_priv->minor; | |
529 | struct drm_device *dev = minor->dev; | |
530 | ||
531 | drm_close_helper(filp); | |
532 | ||
7e13ad89 | 533 | if (atomic_dec_and_mutex_lock(&dev->open_count, &drm_global_mutex)) { |
7a2c65dd | 534 | drm_lastclose(dev); |
7e13ad89 CW |
535 | mutex_unlock(&drm_global_mutex); |
536 | } | |
7a2c65dd CW |
537 | |
538 | drm_minor_release(minor); | |
539 | ||
540 | return 0; | |
541 | } | |
542 | EXPORT_SYMBOL(drm_release_noglobal); | |
543 | ||
bcb877e4 DV |
544 | /** |
545 | * drm_read - read method for DRM file | |
546 | * @filp: file pointer | |
547 | * @buffer: userspace destination pointer for the read | |
548 | * @count: count in bytes to read | |
549 | * @offset: offset to read | |
550 | * | |
b93658f8 | 551 | * This function must be used by drivers as their &file_operations.read |
bcb877e4 DV |
552 | * method iff they use DRM events for asynchronous signalling to userspace. |
553 | * Since events are used by the KMS API for vblank and page flip completion this | |
554 | * means all modern display drivers must use it. | |
555 | * | |
b93658f8 DV |
556 | * @offset is ignored, DRM events are read like a pipe. Therefore drivers also |
557 | * must set the &file_operation.llseek to no_llseek(). Polling support is | |
bcb877e4 DV |
558 | * provided by drm_poll(). |
559 | * | |
560 | * This function will only ever read a full event. Therefore userspace must | |
561 | * supply a big enough buffer to fit any event to ensure forward progress. Since | |
562 | * the maximum event space is currently 4K it's recommended to just use that for | |
563 | * safety. | |
564 | * | |
565 | * RETURNS: | |
566 | * | |
567 | * Number of bytes read (always aligned to full events, and can be 0) or a | |
568 | * negative error code on failure. | |
569 | */ | |
cdd1cf79 CW |
570 | ssize_t drm_read(struct file *filp, char __user *buffer, |
571 | size_t count, loff_t *offset) | |
c9a9c5e0 | 572 | { |
cdd1cf79 | 573 | struct drm_file *file_priv = filp->private_data; |
c9a9c5e0 | 574 | struct drm_device *dev = file_priv->minor->dev; |
9b2c0b7f | 575 | ssize_t ret; |
c9a9c5e0 | 576 | |
9b2c0b7f CW |
577 | ret = mutex_lock_interruptible(&file_priv->event_read_lock); |
578 | if (ret) | |
579 | return ret; | |
580 | ||
cdd1cf79 | 581 | for (;;) { |
83eb64c8 CW |
582 | struct drm_pending_event *e = NULL; |
583 | ||
584 | spin_lock_irq(&dev->event_lock); | |
585 | if (!list_empty(&file_priv->event_list)) { | |
586 | e = list_first_entry(&file_priv->event_list, | |
587 | struct drm_pending_event, link); | |
588 | file_priv->event_space += e->event->length; | |
589 | list_del(&e->link); | |
590 | } | |
591 | spin_unlock_irq(&dev->event_lock); | |
592 | ||
593 | if (e == NULL) { | |
cdd1cf79 CW |
594 | if (ret) |
595 | break; | |
c9a9c5e0 | 596 | |
cdd1cf79 CW |
597 | if (filp->f_flags & O_NONBLOCK) { |
598 | ret = -EAGAIN; | |
599 | break; | |
600 | } | |
c9a9c5e0 | 601 | |
9b2c0b7f | 602 | mutex_unlock(&file_priv->event_read_lock); |
cdd1cf79 CW |
603 | ret = wait_event_interruptible(file_priv->event_wait, |
604 | !list_empty(&file_priv->event_list)); | |
9b2c0b7f CW |
605 | if (ret >= 0) |
606 | ret = mutex_lock_interruptible(&file_priv->event_read_lock); | |
607 | if (ret) | |
608 | return ret; | |
cdd1cf79 | 609 | } else { |
83eb64c8 CW |
610 | unsigned length = e->event->length; |
611 | ||
612 | if (length > count - ret) { | |
613 | put_back_event: | |
614 | spin_lock_irq(&dev->event_lock); | |
615 | file_priv->event_space -= length; | |
616 | list_add(&e->link, &file_priv->event_list); | |
617 | spin_unlock_irq(&dev->event_lock); | |
87189b78 KL |
618 | wake_up_interruptible_poll(&file_priv->event_wait, |
619 | EPOLLIN | EPOLLRDNORM); | |
cdd1cf79 | 620 | break; |
83eb64c8 | 621 | } |
cdd1cf79 | 622 | |
83eb64c8 | 623 | if (copy_to_user(buffer + ret, e->event, length)) { |
cdd1cf79 CW |
624 | if (ret == 0) |
625 | ret = -EFAULT; | |
83eb64c8 | 626 | goto put_back_event; |
cdd1cf79 | 627 | } |
c9a9c5e0 | 628 | |
83eb64c8 | 629 | ret += length; |
1b47aaf9 | 630 | kfree(e); |
c9a9c5e0 | 631 | } |
c9a9c5e0 | 632 | } |
9b2c0b7f | 633 | mutex_unlock(&file_priv->event_read_lock); |
c9a9c5e0 | 634 | |
cdd1cf79 | 635 | return ret; |
c9a9c5e0 KH |
636 | } |
637 | EXPORT_SYMBOL(drm_read); | |
638 | ||
bcb877e4 DV |
639 | /** |
640 | * drm_poll - poll method for DRM file | |
641 | * @filp: file pointer | |
642 | * @wait: poll waiter table | |
643 | * | |
b93658f8 DV |
644 | * This function must be used by drivers as their &file_operations.read method |
645 | * iff they use DRM events for asynchronous signalling to userspace. Since | |
646 | * events are used by the KMS API for vblank and page flip completion this means | |
647 | * all modern display drivers must use it. | |
bcb877e4 DV |
648 | * |
649 | * See also drm_read(). | |
650 | * | |
651 | * RETURNS: | |
652 | * | |
653 | * Mask of POLL flags indicating the current status of the file. | |
654 | */ | |
afc9a42b | 655 | __poll_t drm_poll(struct file *filp, struct poll_table_struct *wait) |
1da177e4 | 656 | { |
c9a9c5e0 | 657 | struct drm_file *file_priv = filp->private_data; |
afc9a42b | 658 | __poll_t mask = 0; |
c9a9c5e0 KH |
659 | |
660 | poll_wait(filp, &file_priv->event_wait, wait); | |
661 | ||
662 | if (!list_empty(&file_priv->event_list)) | |
a9a08845 | 663 | mask |= EPOLLIN | EPOLLRDNORM; |
c9a9c5e0 KH |
664 | |
665 | return mask; | |
1da177e4 | 666 | } |
b5e89ed5 | 667 | EXPORT_SYMBOL(drm_poll); |
2dd500f1 DV |
668 | |
669 | /** | |
4020b220 | 670 | * drm_event_reserve_init_locked - init a DRM event and reserve space for it |
2dd500f1 DV |
671 | * @dev: DRM device |
672 | * @file_priv: DRM file private data | |
673 | * @p: tracking structure for the pending event | |
674 | * @e: actual event data to deliver to userspace | |
675 | * | |
676 | * This function prepares the passed in event for eventual delivery. If the event | |
677 | * doesn't get delivered (because the IOCTL fails later on, before queuing up | |
678 | * anything) then the even must be cancelled and freed using | |
fb740cf2 DV |
679 | * drm_event_cancel_free(). Successfully initialized events should be sent out |
680 | * using drm_send_event() or drm_send_event_locked() to signal completion of the | |
681 | * asynchronous event to userspace. | |
2dd500f1 DV |
682 | * |
683 | * If callers embedded @p into a larger structure it must be allocated with | |
684 | * kmalloc and @p must be the first member element. | |
685 | * | |
4020b220 | 686 | * This is the locked version of drm_event_reserve_init() for callers which |
ef40cbf9 | 687 | * already hold &drm_device.event_lock. |
4020b220 | 688 | * |
2dd500f1 DV |
689 | * RETURNS: |
690 | * | |
691 | * 0 on success or a negative error code on failure. | |
692 | */ | |
4020b220 DV |
693 | int drm_event_reserve_init_locked(struct drm_device *dev, |
694 | struct drm_file *file_priv, | |
695 | struct drm_pending_event *p, | |
696 | struct drm_event *e) | |
2dd500f1 | 697 | { |
4020b220 DV |
698 | if (file_priv->event_space < e->length) |
699 | return -ENOMEM; | |
2dd500f1 DV |
700 | |
701 | file_priv->event_space -= e->length; | |
702 | ||
703 | p->event = e; | |
681047b4 | 704 | list_add(&p->pending_link, &file_priv->pending_event_list); |
2dd500f1 DV |
705 | p->file_priv = file_priv; |
706 | ||
4020b220 DV |
707 | return 0; |
708 | } | |
709 | EXPORT_SYMBOL(drm_event_reserve_init_locked); | |
710 | ||
711 | /** | |
712 | * drm_event_reserve_init - init a DRM event and reserve space for it | |
713 | * @dev: DRM device | |
714 | * @file_priv: DRM file private data | |
715 | * @p: tracking structure for the pending event | |
716 | * @e: actual event data to deliver to userspace | |
717 | * | |
718 | * This function prepares the passed in event for eventual delivery. If the event | |
719 | * doesn't get delivered (because the IOCTL fails later on, before queuing up | |
720 | * anything) then the even must be cancelled and freed using | |
721 | * drm_event_cancel_free(). Successfully initialized events should be sent out | |
722 | * using drm_send_event() or drm_send_event_locked() to signal completion of the | |
723 | * asynchronous event to userspace. | |
724 | * | |
725 | * If callers embedded @p into a larger structure it must be allocated with | |
726 | * kmalloc and @p must be the first member element. | |
727 | * | |
ef40cbf9 | 728 | * Callers which already hold &drm_device.event_lock should use |
20c9ca4f | 729 | * drm_event_reserve_init_locked() instead. |
4020b220 DV |
730 | * |
731 | * RETURNS: | |
732 | * | |
733 | * 0 on success or a negative error code on failure. | |
734 | */ | |
735 | int drm_event_reserve_init(struct drm_device *dev, | |
736 | struct drm_file *file_priv, | |
737 | struct drm_pending_event *p, | |
738 | struct drm_event *e) | |
739 | { | |
740 | unsigned long flags; | |
741 | int ret; | |
742 | ||
743 | spin_lock_irqsave(&dev->event_lock, flags); | |
744 | ret = drm_event_reserve_init_locked(dev, file_priv, p, e); | |
2dd500f1 | 745 | spin_unlock_irqrestore(&dev->event_lock, flags); |
4020b220 | 746 | |
2dd500f1 DV |
747 | return ret; |
748 | } | |
749 | EXPORT_SYMBOL(drm_event_reserve_init); | |
750 | ||
751 | /** | |
1e55a53a | 752 | * drm_event_cancel_free - free a DRM event and release its space |
2dd500f1 DV |
753 | * @dev: DRM device |
754 | * @p: tracking structure for the pending event | |
755 | * | |
756 | * This function frees the event @p initialized with drm_event_reserve_init() | |
b93658f8 DV |
757 | * and releases any allocated space. It is used to cancel an event when the |
758 | * nonblocking operation could not be submitted and needed to be aborted. | |
2dd500f1 DV |
759 | */ |
760 | void drm_event_cancel_free(struct drm_device *dev, | |
761 | struct drm_pending_event *p) | |
762 | { | |
763 | unsigned long flags; | |
948de842 | 764 | |
2dd500f1 | 765 | spin_lock_irqsave(&dev->event_lock, flags); |
681047b4 DV |
766 | if (p->file_priv) { |
767 | p->file_priv->event_space += p->event->length; | |
768 | list_del(&p->pending_link); | |
769 | } | |
2dd500f1 | 770 | spin_unlock_irqrestore(&dev->event_lock, flags); |
838de39f GP |
771 | |
772 | if (p->fence) | |
f54d1867 | 773 | dma_fence_put(p->fence); |
838de39f | 774 | |
1b47aaf9 | 775 | kfree(p); |
2dd500f1 DV |
776 | } |
777 | EXPORT_SYMBOL(drm_event_cancel_free); | |
fb740cf2 DV |
778 | |
779 | /** | |
780 | * drm_send_event_locked - send DRM event to file descriptor | |
781 | * @dev: DRM device | |
782 | * @e: DRM event to deliver | |
783 | * | |
784 | * This function sends the event @e, initialized with drm_event_reserve_init(), | |
785 | * to its associated userspace DRM file. Callers must already hold | |
ef40cbf9 | 786 | * &drm_device.event_lock, see drm_send_event() for the unlocked version. |
681047b4 DV |
787 | * |
788 | * Note that the core will take care of unlinking and disarming events when the | |
789 | * corresponding DRM file is closed. Drivers need not worry about whether the | |
790 | * DRM file for this event still exists and can call this function upon | |
791 | * completion of the asynchronous work unconditionally. | |
fb740cf2 DV |
792 | */ |
793 | void drm_send_event_locked(struct drm_device *dev, struct drm_pending_event *e) | |
794 | { | |
795 | assert_spin_locked(&dev->event_lock); | |
796 | ||
3b24f7d6 | 797 | if (e->completion) { |
3b24f7d6 | 798 | complete_all(e->completion); |
24835e44 | 799 | e->completion_release(e->completion); |
3b24f7d6 DV |
800 | e->completion = NULL; |
801 | } | |
802 | ||
1b47aaf9 | 803 | if (e->fence) { |
f54d1867 CW |
804 | dma_fence_signal(e->fence); |
805 | dma_fence_put(e->fence); | |
1b47aaf9 GP |
806 | } |
807 | ||
681047b4 | 808 | if (!e->file_priv) { |
1b47aaf9 | 809 | kfree(e); |
681047b4 DV |
810 | return; |
811 | } | |
812 | ||
813 | list_del(&e->pending_link); | |
fb740cf2 DV |
814 | list_add_tail(&e->link, |
815 | &e->file_priv->event_list); | |
87189b78 KL |
816 | wake_up_interruptible_poll(&e->file_priv->event_wait, |
817 | EPOLLIN | EPOLLRDNORM); | |
fb740cf2 DV |
818 | } |
819 | EXPORT_SYMBOL(drm_send_event_locked); | |
820 | ||
821 | /** | |
822 | * drm_send_event - send DRM event to file descriptor | |
823 | * @dev: DRM device | |
824 | * @e: DRM event to deliver | |
825 | * | |
826 | * This function sends the event @e, initialized with drm_event_reserve_init(), | |
ef40cbf9 DV |
827 | * to its associated userspace DRM file. This function acquires |
828 | * &drm_device.event_lock, see drm_send_event_locked() for callers which already | |
829 | * hold this lock. | |
681047b4 DV |
830 | * |
831 | * Note that the core will take care of unlinking and disarming events when the | |
832 | * corresponding DRM file is closed. Drivers need not worry about whether the | |
833 | * DRM file for this event still exists and can call this function upon | |
834 | * completion of the asynchronous work unconditionally. | |
fb740cf2 DV |
835 | */ |
836 | void drm_send_event(struct drm_device *dev, struct drm_pending_event *e) | |
837 | { | |
838 | unsigned long irqflags; | |
839 | ||
840 | spin_lock_irqsave(&dev->event_lock, irqflags); | |
841 | drm_send_event_locked(dev, e); | |
842 | spin_unlock_irqrestore(&dev->event_lock, irqflags); | |
843 | } | |
844 | EXPORT_SYMBOL(drm_send_event); | |
4748aa16 CW |
845 | |
846 | /** | |
847 | * mock_drm_getfile - Create a new struct file for the drm device | |
848 | * @minor: drm minor to wrap (e.g. #drm_device.primary) | |
849 | * @flags: file creation mode (O_RDWR etc) | |
850 | * | |
851 | * This create a new struct file that wraps a DRM file context around a | |
852 | * DRM minor. This mimicks userspace opening e.g. /dev/dri/card0, but without | |
853 | * invoking userspace. The struct file may be operated on using its f_op | |
854 | * (the drm_device.driver.fops) to mimick userspace operations, or be supplied | |
855 | * to userspace facing functions as an internal/anonymous client. | |
856 | * | |
857 | * RETURNS: | |
858 | * Pointer to newly created struct file, ERR_PTR on failure. | |
859 | */ | |
860 | struct file *mock_drm_getfile(struct drm_minor *minor, unsigned int flags) | |
861 | { | |
862 | struct drm_device *dev = minor->dev; | |
863 | struct drm_file *priv; | |
864 | struct file *file; | |
865 | ||
866 | priv = drm_file_alloc(minor); | |
867 | if (IS_ERR(priv)) | |
868 | return ERR_CAST(priv); | |
869 | ||
870 | file = anon_inode_getfile("drm", dev->driver->fops, priv, flags); | |
871 | if (IS_ERR(file)) { | |
872 | drm_file_free(priv); | |
873 | return file; | |
874 | } | |
875 | ||
876 | /* Everyone shares a single global address space */ | |
877 | file->f_mapping = dev->anon_inode->i_mapping; | |
878 | ||
879 | drm_dev_get(dev); | |
880 | priv->filp = file; | |
881 | ||
882 | return file; | |
883 | } | |
884 | EXPORT_SYMBOL_FOR_TESTS_ONLY(mock_drm_getfile); | |
b1823416 THV |
885 | |
886 | #ifdef CONFIG_MMU | |
887 | #ifdef CONFIG_TRANSPARENT_HUGEPAGE | |
888 | /* | |
889 | * drm_addr_inflate() attempts to construct an aligned area by inflating | |
890 | * the area size and skipping the unaligned start of the area. | |
891 | * adapted from shmem_get_unmapped_area() | |
892 | */ | |
893 | static unsigned long drm_addr_inflate(unsigned long addr, | |
894 | unsigned long len, | |
895 | unsigned long pgoff, | |
896 | unsigned long flags, | |
897 | unsigned long huge_size) | |
898 | { | |
899 | unsigned long offset, inflated_len; | |
900 | unsigned long inflated_addr; | |
901 | unsigned long inflated_offset; | |
902 | ||
903 | offset = (pgoff << PAGE_SHIFT) & (huge_size - 1); | |
904 | if (offset && offset + len < 2 * huge_size) | |
905 | return addr; | |
906 | if ((addr & (huge_size - 1)) == offset) | |
907 | return addr; | |
908 | ||
909 | inflated_len = len + huge_size - PAGE_SIZE; | |
910 | if (inflated_len > TASK_SIZE) | |
911 | return addr; | |
912 | if (inflated_len < len) | |
913 | return addr; | |
914 | ||
915 | inflated_addr = current->mm->get_unmapped_area(NULL, 0, inflated_len, | |
916 | 0, flags); | |
917 | if (IS_ERR_VALUE(inflated_addr)) | |
918 | return addr; | |
919 | if (inflated_addr & ~PAGE_MASK) | |
920 | return addr; | |
921 | ||
922 | inflated_offset = inflated_addr & (huge_size - 1); | |
923 | inflated_addr += offset - inflated_offset; | |
924 | if (inflated_offset > offset) | |
925 | inflated_addr += huge_size; | |
926 | ||
927 | if (inflated_addr > TASK_SIZE - len) | |
928 | return addr; | |
929 | ||
930 | return inflated_addr; | |
931 | } | |
932 | ||
933 | /** | |
934 | * drm_get_unmapped_area() - Get an unused user-space virtual memory area | |
935 | * suitable for huge page table entries. | |
936 | * @file: The struct file representing the address space being mmap()'d. | |
937 | * @uaddr: Start address suggested by user-space. | |
938 | * @len: Length of the area. | |
939 | * @pgoff: The page offset into the address space. | |
940 | * @flags: mmap flags | |
941 | * @mgr: The address space manager used by the drm driver. This argument can | |
942 | * probably be removed at some point when all drivers use the same | |
943 | * address space manager. | |
944 | * | |
945 | * This function attempts to find an unused user-space virtual memory area | |
946 | * that can accommodate the size we want to map, and that is properly | |
947 | * aligned to facilitate huge page table entries matching actual | |
948 | * huge pages or huge page aligned memory in buffer objects. Buffer objects | |
949 | * are assumed to start at huge page boundary pfns (io memory) or be | |
950 | * populated by huge pages aligned to the start of the buffer object | |
951 | * (system- or coherent memory). Adapted from shmem_get_unmapped_area. | |
952 | * | |
953 | * Return: aligned user-space address. | |
954 | */ | |
955 | unsigned long drm_get_unmapped_area(struct file *file, | |
956 | unsigned long uaddr, unsigned long len, | |
957 | unsigned long pgoff, unsigned long flags, | |
958 | struct drm_vma_offset_manager *mgr) | |
959 | { | |
960 | unsigned long addr; | |
961 | unsigned long inflated_addr; | |
962 | struct drm_vma_offset_node *node; | |
963 | ||
964 | if (len > TASK_SIZE) | |
965 | return -ENOMEM; | |
966 | ||
967 | /* | |
968 | * @pgoff is the file page-offset the huge page boundaries of | |
969 | * which typically aligns to physical address huge page boundaries. | |
970 | * That's not true for DRM, however, where physical address huge | |
971 | * page boundaries instead are aligned with the offset from | |
972 | * buffer object start. So adjust @pgoff to be the offset from | |
973 | * buffer object start. | |
974 | */ | |
975 | drm_vma_offset_lock_lookup(mgr); | |
976 | node = drm_vma_offset_lookup_locked(mgr, pgoff, 1); | |
977 | if (node) | |
978 | pgoff -= node->vm_node.start; | |
979 | drm_vma_offset_unlock_lookup(mgr); | |
980 | ||
981 | addr = current->mm->get_unmapped_area(file, uaddr, len, pgoff, flags); | |
982 | if (IS_ERR_VALUE(addr)) | |
983 | return addr; | |
984 | if (addr & ~PAGE_MASK) | |
985 | return addr; | |
986 | if (addr > TASK_SIZE - len) | |
987 | return addr; | |
988 | ||
989 | if (len < HPAGE_PMD_SIZE) | |
990 | return addr; | |
991 | if (flags & MAP_FIXED) | |
992 | return addr; | |
993 | /* | |
994 | * Our priority is to support MAP_SHARED mapped hugely; | |
995 | * and support MAP_PRIVATE mapped hugely too, until it is COWed. | |
996 | * But if caller specified an address hint, respect that as before. | |
997 | */ | |
998 | if (uaddr) | |
999 | return addr; | |
1000 | ||
1001 | inflated_addr = drm_addr_inflate(addr, len, pgoff, flags, | |
1002 | HPAGE_PMD_SIZE); | |
1003 | ||
1004 | if (IS_ENABLED(CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD) && | |
1005 | len >= HPAGE_PUD_SIZE) | |
1006 | inflated_addr = drm_addr_inflate(inflated_addr, len, pgoff, | |
1007 | flags, HPAGE_PUD_SIZE); | |
1008 | return inflated_addr; | |
1009 | } | |
1010 | #else /* CONFIG_TRANSPARENT_HUGEPAGE */ | |
1011 | unsigned long drm_get_unmapped_area(struct file *file, | |
1012 | unsigned long uaddr, unsigned long len, | |
1013 | unsigned long pgoff, unsigned long flags, | |
1014 | struct drm_vma_offset_manager *mgr) | |
1015 | { | |
1016 | return current->mm->get_unmapped_area(file, uaddr, len, pgoff, flags); | |
1017 | } | |
1018 | #endif /* CONFIG_TRANSPARENT_HUGEPAGE */ | |
9431042d | 1019 | EXPORT_SYMBOL_GPL(drm_get_unmapped_area); |
b1823416 | 1020 | #endif /* CONFIG_MMU */ |