Commit | Line | Data |
---|---|---|
1da177e4 | 1 | /** |
b5e89ed5 | 2 | * \file drm_bufs.c |
1da177e4 | 3 | * Generic buffer template |
b5e89ed5 | 4 | * |
1da177e4 LT |
5 | * \author Rickard E. (Rik) Faith <faith@valinux.com> |
6 | * \author Gareth Hughes <gareth@valinux.com> | |
7 | */ | |
8 | ||
9 | /* | |
10 | * Created: Thu Nov 23 03:10:50 2000 by gareth@valinux.com | |
11 | * | |
12 | * Copyright 1999, 2000 Precision Insight, Inc., Cedar Park, Texas. | |
13 | * Copyright 2000 VA Linux Systems, Inc., Sunnyvale, California. | |
14 | * All Rights Reserved. | |
15 | * | |
16 | * Permission is hereby granted, free of charge, to any person obtaining a | |
17 | * copy of this software and associated documentation files (the "Software"), | |
18 | * to deal in the Software without restriction, including without limitation | |
19 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, | |
20 | * and/or sell copies of the Software, and to permit persons to whom the | |
21 | * Software is furnished to do so, subject to the following conditions: | |
22 | * | |
23 | * The above copyright notice and this permission notice (including the next | |
24 | * paragraph) shall be included in all copies or substantial portions of the | |
25 | * Software. | |
26 | * | |
27 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
28 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
29 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |
30 | * VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR | |
31 | * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, | |
32 | * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR | |
33 | * OTHER DEALINGS IN THE SOFTWARE. | |
34 | */ | |
35 | ||
36 | #include <linux/vmalloc.h> | |
37 | #include "drmP.h" | |
38 | ||
d985c108 | 39 | unsigned long drm_get_resource_start(drm_device_t *dev, unsigned int resource) |
1da177e4 | 40 | { |
836cf046 DA |
41 | return pci_resource_start(dev->pdev, resource); |
42 | } | |
43 | EXPORT_SYMBOL(drm_get_resource_start); | |
1da177e4 | 44 | |
d985c108 | 45 | unsigned long drm_get_resource_len(drm_device_t *dev, unsigned int resource) |
836cf046 DA |
46 | { |
47 | return pci_resource_len(dev->pdev, resource); | |
48 | } | |
b5e89ed5 | 49 | |
836cf046 | 50 | EXPORT_SYMBOL(drm_get_resource_len); |
1da177e4 | 51 | |
d985c108 DA |
52 | static drm_map_list_t *drm_find_matching_map(drm_device_t *dev, |
53 | drm_local_map_t *map) | |
836cf046 | 54 | { |
bd1b331f DA |
55 | drm_map_list_t *entry; |
56 | list_for_each_entry(entry, &dev->maplist, head) { | |
836cf046 | 57 | if (entry->map && map->type == entry->map->type && |
54ba2f76 DA |
58 | ((entry->map->offset == map->offset) || |
59 | (map->type == _DRM_SHM && map->flags==_DRM_CONTAINS_LOCK))) { | |
89625eb1 | 60 | return entry; |
836cf046 DA |
61 | } |
62 | } | |
63 | ||
64 | return NULL; | |
1da177e4 | 65 | } |
1da177e4 | 66 | |
fb41e54b AB |
67 | static int drm_map_handle(drm_device_t *dev, drm_hash_item_t *hash, |
68 | unsigned long user_token, int hashed_handle) | |
d1f2b55a | 69 | { |
8d153f71 | 70 | int use_hashed_handle; |
c2604ce0 | 71 | #if (BITS_PER_LONG == 64) |
8d153f71 TH |
72 | use_hashed_handle = ((user_token & 0xFFFFFFFF00000000UL) || hashed_handle); |
73 | #elif (BITS_PER_LONG == 32) | |
74 | use_hashed_handle = hashed_handle; | |
75 | #else | |
76 | #error Unsupported long size. Neither 64 nor 32 bits. | |
77 | #endif | |
d1f2b55a | 78 | |
e08870c8 TH |
79 | if (!use_hashed_handle) { |
80 | int ret; | |
1545085a | 81 | hash->key = user_token >> PAGE_SHIFT; |
e08870c8 TH |
82 | ret = drm_ht_insert_item(&dev->map_hash, hash); |
83 | if (ret != -EINVAL) | |
84 | return ret; | |
d1f2b55a | 85 | } |
e08870c8 TH |
86 | return drm_ht_just_insert_please(&dev->map_hash, hash, |
87 | user_token, 32 - PAGE_SHIFT - 3, | |
1545085a | 88 | 0, DRM_MAP_HASH_OFFSET >> PAGE_SHIFT); |
d1f2b55a | 89 | } |
9a186645 | 90 | |
1da177e4 LT |
91 | /** |
92 | * Ioctl to specify a range of memory that is available for mapping by a non-root process. | |
93 | * | |
94 | * \param inode device inode. | |
95 | * \param filp file pointer. | |
96 | * \param cmd command. | |
97 | * \param arg pointer to a drm_map structure. | |
98 | * \return zero on success or a negative value on error. | |
99 | * | |
100 | * Adjusts the memory offset to its absolute value according to the mapping | |
101 | * type. Adds the map to the map list drm_device::maplist. Adds MTRR's where | |
102 | * applicable and if supported by the kernel. | |
103 | */ | |
b3a83639 DA |
104 | static int drm_addmap_core(drm_device_t * dev, unsigned int offset, |
105 | unsigned int size, drm_map_type_t type, | |
106 | drm_map_flags_t flags, drm_map_list_t ** maplist) | |
1da177e4 | 107 | { |
1da177e4 | 108 | drm_map_t *map; |
1da177e4 | 109 | drm_map_list_t *list; |
9c8da5eb | 110 | drm_dma_handle_t *dmah; |
8d153f71 TH |
111 | unsigned long user_token; |
112 | int ret; | |
1da177e4 | 113 | |
b5e89ed5 DA |
114 | map = drm_alloc(sizeof(*map), DRM_MEM_MAPS); |
115 | if (!map) | |
1da177e4 LT |
116 | return -ENOMEM; |
117 | ||
7ab98401 DA |
118 | map->offset = offset; |
119 | map->size = size; | |
120 | map->flags = flags; | |
121 | map->type = type; | |
1da177e4 LT |
122 | |
123 | /* Only allow shared memory to be removable since we only keep enough | |
124 | * book keeping information about shared memory to allow for removal | |
125 | * when processes fork. | |
126 | */ | |
b5e89ed5 DA |
127 | if ((map->flags & _DRM_REMOVABLE) && map->type != _DRM_SHM) { |
128 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); | |
1da177e4 LT |
129 | return -EINVAL; |
130 | } | |
b5e89ed5 DA |
131 | DRM_DEBUG("offset = 0x%08lx, size = 0x%08lx, type = %d\n", |
132 | map->offset, map->size, map->type); | |
133 | if ((map->offset & (~PAGE_MASK)) || (map->size & (~PAGE_MASK))) { | |
134 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); | |
1da177e4 LT |
135 | return -EINVAL; |
136 | } | |
b5e89ed5 | 137 | map->mtrr = -1; |
1da177e4 LT |
138 | map->handle = NULL; |
139 | ||
b5e89ed5 | 140 | switch (map->type) { |
1da177e4 LT |
141 | case _DRM_REGISTERS: |
142 | case _DRM_FRAME_BUFFER: | |
88f399cd | 143 | #if !defined(__sparc__) && !defined(__alpha__) && !defined(__ia64__) && !defined(__powerpc64__) && !defined(__x86_64__) |
8d2ea625 | 144 | if (map->offset + (map->size-1) < map->offset || |
b5e89ed5 DA |
145 | map->offset < virt_to_phys(high_memory)) { |
146 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); | |
1da177e4 LT |
147 | return -EINVAL; |
148 | } | |
149 | #endif | |
150 | #ifdef __alpha__ | |
151 | map->offset += dev->hose->mem_space->start; | |
152 | #endif | |
836cf046 DA |
153 | /* Some drivers preinitialize some maps, without the X Server |
154 | * needing to be aware of it. Therefore, we just return success | |
155 | * when the server tries to create a duplicate map. | |
156 | */ | |
89625eb1 DA |
157 | list = drm_find_matching_map(dev, map); |
158 | if (list != NULL) { | |
159 | if (list->map->size != map->size) { | |
836cf046 | 160 | DRM_DEBUG("Matching maps of type %d with " |
b5e89ed5 DA |
161 | "mismatched sizes, (%ld vs %ld)\n", |
162 | map->type, map->size, | |
163 | list->map->size); | |
89625eb1 | 164 | list->map->size = map->size; |
836cf046 DA |
165 | } |
166 | ||
167 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); | |
89625eb1 | 168 | *maplist = list; |
836cf046 DA |
169 | return 0; |
170 | } | |
171 | ||
1da177e4 | 172 | if (drm_core_has_MTRR(dev)) { |
b5e89ed5 DA |
173 | if (map->type == _DRM_FRAME_BUFFER || |
174 | (map->flags & _DRM_WRITE_COMBINING)) { | |
175 | map->mtrr = mtrr_add(map->offset, map->size, | |
176 | MTRR_TYPE_WRCOMB, 1); | |
1da177e4 LT |
177 | } |
178 | } | |
179 | if (map->type == _DRM_REGISTERS) | |
004a7727 | 180 | map->handle = ioremap(map->offset, map->size); |
1da177e4 | 181 | break; |
1da177e4 | 182 | case _DRM_SHM: |
54ba2f76 DA |
183 | list = drm_find_matching_map(dev, map); |
184 | if (list != NULL) { | |
185 | if(list->map->size != map->size) { | |
186 | DRM_DEBUG("Matching maps of type %d with " | |
187 | "mismatched sizes, (%ld vs %ld)\n", | |
188 | map->type, map->size, list->map->size); | |
189 | list->map->size = map->size; | |
190 | } | |
191 | ||
192 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); | |
193 | *maplist = list; | |
194 | return 0; | |
195 | } | |
f239b7b0 | 196 | map->handle = vmalloc_user(map->size); |
b5e89ed5 DA |
197 | DRM_DEBUG("%lu %d %p\n", |
198 | map->size, drm_order(map->size), map->handle); | |
199 | if (!map->handle) { | |
200 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); | |
1da177e4 LT |
201 | return -ENOMEM; |
202 | } | |
203 | map->offset = (unsigned long)map->handle; | |
b5e89ed5 | 204 | if (map->flags & _DRM_CONTAINS_LOCK) { |
1da177e4 LT |
205 | /* Prevent a 2nd X Server from creating a 2nd lock */ |
206 | if (dev->lock.hw_lock != NULL) { | |
b5e89ed5 DA |
207 | vfree(map->handle); |
208 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); | |
1da177e4 LT |
209 | return -EBUSY; |
210 | } | |
b5e89ed5 | 211 | dev->sigdata.lock = dev->lock.hw_lock = map->handle; /* Pointer to lock */ |
1da177e4 LT |
212 | } |
213 | break; | |
54ba2f76 DA |
214 | case _DRM_AGP: { |
215 | drm_agp_mem_t *entry; | |
216 | int valid = 0; | |
217 | ||
218 | if (!drm_core_has_AGP(dev)) { | |
219 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); | |
220 | return -EINVAL; | |
221 | } | |
1da177e4 | 222 | #ifdef __alpha__ |
54ba2f76 | 223 | map->offset += dev->hose->mem_space->start; |
1da177e4 | 224 | #endif |
54ba2f76 DA |
225 | /* Note: dev->agp->base may actually be 0 when the DRM |
226 | * is not in control of AGP space. But if user space is | |
227 | * it should already have added the AGP base itself. | |
228 | */ | |
229 | map->offset += dev->agp->base; | |
230 | map->mtrr = dev->agp->agp_mtrr; /* for getmap */ | |
231 | ||
232 | /* This assumes the DRM is in total control of AGP space. | |
233 | * It's not always the case as AGP can be in the control | |
234 | * of user space (i.e. i810 driver). So this loop will get | |
235 | * skipped and we double check that dev->agp->memory is | |
236 | * actually set as well as being invalid before EPERM'ing | |
237 | */ | |
bd1b331f | 238 | list_for_each_entry(entry, &dev->agp->memory, head) { |
54ba2f76 DA |
239 | if ((map->offset >= entry->bound) && |
240 | (map->offset + map->size <= entry->bound + entry->pages * PAGE_SIZE)) { | |
241 | valid = 1; | |
242 | break; | |
243 | } | |
1da177e4 | 244 | } |
bd1b331f | 245 | if (!list_empty(&dev->agp->memory) && !valid) { |
54ba2f76 DA |
246 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); |
247 | return -EPERM; | |
248 | } | |
249 | DRM_DEBUG("AGP offset = 0x%08lx, size = 0x%08lx\n", map->offset, map->size); | |
250 | ||
1da177e4 | 251 | break; |
54ba2f76 | 252 | } |
1da177e4 LT |
253 | case _DRM_SCATTER_GATHER: |
254 | if (!dev->sg) { | |
255 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); | |
256 | return -EINVAL; | |
257 | } | |
d1f2b55a | 258 | map->offset += (unsigned long)dev->sg->virtual; |
1da177e4 | 259 | break; |
b5e89ed5 | 260 | case _DRM_CONSISTENT: |
2d0f9eaf | 261 | /* dma_addr_t is 64bit on i386 with CONFIG_HIGHMEM64G, |
9c8da5eb | 262 | * As we're limiting the address to 2^32-1 (or less), |
2d0f9eaf DA |
263 | * casting it down to 32 bits is no problem, but we |
264 | * need to point to a 64bit variable first. */ | |
9c8da5eb DA |
265 | dmah = drm_pci_alloc(dev, map->size, map->size, 0xffffffffUL); |
266 | if (!dmah) { | |
2d0f9eaf DA |
267 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); |
268 | return -ENOMEM; | |
269 | } | |
9c8da5eb DA |
270 | map->handle = dmah->vaddr; |
271 | map->offset = (unsigned long)dmah->busaddr; | |
272 | kfree(dmah); | |
2d0f9eaf | 273 | break; |
1da177e4 | 274 | default: |
b5e89ed5 | 275 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); |
1da177e4 LT |
276 | return -EINVAL; |
277 | } | |
278 | ||
279 | list = drm_alloc(sizeof(*list), DRM_MEM_MAPS); | |
b5e89ed5 | 280 | if (!list) { |
85abb3f9 | 281 | if (map->type == _DRM_REGISTERS) |
004a7727 | 282 | iounmap(map->handle); |
1da177e4 LT |
283 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); |
284 | return -EINVAL; | |
285 | } | |
286 | memset(list, 0, sizeof(*list)); | |
287 | list->map = map; | |
288 | ||
30e2fb18 | 289 | mutex_lock(&dev->struct_mutex); |
bd1b331f | 290 | list_add(&list->head, &dev->maplist); |
8d153f71 | 291 | |
d1f2b55a | 292 | /* Assign a 32-bit handle */ |
30e2fb18 | 293 | /* We do it here so that dev->struct_mutex protects the increment */ |
8d153f71 TH |
294 | user_token = (map->type == _DRM_SHM) ? (unsigned long)map->handle : |
295 | map->offset; | |
a1d0fcf5 | 296 | ret = drm_map_handle(dev, &list->hash, user_token, 0); |
8d153f71 | 297 | if (ret) { |
85abb3f9 | 298 | if (map->type == _DRM_REGISTERS) |
004a7727 | 299 | iounmap(map->handle); |
8d153f71 TH |
300 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); |
301 | drm_free(list, sizeof(*list), DRM_MEM_MAPS); | |
302 | mutex_unlock(&dev->struct_mutex); | |
303 | return ret; | |
304 | } | |
305 | ||
1545085a | 306 | list->user_token = list->hash.key << PAGE_SHIFT; |
30e2fb18 | 307 | mutex_unlock(&dev->struct_mutex); |
1da177e4 | 308 | |
89625eb1 | 309 | *maplist = list; |
7ab98401 | 310 | return 0; |
54ba2f76 | 311 | } |
89625eb1 | 312 | |
b5e89ed5 | 313 | int drm_addmap(drm_device_t * dev, unsigned int offset, |
89625eb1 | 314 | unsigned int size, drm_map_type_t type, |
b5e89ed5 | 315 | drm_map_flags_t flags, drm_local_map_t ** map_ptr) |
89625eb1 DA |
316 | { |
317 | drm_map_list_t *list; | |
318 | int rc; | |
319 | ||
320 | rc = drm_addmap_core(dev, offset, size, type, flags, &list); | |
321 | if (!rc) | |
322 | *map_ptr = list->map; | |
323 | return rc; | |
324 | } | |
b5e89ed5 | 325 | |
7ab98401 DA |
326 | EXPORT_SYMBOL(drm_addmap); |
327 | ||
328 | int drm_addmap_ioctl(struct inode *inode, struct file *filp, | |
329 | unsigned int cmd, unsigned long arg) | |
330 | { | |
331 | drm_file_t *priv = filp->private_data; | |
332 | drm_device_t *dev = priv->head->dev; | |
333 | drm_map_t map; | |
89625eb1 | 334 | drm_map_list_t *maplist; |
7ab98401 DA |
335 | drm_map_t __user *argp = (void __user *)arg; |
336 | int err; | |
337 | ||
338 | if (!(filp->f_mode & 3)) | |
339 | return -EACCES; /* Require read/write */ | |
340 | ||
b5e89ed5 | 341 | if (copy_from_user(&map, argp, sizeof(map))) { |
1da177e4 | 342 | return -EFAULT; |
7ab98401 DA |
343 | } |
344 | ||
d985c108 DA |
345 | if (!(capable(CAP_SYS_ADMIN) || map.type == _DRM_AGP)) |
346 | return -EPERM; | |
347 | ||
89625eb1 DA |
348 | err = drm_addmap_core(dev, map.offset, map.size, map.type, map.flags, |
349 | &maplist); | |
7ab98401 | 350 | |
b5e89ed5 | 351 | if (err) |
7ab98401 | 352 | return err; |
d1f2b55a | 353 | |
89625eb1 | 354 | if (copy_to_user(argp, maplist->map, sizeof(drm_map_t))) |
d1f2b55a | 355 | return -EFAULT; |
67e1a014 DA |
356 | |
357 | /* avoid a warning on 64-bit, this casting isn't very nice, but the API is set so too late */ | |
358 | if (put_user((void *)(unsigned long)maplist->user_token, &argp->handle)) | |
d1f2b55a | 359 | return -EFAULT; |
1da177e4 | 360 | return 0; |
88f399cd | 361 | } |
1da177e4 | 362 | |
1da177e4 LT |
363 | /** |
364 | * Remove a map private from list and deallocate resources if the mapping | |
365 | * isn't in use. | |
366 | * | |
367 | * \param inode device inode. | |
368 | * \param filp file pointer. | |
369 | * \param cmd command. | |
370 | * \param arg pointer to a drm_map_t structure. | |
371 | * \return zero on success or a negative value on error. | |
372 | * | |
373 | * Searches the map on drm_device::maplist, removes it from the list, see if | |
374 | * its being used, and free any associate resource (such as MTRR's) if it's not | |
375 | * being on use. | |
376 | * | |
7ab98401 | 377 | * \sa drm_addmap |
1da177e4 | 378 | */ |
d985c108 | 379 | int drm_rmmap_locked(drm_device_t *dev, drm_local_map_t *map) |
1da177e4 | 380 | { |
bd1b331f | 381 | drm_map_list_t *r_list = NULL, *list_t; |
836cf046 | 382 | drm_dma_handle_t dmah; |
bd1b331f | 383 | int found = 0; |
1da177e4 | 384 | |
836cf046 | 385 | /* Find the list entry for the map and remove it */ |
bd1b331f | 386 | list_for_each_entry_safe(r_list, list_t, &dev->maplist, head) { |
836cf046 | 387 | if (r_list->map == map) { |
bd1b331f | 388 | list_del(&r_list->head); |
1545085a TH |
389 | drm_ht_remove_key(&dev->map_hash, |
390 | r_list->user_token >> PAGE_SHIFT); | |
bd1b331f DA |
391 | drm_free(r_list, sizeof(*r_list), DRM_MEM_MAPS); |
392 | found = 1; | |
836cf046 DA |
393 | break; |
394 | } | |
1da177e4 LT |
395 | } |
396 | ||
bd1b331f | 397 | if (!found) |
1da177e4 | 398 | return -EINVAL; |
1da177e4 | 399 | |
836cf046 DA |
400 | switch (map->type) { |
401 | case _DRM_REGISTERS: | |
004a7727 | 402 | iounmap(map->handle); |
836cf046 DA |
403 | /* FALLTHROUGH */ |
404 | case _DRM_FRAME_BUFFER: | |
405 | if (drm_core_has_MTRR(dev) && map->mtrr >= 0) { | |
406 | int retcode; | |
b5e89ed5 DA |
407 | retcode = mtrr_del(map->mtrr, map->offset, map->size); |
408 | DRM_DEBUG("mtrr_del=%d\n", retcode); | |
1da177e4 | 409 | } |
836cf046 DA |
410 | break; |
411 | case _DRM_SHM: | |
412 | vfree(map->handle); | |
413 | break; | |
414 | case _DRM_AGP: | |
415 | case _DRM_SCATTER_GATHER: | |
416 | break; | |
417 | case _DRM_CONSISTENT: | |
418 | dmah.vaddr = map->handle; | |
419 | dmah.busaddr = map->offset; | |
420 | dmah.size = map->size; | |
421 | __drm_pci_free(dev, &dmah); | |
422 | break; | |
1da177e4 | 423 | } |
836cf046 DA |
424 | drm_free(map, sizeof(*map), DRM_MEM_MAPS); |
425 | ||
1da177e4 LT |
426 | return 0; |
427 | } | |
836cf046 | 428 | |
d985c108 | 429 | int drm_rmmap(drm_device_t *dev, drm_local_map_t *map) |
836cf046 DA |
430 | { |
431 | int ret; | |
432 | ||
30e2fb18 | 433 | mutex_lock(&dev->struct_mutex); |
836cf046 | 434 | ret = drm_rmmap_locked(dev, map); |
30e2fb18 | 435 | mutex_unlock(&dev->struct_mutex); |
836cf046 DA |
436 | |
437 | return ret; | |
438 | } | |
7ab98401 | 439 | |
836cf046 DA |
440 | /* The rmmap ioctl appears to be unnecessary. All mappings are torn down on |
441 | * the last close of the device, and this is necessary for cleanup when things | |
442 | * exit uncleanly. Therefore, having userland manually remove mappings seems | |
443 | * like a pointless exercise since they're going away anyway. | |
444 | * | |
445 | * One use case might be after addmap is allowed for normal users for SHM and | |
446 | * gets used by drivers that the server doesn't need to care about. This seems | |
447 | * unlikely. | |
448 | */ | |
7ab98401 DA |
449 | int drm_rmmap_ioctl(struct inode *inode, struct file *filp, |
450 | unsigned int cmd, unsigned long arg) | |
451 | { | |
452 | drm_file_t *priv = filp->private_data; | |
453 | drm_device_t *dev = priv->head->dev; | |
454 | drm_map_t request; | |
836cf046 | 455 | drm_local_map_t *map = NULL; |
bd1b331f | 456 | drm_map_list_t *r_list; |
836cf046 | 457 | int ret; |
7ab98401 | 458 | |
b5e89ed5 | 459 | if (copy_from_user(&request, (drm_map_t __user *) arg, sizeof(request))) { |
7ab98401 DA |
460 | return -EFAULT; |
461 | } | |
462 | ||
30e2fb18 | 463 | mutex_lock(&dev->struct_mutex); |
bd1b331f | 464 | list_for_each_entry(r_list, &dev->maplist, head) { |
836cf046 | 465 | if (r_list->map && |
b5e89ed5 | 466 | r_list->user_token == (unsigned long)request.handle && |
836cf046 DA |
467 | r_list->map->flags & _DRM_REMOVABLE) { |
468 | map = r_list->map; | |
469 | break; | |
470 | } | |
471 | } | |
472 | ||
473 | /* List has wrapped around to the head pointer, or its empty we didn't | |
474 | * find anything. | |
475 | */ | |
bd1b331f | 476 | if (list_empty(&dev->maplist) || !map) { |
30e2fb18 | 477 | mutex_unlock(&dev->struct_mutex); |
836cf046 DA |
478 | return -EINVAL; |
479 | } | |
480 | ||
7a3f1f21 TH |
481 | if (!map) { |
482 | mutex_unlock(&dev->struct_mutex); | |
836cf046 | 483 | return -EINVAL; |
7a3f1f21 | 484 | } |
836cf046 DA |
485 | |
486 | /* Register and framebuffer maps are permanent */ | |
487 | if ((map->type == _DRM_REGISTERS) || (map->type == _DRM_FRAME_BUFFER)) { | |
30e2fb18 | 488 | mutex_unlock(&dev->struct_mutex); |
836cf046 DA |
489 | return 0; |
490 | } | |
491 | ||
492 | ret = drm_rmmap_locked(dev, map); | |
493 | ||
30e2fb18 | 494 | mutex_unlock(&dev->struct_mutex); |
836cf046 DA |
495 | |
496 | return ret; | |
7ab98401 | 497 | } |
1da177e4 LT |
498 | |
499 | /** | |
500 | * Cleanup after an error on one of the addbufs() functions. | |
501 | * | |
836cf046 | 502 | * \param dev DRM device. |
1da177e4 LT |
503 | * \param entry buffer entry where the error occurred. |
504 | * | |
505 | * Frees any pages and buffers associated with the given entry. | |
506 | */ | |
b5e89ed5 | 507 | static void drm_cleanup_buf_error(drm_device_t * dev, drm_buf_entry_t * entry) |
1da177e4 LT |
508 | { |
509 | int i; | |
510 | ||
511 | if (entry->seg_count) { | |
512 | for (i = 0; i < entry->seg_count; i++) { | |
513 | if (entry->seglist[i]) { | |
ddf19b97 | 514 | drm_pci_free(dev, entry->seglist[i]); |
1da177e4 LT |
515 | } |
516 | } | |
517 | drm_free(entry->seglist, | |
b5e89ed5 DA |
518 | entry->seg_count * |
519 | sizeof(*entry->seglist), DRM_MEM_SEGS); | |
1da177e4 LT |
520 | |
521 | entry->seg_count = 0; | |
522 | } | |
523 | ||
b5e89ed5 DA |
524 | if (entry->buf_count) { |
525 | for (i = 0; i < entry->buf_count; i++) { | |
1da177e4 LT |
526 | if (entry->buflist[i].dev_private) { |
527 | drm_free(entry->buflist[i].dev_private, | |
b5e89ed5 DA |
528 | entry->buflist[i].dev_priv_size, |
529 | DRM_MEM_BUFS); | |
1da177e4 LT |
530 | } |
531 | } | |
532 | drm_free(entry->buflist, | |
b5e89ed5 DA |
533 | entry->buf_count * |
534 | sizeof(*entry->buflist), DRM_MEM_BUFS); | |
1da177e4 LT |
535 | |
536 | entry->buf_count = 0; | |
537 | } | |
538 | } | |
539 | ||
540 | #if __OS_HAS_AGP | |
541 | /** | |
d59431bf | 542 | * Add AGP buffers for DMA transfers. |
1da177e4 | 543 | * |
d59431bf DA |
544 | * \param dev drm_device_t to which the buffers are to be added. |
545 | * \param request pointer to a drm_buf_desc_t describing the request. | |
1da177e4 | 546 | * \return zero on success or a negative number on failure. |
b5e89ed5 | 547 | * |
1da177e4 LT |
548 | * After some sanity checks creates a drm_buf structure for each buffer and |
549 | * reallocates the buffer list of the same size order to accommodate the new | |
550 | * buffers. | |
551 | */ | |
b5e89ed5 | 552 | int drm_addbufs_agp(drm_device_t * dev, drm_buf_desc_t * request) |
1da177e4 | 553 | { |
1da177e4 | 554 | drm_device_dma_t *dma = dev->dma; |
1da177e4 | 555 | drm_buf_entry_t *entry; |
54ba2f76 | 556 | drm_agp_mem_t *agp_entry; |
1da177e4 LT |
557 | drm_buf_t *buf; |
558 | unsigned long offset; | |
559 | unsigned long agp_offset; | |
560 | int count; | |
561 | int order; | |
562 | int size; | |
563 | int alignment; | |
564 | int page_order; | |
565 | int total; | |
566 | int byte_count; | |
54ba2f76 | 567 | int i, valid; |
1da177e4 | 568 | drm_buf_t **temp_buflist; |
1da177e4 | 569 | |
b5e89ed5 DA |
570 | if (!dma) |
571 | return -EINVAL; | |
1da177e4 | 572 | |
d59431bf DA |
573 | count = request->count; |
574 | order = drm_order(request->size); | |
1da177e4 LT |
575 | size = 1 << order; |
576 | ||
b5e89ed5 DA |
577 | alignment = (request->flags & _DRM_PAGE_ALIGN) |
578 | ? PAGE_ALIGN(size) : size; | |
1da177e4 LT |
579 | page_order = order - PAGE_SHIFT > 0 ? order - PAGE_SHIFT : 0; |
580 | total = PAGE_SIZE << page_order; | |
581 | ||
582 | byte_count = 0; | |
d59431bf | 583 | agp_offset = dev->agp->base + request->agp_start; |
1da177e4 | 584 | |
b5e89ed5 DA |
585 | DRM_DEBUG("count: %d\n", count); |
586 | DRM_DEBUG("order: %d\n", order); | |
587 | DRM_DEBUG("size: %d\n", size); | |
d985c108 | 588 | DRM_DEBUG("agp_offset: %lx\n", agp_offset); |
b5e89ed5 DA |
589 | DRM_DEBUG("alignment: %d\n", alignment); |
590 | DRM_DEBUG("page_order: %d\n", page_order); | |
591 | DRM_DEBUG("total: %d\n", total); | |
1da177e4 | 592 | |
b5e89ed5 DA |
593 | if (order < DRM_MIN_ORDER || order > DRM_MAX_ORDER) |
594 | return -EINVAL; | |
595 | if (dev->queue_count) | |
596 | return -EBUSY; /* Not while in use */ | |
1da177e4 | 597 | |
54ba2f76 DA |
598 | /* Make sure buffers are located in AGP memory that we own */ |
599 | valid = 0; | |
bd1b331f | 600 | list_for_each_entry(agp_entry, &dev->agp->memory, head) { |
54ba2f76 DA |
601 | if ((agp_offset >= agp_entry->bound) && |
602 | (agp_offset + total * count <= agp_entry->bound + agp_entry->pages * PAGE_SIZE)) { | |
603 | valid = 1; | |
604 | break; | |
605 | } | |
606 | } | |
bd1b331f | 607 | if (!list_empty(&dev->agp->memory) && !valid) { |
54ba2f76 DA |
608 | DRM_DEBUG("zone invalid\n"); |
609 | return -EINVAL; | |
610 | } | |
b5e89ed5 DA |
611 | spin_lock(&dev->count_lock); |
612 | if (dev->buf_use) { | |
613 | spin_unlock(&dev->count_lock); | |
1da177e4 LT |
614 | return -EBUSY; |
615 | } | |
b5e89ed5 DA |
616 | atomic_inc(&dev->buf_alloc); |
617 | spin_unlock(&dev->count_lock); | |
1da177e4 | 618 | |
30e2fb18 | 619 | mutex_lock(&dev->struct_mutex); |
1da177e4 | 620 | entry = &dma->bufs[order]; |
b5e89ed5 | 621 | if (entry->buf_count) { |
30e2fb18 | 622 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 DA |
623 | atomic_dec(&dev->buf_alloc); |
624 | return -ENOMEM; /* May only call once for each order */ | |
1da177e4 LT |
625 | } |
626 | ||
627 | if (count < 0 || count > 4096) { | |
30e2fb18 | 628 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 629 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
630 | return -EINVAL; |
631 | } | |
632 | ||
b5e89ed5 DA |
633 | entry->buflist = drm_alloc(count * sizeof(*entry->buflist), |
634 | DRM_MEM_BUFS); | |
635 | if (!entry->buflist) { | |
30e2fb18 | 636 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 637 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
638 | return -ENOMEM; |
639 | } | |
b5e89ed5 | 640 | memset(entry->buflist, 0, count * sizeof(*entry->buflist)); |
1da177e4 LT |
641 | |
642 | entry->buf_size = size; | |
643 | entry->page_order = page_order; | |
644 | ||
645 | offset = 0; | |
646 | ||
b5e89ed5 DA |
647 | while (entry->buf_count < count) { |
648 | buf = &entry->buflist[entry->buf_count]; | |
649 | buf->idx = dma->buf_count + entry->buf_count; | |
650 | buf->total = alignment; | |
651 | buf->order = order; | |
652 | buf->used = 0; | |
1da177e4 | 653 | |
b5e89ed5 | 654 | buf->offset = (dma->byte_count + offset); |
1da177e4 LT |
655 | buf->bus_address = agp_offset + offset; |
656 | buf->address = (void *)(agp_offset + offset); | |
b5e89ed5 | 657 | buf->next = NULL; |
1da177e4 LT |
658 | buf->waiting = 0; |
659 | buf->pending = 0; | |
b5e89ed5 DA |
660 | init_waitqueue_head(&buf->dma_wait); |
661 | buf->filp = NULL; | |
1da177e4 LT |
662 | |
663 | buf->dev_priv_size = dev->driver->dev_priv_size; | |
b5e89ed5 DA |
664 | buf->dev_private = drm_alloc(buf->dev_priv_size, DRM_MEM_BUFS); |
665 | if (!buf->dev_private) { | |
1da177e4 LT |
666 | /* Set count correctly so we free the proper amount. */ |
667 | entry->buf_count = count; | |
b5e89ed5 | 668 | drm_cleanup_buf_error(dev, entry); |
30e2fb18 | 669 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 670 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
671 | return -ENOMEM; |
672 | } | |
b5e89ed5 | 673 | memset(buf->dev_private, 0, buf->dev_priv_size); |
1da177e4 | 674 | |
b5e89ed5 | 675 | DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address); |
1da177e4 LT |
676 | |
677 | offset += alignment; | |
678 | entry->buf_count++; | |
679 | byte_count += PAGE_SIZE << page_order; | |
680 | } | |
681 | ||
b5e89ed5 | 682 | DRM_DEBUG("byte_count: %d\n", byte_count); |
1da177e4 | 683 | |
b5e89ed5 DA |
684 | temp_buflist = drm_realloc(dma->buflist, |
685 | dma->buf_count * sizeof(*dma->buflist), | |
686 | (dma->buf_count + entry->buf_count) | |
687 | * sizeof(*dma->buflist), DRM_MEM_BUFS); | |
688 | if (!temp_buflist) { | |
1da177e4 | 689 | /* Free the entry because it isn't valid */ |
b5e89ed5 | 690 | drm_cleanup_buf_error(dev, entry); |
30e2fb18 | 691 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 692 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
693 | return -ENOMEM; |
694 | } | |
695 | dma->buflist = temp_buflist; | |
696 | ||
b5e89ed5 | 697 | for (i = 0; i < entry->buf_count; i++) { |
1da177e4 LT |
698 | dma->buflist[i + dma->buf_count] = &entry->buflist[i]; |
699 | } | |
700 | ||
701 | dma->buf_count += entry->buf_count; | |
d985c108 DA |
702 | dma->seg_count += entry->seg_count; |
703 | dma->page_count += byte_count >> PAGE_SHIFT; | |
1da177e4 LT |
704 | dma->byte_count += byte_count; |
705 | ||
b5e89ed5 DA |
706 | DRM_DEBUG("dma->buf_count : %d\n", dma->buf_count); |
707 | DRM_DEBUG("entry->buf_count : %d\n", entry->buf_count); | |
1da177e4 | 708 | |
30e2fb18 | 709 | mutex_unlock(&dev->struct_mutex); |
1da177e4 | 710 | |
d59431bf DA |
711 | request->count = entry->buf_count; |
712 | request->size = size; | |
1da177e4 LT |
713 | |
714 | dma->flags = _DRM_DMA_USE_AGP; | |
715 | ||
b5e89ed5 | 716 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
717 | return 0; |
718 | } | |
d84f76d3 | 719 | EXPORT_SYMBOL(drm_addbufs_agp); |
b5e89ed5 | 720 | #endif /* __OS_HAS_AGP */ |
1da177e4 | 721 | |
b5e89ed5 | 722 | int drm_addbufs_pci(drm_device_t * dev, drm_buf_desc_t * request) |
1da177e4 | 723 | { |
1da177e4 | 724 | drm_device_dma_t *dma = dev->dma; |
1da177e4 LT |
725 | int count; |
726 | int order; | |
727 | int size; | |
728 | int total; | |
729 | int page_order; | |
730 | drm_buf_entry_t *entry; | |
ddf19b97 | 731 | drm_dma_handle_t *dmah; |
1da177e4 LT |
732 | drm_buf_t *buf; |
733 | int alignment; | |
734 | unsigned long offset; | |
735 | int i; | |
736 | int byte_count; | |
737 | int page_count; | |
738 | unsigned long *temp_pagelist; | |
739 | drm_buf_t **temp_buflist; | |
1da177e4 | 740 | |
b5e89ed5 DA |
741 | if (!drm_core_check_feature(dev, DRIVER_PCI_DMA)) |
742 | return -EINVAL; | |
d985c108 | 743 | |
b5e89ed5 DA |
744 | if (!dma) |
745 | return -EINVAL; | |
1da177e4 | 746 | |
d985c108 DA |
747 | if (!capable(CAP_SYS_ADMIN)) |
748 | return -EPERM; | |
749 | ||
d59431bf DA |
750 | count = request->count; |
751 | order = drm_order(request->size); | |
1da177e4 LT |
752 | size = 1 << order; |
753 | ||
b5e89ed5 DA |
754 | DRM_DEBUG("count=%d, size=%d (%d), order=%d, queue_count=%d\n", |
755 | request->count, request->size, size, order, dev->queue_count); | |
1da177e4 | 756 | |
b5e89ed5 DA |
757 | if (order < DRM_MIN_ORDER || order > DRM_MAX_ORDER) |
758 | return -EINVAL; | |
759 | if (dev->queue_count) | |
760 | return -EBUSY; /* Not while in use */ | |
1da177e4 | 761 | |
d59431bf | 762 | alignment = (request->flags & _DRM_PAGE_ALIGN) |
b5e89ed5 | 763 | ? PAGE_ALIGN(size) : size; |
1da177e4 LT |
764 | page_order = order - PAGE_SHIFT > 0 ? order - PAGE_SHIFT : 0; |
765 | total = PAGE_SIZE << page_order; | |
766 | ||
b5e89ed5 DA |
767 | spin_lock(&dev->count_lock); |
768 | if (dev->buf_use) { | |
769 | spin_unlock(&dev->count_lock); | |
1da177e4 LT |
770 | return -EBUSY; |
771 | } | |
b5e89ed5 DA |
772 | atomic_inc(&dev->buf_alloc); |
773 | spin_unlock(&dev->count_lock); | |
1da177e4 | 774 | |
30e2fb18 | 775 | mutex_lock(&dev->struct_mutex); |
1da177e4 | 776 | entry = &dma->bufs[order]; |
b5e89ed5 | 777 | if (entry->buf_count) { |
30e2fb18 | 778 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 779 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
780 | return -ENOMEM; /* May only call once for each order */ |
781 | } | |
782 | ||
783 | if (count < 0 || count > 4096) { | |
30e2fb18 | 784 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 785 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
786 | return -EINVAL; |
787 | } | |
788 | ||
b5e89ed5 DA |
789 | entry->buflist = drm_alloc(count * sizeof(*entry->buflist), |
790 | DRM_MEM_BUFS); | |
791 | if (!entry->buflist) { | |
30e2fb18 | 792 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 793 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
794 | return -ENOMEM; |
795 | } | |
b5e89ed5 DA |
796 | memset(entry->buflist, 0, count * sizeof(*entry->buflist)); |
797 | ||
798 | entry->seglist = drm_alloc(count * sizeof(*entry->seglist), | |
799 | DRM_MEM_SEGS); | |
800 | if (!entry->seglist) { | |
801 | drm_free(entry->buflist, | |
802 | count * sizeof(*entry->buflist), DRM_MEM_BUFS); | |
30e2fb18 | 803 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 804 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
805 | return -ENOMEM; |
806 | } | |
b5e89ed5 | 807 | memset(entry->seglist, 0, count * sizeof(*entry->seglist)); |
1da177e4 LT |
808 | |
809 | /* Keep the original pagelist until we know all the allocations | |
810 | * have succeeded | |
811 | */ | |
b5e89ed5 DA |
812 | temp_pagelist = drm_alloc((dma->page_count + (count << page_order)) |
813 | * sizeof(*dma->pagelist), DRM_MEM_PAGES); | |
1da177e4 | 814 | if (!temp_pagelist) { |
b5e89ed5 DA |
815 | drm_free(entry->buflist, |
816 | count * sizeof(*entry->buflist), DRM_MEM_BUFS); | |
817 | drm_free(entry->seglist, | |
818 | count * sizeof(*entry->seglist), DRM_MEM_SEGS); | |
30e2fb18 | 819 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 820 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
821 | return -ENOMEM; |
822 | } | |
823 | memcpy(temp_pagelist, | |
b5e89ed5 DA |
824 | dma->pagelist, dma->page_count * sizeof(*dma->pagelist)); |
825 | DRM_DEBUG("pagelist: %d entries\n", | |
826 | dma->page_count + (count << page_order)); | |
1da177e4 | 827 | |
b5e89ed5 | 828 | entry->buf_size = size; |
1da177e4 LT |
829 | entry->page_order = page_order; |
830 | byte_count = 0; | |
831 | page_count = 0; | |
832 | ||
b5e89ed5 | 833 | while (entry->buf_count < count) { |
ddf19b97 DA |
834 | |
835 | dmah = drm_pci_alloc(dev, PAGE_SIZE << page_order, 0x1000, 0xfffffffful); | |
836 | ||
837 | if (!dmah) { | |
1da177e4 LT |
838 | /* Set count correctly so we free the proper amount. */ |
839 | entry->buf_count = count; | |
840 | entry->seg_count = count; | |
841 | drm_cleanup_buf_error(dev, entry); | |
b5e89ed5 DA |
842 | drm_free(temp_pagelist, |
843 | (dma->page_count + (count << page_order)) | |
844 | * sizeof(*dma->pagelist), DRM_MEM_PAGES); | |
30e2fb18 | 845 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 846 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
847 | return -ENOMEM; |
848 | } | |
ddf19b97 | 849 | entry->seglist[entry->seg_count++] = dmah; |
b5e89ed5 DA |
850 | for (i = 0; i < (1 << page_order); i++) { |
851 | DRM_DEBUG("page %d @ 0x%08lx\n", | |
852 | dma->page_count + page_count, | |
ddf19b97 | 853 | (unsigned long)dmah->vaddr + PAGE_SIZE * i); |
1da177e4 | 854 | temp_pagelist[dma->page_count + page_count++] |
ddf19b97 | 855 | = (unsigned long)dmah->vaddr + PAGE_SIZE * i; |
1da177e4 | 856 | } |
b5e89ed5 DA |
857 | for (offset = 0; |
858 | offset + size <= total && entry->buf_count < count; | |
859 | offset += alignment, ++entry->buf_count) { | |
860 | buf = &entry->buflist[entry->buf_count]; | |
861 | buf->idx = dma->buf_count + entry->buf_count; | |
862 | buf->total = alignment; | |
863 | buf->order = order; | |
864 | buf->used = 0; | |
865 | buf->offset = (dma->byte_count + byte_count + offset); | |
ddf19b97 DA |
866 | buf->address = (void *)(dmah->vaddr + offset); |
867 | buf->bus_address = dmah->busaddr + offset; | |
b5e89ed5 | 868 | buf->next = NULL; |
1da177e4 LT |
869 | buf->waiting = 0; |
870 | buf->pending = 0; | |
b5e89ed5 DA |
871 | init_waitqueue_head(&buf->dma_wait); |
872 | buf->filp = NULL; | |
1da177e4 LT |
873 | |
874 | buf->dev_priv_size = dev->driver->dev_priv_size; | |
b5e89ed5 DA |
875 | buf->dev_private = drm_alloc(buf->dev_priv_size, |
876 | DRM_MEM_BUFS); | |
877 | if (!buf->dev_private) { | |
1da177e4 LT |
878 | /* Set count correctly so we free the proper amount. */ |
879 | entry->buf_count = count; | |
880 | entry->seg_count = count; | |
b5e89ed5 DA |
881 | drm_cleanup_buf_error(dev, entry); |
882 | drm_free(temp_pagelist, | |
883 | (dma->page_count + | |
884 | (count << page_order)) | |
885 | * sizeof(*dma->pagelist), | |
886 | DRM_MEM_PAGES); | |
30e2fb18 | 887 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 888 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
889 | return -ENOMEM; |
890 | } | |
b5e89ed5 | 891 | memset(buf->dev_private, 0, buf->dev_priv_size); |
1da177e4 | 892 | |
b5e89ed5 DA |
893 | DRM_DEBUG("buffer %d @ %p\n", |
894 | entry->buf_count, buf->address); | |
1da177e4 LT |
895 | } |
896 | byte_count += PAGE_SIZE << page_order; | |
897 | } | |
898 | ||
b5e89ed5 DA |
899 | temp_buflist = drm_realloc(dma->buflist, |
900 | dma->buf_count * sizeof(*dma->buflist), | |
901 | (dma->buf_count + entry->buf_count) | |
902 | * sizeof(*dma->buflist), DRM_MEM_BUFS); | |
1da177e4 LT |
903 | if (!temp_buflist) { |
904 | /* Free the entry because it isn't valid */ | |
b5e89ed5 DA |
905 | drm_cleanup_buf_error(dev, entry); |
906 | drm_free(temp_pagelist, | |
907 | (dma->page_count + (count << page_order)) | |
908 | * sizeof(*dma->pagelist), DRM_MEM_PAGES); | |
30e2fb18 | 909 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 910 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
911 | return -ENOMEM; |
912 | } | |
913 | dma->buflist = temp_buflist; | |
914 | ||
b5e89ed5 | 915 | for (i = 0; i < entry->buf_count; i++) { |
1da177e4 LT |
916 | dma->buflist[i + dma->buf_count] = &entry->buflist[i]; |
917 | } | |
918 | ||
919 | /* No allocations failed, so now we can replace the orginal pagelist | |
920 | * with the new one. | |
921 | */ | |
922 | if (dma->page_count) { | |
923 | drm_free(dma->pagelist, | |
b5e89ed5 DA |
924 | dma->page_count * sizeof(*dma->pagelist), |
925 | DRM_MEM_PAGES); | |
1da177e4 LT |
926 | } |
927 | dma->pagelist = temp_pagelist; | |
928 | ||
929 | dma->buf_count += entry->buf_count; | |
930 | dma->seg_count += entry->seg_count; | |
931 | dma->page_count += entry->seg_count << page_order; | |
932 | dma->byte_count += PAGE_SIZE * (entry->seg_count << page_order); | |
933 | ||
30e2fb18 | 934 | mutex_unlock(&dev->struct_mutex); |
1da177e4 | 935 | |
d59431bf DA |
936 | request->count = entry->buf_count; |
937 | request->size = size; | |
1da177e4 | 938 | |
3417f33e GS |
939 | if (request->flags & _DRM_PCI_BUFFER_RO) |
940 | dma->flags = _DRM_DMA_USE_PCI_RO; | |
941 | ||
b5e89ed5 | 942 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
943 | return 0; |
944 | ||
945 | } | |
d84f76d3 | 946 | EXPORT_SYMBOL(drm_addbufs_pci); |
1da177e4 | 947 | |
b5e89ed5 | 948 | static int drm_addbufs_sg(drm_device_t * dev, drm_buf_desc_t * request) |
1da177e4 | 949 | { |
1da177e4 | 950 | drm_device_dma_t *dma = dev->dma; |
1da177e4 LT |
951 | drm_buf_entry_t *entry; |
952 | drm_buf_t *buf; | |
953 | unsigned long offset; | |
954 | unsigned long agp_offset; | |
955 | int count; | |
956 | int order; | |
957 | int size; | |
958 | int alignment; | |
959 | int page_order; | |
960 | int total; | |
961 | int byte_count; | |
962 | int i; | |
963 | drm_buf_t **temp_buflist; | |
964 | ||
b5e89ed5 DA |
965 | if (!drm_core_check_feature(dev, DRIVER_SG)) |
966 | return -EINVAL; | |
967 | ||
968 | if (!dma) | |
969 | return -EINVAL; | |
1da177e4 | 970 | |
d985c108 DA |
971 | if (!capable(CAP_SYS_ADMIN)) |
972 | return -EPERM; | |
973 | ||
d59431bf DA |
974 | count = request->count; |
975 | order = drm_order(request->size); | |
1da177e4 LT |
976 | size = 1 << order; |
977 | ||
b5e89ed5 DA |
978 | alignment = (request->flags & _DRM_PAGE_ALIGN) |
979 | ? PAGE_ALIGN(size) : size; | |
1da177e4 LT |
980 | page_order = order - PAGE_SHIFT > 0 ? order - PAGE_SHIFT : 0; |
981 | total = PAGE_SIZE << page_order; | |
982 | ||
983 | byte_count = 0; | |
d59431bf | 984 | agp_offset = request->agp_start; |
1da177e4 | 985 | |
b5e89ed5 DA |
986 | DRM_DEBUG("count: %d\n", count); |
987 | DRM_DEBUG("order: %d\n", order); | |
988 | DRM_DEBUG("size: %d\n", size); | |
989 | DRM_DEBUG("agp_offset: %lu\n", agp_offset); | |
990 | DRM_DEBUG("alignment: %d\n", alignment); | |
991 | DRM_DEBUG("page_order: %d\n", page_order); | |
992 | DRM_DEBUG("total: %d\n", total); | |
1da177e4 | 993 | |
b5e89ed5 DA |
994 | if (order < DRM_MIN_ORDER || order > DRM_MAX_ORDER) |
995 | return -EINVAL; | |
996 | if (dev->queue_count) | |
997 | return -EBUSY; /* Not while in use */ | |
1da177e4 | 998 | |
b5e89ed5 DA |
999 | spin_lock(&dev->count_lock); |
1000 | if (dev->buf_use) { | |
1001 | spin_unlock(&dev->count_lock); | |
1da177e4 LT |
1002 | return -EBUSY; |
1003 | } | |
b5e89ed5 DA |
1004 | atomic_inc(&dev->buf_alloc); |
1005 | spin_unlock(&dev->count_lock); | |
1da177e4 | 1006 | |
30e2fb18 | 1007 | mutex_lock(&dev->struct_mutex); |
1da177e4 | 1008 | entry = &dma->bufs[order]; |
b5e89ed5 | 1009 | if (entry->buf_count) { |
30e2fb18 | 1010 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 DA |
1011 | atomic_dec(&dev->buf_alloc); |
1012 | return -ENOMEM; /* May only call once for each order */ | |
1da177e4 LT |
1013 | } |
1014 | ||
1015 | if (count < 0 || count > 4096) { | |
30e2fb18 | 1016 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 1017 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
1018 | return -EINVAL; |
1019 | } | |
1020 | ||
b5e89ed5 DA |
1021 | entry->buflist = drm_alloc(count * sizeof(*entry->buflist), |
1022 | DRM_MEM_BUFS); | |
1023 | if (!entry->buflist) { | |
30e2fb18 | 1024 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 1025 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
1026 | return -ENOMEM; |
1027 | } | |
b5e89ed5 | 1028 | memset(entry->buflist, 0, count * sizeof(*entry->buflist)); |
1da177e4 LT |
1029 | |
1030 | entry->buf_size = size; | |
1031 | entry->page_order = page_order; | |
1032 | ||
1033 | offset = 0; | |
1034 | ||
b5e89ed5 DA |
1035 | while (entry->buf_count < count) { |
1036 | buf = &entry->buflist[entry->buf_count]; | |
1037 | buf->idx = dma->buf_count + entry->buf_count; | |
1038 | buf->total = alignment; | |
1039 | buf->order = order; | |
1040 | buf->used = 0; | |
1da177e4 | 1041 | |
b5e89ed5 | 1042 | buf->offset = (dma->byte_count + offset); |
1da177e4 | 1043 | buf->bus_address = agp_offset + offset; |
b5e89ed5 | 1044 | buf->address = (void *)(agp_offset + offset |
d1f2b55a | 1045 | + (unsigned long)dev->sg->virtual); |
b5e89ed5 | 1046 | buf->next = NULL; |
1da177e4 LT |
1047 | buf->waiting = 0; |
1048 | buf->pending = 0; | |
b5e89ed5 DA |
1049 | init_waitqueue_head(&buf->dma_wait); |
1050 | buf->filp = NULL; | |
1da177e4 LT |
1051 | |
1052 | buf->dev_priv_size = dev->driver->dev_priv_size; | |
b5e89ed5 DA |
1053 | buf->dev_private = drm_alloc(buf->dev_priv_size, DRM_MEM_BUFS); |
1054 | if (!buf->dev_private) { | |
1da177e4 LT |
1055 | /* Set count correctly so we free the proper amount. */ |
1056 | entry->buf_count = count; | |
b5e89ed5 | 1057 | drm_cleanup_buf_error(dev, entry); |
30e2fb18 | 1058 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 1059 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
1060 | return -ENOMEM; |
1061 | } | |
1062 | ||
b5e89ed5 | 1063 | memset(buf->dev_private, 0, buf->dev_priv_size); |
1da177e4 | 1064 | |
b5e89ed5 | 1065 | DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address); |
1da177e4 LT |
1066 | |
1067 | offset += alignment; | |
1068 | entry->buf_count++; | |
1069 | byte_count += PAGE_SIZE << page_order; | |
1070 | } | |
1071 | ||
b5e89ed5 | 1072 | DRM_DEBUG("byte_count: %d\n", byte_count); |
1da177e4 | 1073 | |
b5e89ed5 DA |
1074 | temp_buflist = drm_realloc(dma->buflist, |
1075 | dma->buf_count * sizeof(*dma->buflist), | |
1076 | (dma->buf_count + entry->buf_count) | |
1077 | * sizeof(*dma->buflist), DRM_MEM_BUFS); | |
1078 | if (!temp_buflist) { | |
1da177e4 | 1079 | /* Free the entry because it isn't valid */ |
b5e89ed5 | 1080 | drm_cleanup_buf_error(dev, entry); |
30e2fb18 | 1081 | mutex_unlock(&dev->struct_mutex); |
b5e89ed5 | 1082 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
1083 | return -ENOMEM; |
1084 | } | |
1085 | dma->buflist = temp_buflist; | |
1086 | ||
b5e89ed5 | 1087 | for (i = 0; i < entry->buf_count; i++) { |
1da177e4 LT |
1088 | dma->buflist[i + dma->buf_count] = &entry->buflist[i]; |
1089 | } | |
1090 | ||
1091 | dma->buf_count += entry->buf_count; | |
d985c108 DA |
1092 | dma->seg_count += entry->seg_count; |
1093 | dma->page_count += byte_count >> PAGE_SHIFT; | |
1da177e4 LT |
1094 | dma->byte_count += byte_count; |
1095 | ||
b5e89ed5 DA |
1096 | DRM_DEBUG("dma->buf_count : %d\n", dma->buf_count); |
1097 | DRM_DEBUG("entry->buf_count : %d\n", entry->buf_count); | |
1da177e4 | 1098 | |
30e2fb18 | 1099 | mutex_unlock(&dev->struct_mutex); |
1da177e4 | 1100 | |
d59431bf DA |
1101 | request->count = entry->buf_count; |
1102 | request->size = size; | |
1da177e4 LT |
1103 | |
1104 | dma->flags = _DRM_DMA_USE_SG; | |
1105 | ||
b5e89ed5 | 1106 | atomic_dec(&dev->buf_alloc); |
1da177e4 LT |
1107 | return 0; |
1108 | } | |
1109 | ||
5d23fafb | 1110 | static int drm_addbufs_fb(drm_device_t * dev, drm_buf_desc_t * request) |
b84397d6 | 1111 | { |
b84397d6 | 1112 | drm_device_dma_t *dma = dev->dma; |
b84397d6 DA |
1113 | drm_buf_entry_t *entry; |
1114 | drm_buf_t *buf; | |
1115 | unsigned long offset; | |
1116 | unsigned long agp_offset; | |
1117 | int count; | |
1118 | int order; | |
1119 | int size; | |
1120 | int alignment; | |
1121 | int page_order; | |
1122 | int total; | |
1123 | int byte_count; | |
1124 | int i; | |
1125 | drm_buf_t **temp_buflist; | |
b84397d6 DA |
1126 | |
1127 | if (!drm_core_check_feature(dev, DRIVER_FB_DMA)) | |
1128 | return -EINVAL; | |
b5e89ed5 | 1129 | |
b84397d6 DA |
1130 | if (!dma) |
1131 | return -EINVAL; | |
1132 | ||
d985c108 DA |
1133 | if (!capable(CAP_SYS_ADMIN)) |
1134 | return -EPERM; | |
1135 | ||
d59431bf DA |
1136 | count = request->count; |
1137 | order = drm_order(request->size); | |
b84397d6 DA |
1138 | size = 1 << order; |
1139 | ||
d59431bf | 1140 | alignment = (request->flags & _DRM_PAGE_ALIGN) |
b84397d6 DA |
1141 | ? PAGE_ALIGN(size) : size; |
1142 | page_order = order - PAGE_SHIFT > 0 ? order - PAGE_SHIFT : 0; | |
1143 | total = PAGE_SIZE << page_order; | |
1144 | ||
1145 | byte_count = 0; | |
d59431bf | 1146 | agp_offset = request->agp_start; |
b84397d6 DA |
1147 | |
1148 | DRM_DEBUG("count: %d\n", count); | |
1149 | DRM_DEBUG("order: %d\n", order); | |
1150 | DRM_DEBUG("size: %d\n", size); | |
1151 | DRM_DEBUG("agp_offset: %lu\n", agp_offset); | |
1152 | DRM_DEBUG("alignment: %d\n", alignment); | |
1153 | DRM_DEBUG("page_order: %d\n", page_order); | |
1154 | DRM_DEBUG("total: %d\n", total); | |
1155 | ||
1156 | if (order < DRM_MIN_ORDER || order > DRM_MAX_ORDER) | |
1157 | return -EINVAL; | |
1158 | if (dev->queue_count) | |
1159 | return -EBUSY; /* Not while in use */ | |
1160 | ||
1161 | spin_lock(&dev->count_lock); | |
1162 | if (dev->buf_use) { | |
1163 | spin_unlock(&dev->count_lock); | |
1164 | return -EBUSY; | |
1165 | } | |
1166 | atomic_inc(&dev->buf_alloc); | |
1167 | spin_unlock(&dev->count_lock); | |
1168 | ||
30e2fb18 | 1169 | mutex_lock(&dev->struct_mutex); |
b84397d6 DA |
1170 | entry = &dma->bufs[order]; |
1171 | if (entry->buf_count) { | |
30e2fb18 | 1172 | mutex_unlock(&dev->struct_mutex); |
b84397d6 DA |
1173 | atomic_dec(&dev->buf_alloc); |
1174 | return -ENOMEM; /* May only call once for each order */ | |
1175 | } | |
1176 | ||
1177 | if (count < 0 || count > 4096) { | |
30e2fb18 | 1178 | mutex_unlock(&dev->struct_mutex); |
b84397d6 DA |
1179 | atomic_dec(&dev->buf_alloc); |
1180 | return -EINVAL; | |
1181 | } | |
1182 | ||
1183 | entry->buflist = drm_alloc(count * sizeof(*entry->buflist), | |
1184 | DRM_MEM_BUFS); | |
1185 | if (!entry->buflist) { | |
30e2fb18 | 1186 | mutex_unlock(&dev->struct_mutex); |
b84397d6 DA |
1187 | atomic_dec(&dev->buf_alloc); |
1188 | return -ENOMEM; | |
1189 | } | |
1190 | memset(entry->buflist, 0, count * sizeof(*entry->buflist)); | |
1191 | ||
1192 | entry->buf_size = size; | |
1193 | entry->page_order = page_order; | |
1194 | ||
1195 | offset = 0; | |
1196 | ||
1197 | while (entry->buf_count < count) { | |
1198 | buf = &entry->buflist[entry->buf_count]; | |
1199 | buf->idx = dma->buf_count + entry->buf_count; | |
1200 | buf->total = alignment; | |
1201 | buf->order = order; | |
1202 | buf->used = 0; | |
1203 | ||
1204 | buf->offset = (dma->byte_count + offset); | |
1205 | buf->bus_address = agp_offset + offset; | |
1206 | buf->address = (void *)(agp_offset + offset); | |
1207 | buf->next = NULL; | |
1208 | buf->waiting = 0; | |
1209 | buf->pending = 0; | |
1210 | init_waitqueue_head(&buf->dma_wait); | |
1211 | buf->filp = NULL; | |
1212 | ||
1213 | buf->dev_priv_size = dev->driver->dev_priv_size; | |
1214 | buf->dev_private = drm_alloc(buf->dev_priv_size, DRM_MEM_BUFS); | |
1215 | if (!buf->dev_private) { | |
1216 | /* Set count correctly so we free the proper amount. */ | |
1217 | entry->buf_count = count; | |
1218 | drm_cleanup_buf_error(dev, entry); | |
30e2fb18 | 1219 | mutex_unlock(&dev->struct_mutex); |
b84397d6 DA |
1220 | atomic_dec(&dev->buf_alloc); |
1221 | return -ENOMEM; | |
1222 | } | |
1223 | memset(buf->dev_private, 0, buf->dev_priv_size); | |
1224 | ||
1225 | DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address); | |
1226 | ||
1227 | offset += alignment; | |
1228 | entry->buf_count++; | |
1229 | byte_count += PAGE_SIZE << page_order; | |
1230 | } | |
1231 | ||
1232 | DRM_DEBUG("byte_count: %d\n", byte_count); | |
1233 | ||
1234 | temp_buflist = drm_realloc(dma->buflist, | |
1235 | dma->buf_count * sizeof(*dma->buflist), | |
1236 | (dma->buf_count + entry->buf_count) | |
1237 | * sizeof(*dma->buflist), DRM_MEM_BUFS); | |
1238 | if (!temp_buflist) { | |
1239 | /* Free the entry because it isn't valid */ | |
1240 | drm_cleanup_buf_error(dev, entry); | |
30e2fb18 | 1241 | mutex_unlock(&dev->struct_mutex); |
b84397d6 DA |
1242 | atomic_dec(&dev->buf_alloc); |
1243 | return -ENOMEM; | |
1244 | } | |
1245 | dma->buflist = temp_buflist; | |
1246 | ||
1247 | for (i = 0; i < entry->buf_count; i++) { | |
1248 | dma->buflist[i + dma->buf_count] = &entry->buflist[i]; | |
1249 | } | |
1250 | ||
1251 | dma->buf_count += entry->buf_count; | |
d985c108 DA |
1252 | dma->seg_count += entry->seg_count; |
1253 | dma->page_count += byte_count >> PAGE_SHIFT; | |
b84397d6 DA |
1254 | dma->byte_count += byte_count; |
1255 | ||
1256 | DRM_DEBUG("dma->buf_count : %d\n", dma->buf_count); | |
1257 | DRM_DEBUG("entry->buf_count : %d\n", entry->buf_count); | |
1258 | ||
30e2fb18 | 1259 | mutex_unlock(&dev->struct_mutex); |
b84397d6 | 1260 | |
d59431bf DA |
1261 | request->count = entry->buf_count; |
1262 | request->size = size; | |
b84397d6 DA |
1263 | |
1264 | dma->flags = _DRM_DMA_USE_FB; | |
1265 | ||
1266 | atomic_dec(&dev->buf_alloc); | |
1267 | return 0; | |
1268 | } | |
d985c108 | 1269 | |
b84397d6 | 1270 | |
1da177e4 LT |
1271 | /** |
1272 | * Add buffers for DMA transfers (ioctl). | |
1273 | * | |
1274 | * \param inode device inode. | |
1275 | * \param filp file pointer. | |
1276 | * \param cmd command. | |
1277 | * \param arg pointer to a drm_buf_desc_t request. | |
1278 | * \return zero on success or a negative number on failure. | |
1279 | * | |
1280 | * According with the memory type specified in drm_buf_desc::flags and the | |
1281 | * build options, it dispatches the call either to addbufs_agp(), | |
1282 | * addbufs_sg() or addbufs_pci() for AGP, scatter-gather or consistent | |
1283 | * PCI memory respectively. | |
1284 | */ | |
b5e89ed5 DA |
1285 | int drm_addbufs(struct inode *inode, struct file *filp, |
1286 | unsigned int cmd, unsigned long arg) | |
1da177e4 LT |
1287 | { |
1288 | drm_buf_desc_t request; | |
1289 | drm_file_t *priv = filp->private_data; | |
1290 | drm_device_t *dev = priv->head->dev; | |
d59431bf | 1291 | int ret; |
b5e89ed5 | 1292 | |
1da177e4 LT |
1293 | if (!drm_core_check_feature(dev, DRIVER_HAVE_DMA)) |
1294 | return -EINVAL; | |
1295 | ||
b5e89ed5 DA |
1296 | if (copy_from_user(&request, (drm_buf_desc_t __user *) arg, |
1297 | sizeof(request))) | |
1da177e4 LT |
1298 | return -EFAULT; |
1299 | ||
1300 | #if __OS_HAS_AGP | |
b5e89ed5 DA |
1301 | if (request.flags & _DRM_AGP_BUFFER) |
1302 | ret = drm_addbufs_agp(dev, &request); | |
1da177e4 LT |
1303 | else |
1304 | #endif | |
b5e89ed5 DA |
1305 | if (request.flags & _DRM_SG_BUFFER) |
1306 | ret = drm_addbufs_sg(dev, &request); | |
1307 | else if (request.flags & _DRM_FB_BUFFER) | |
1308 | ret = drm_addbufs_fb(dev, &request); | |
1da177e4 | 1309 | else |
b5e89ed5 | 1310 | ret = drm_addbufs_pci(dev, &request); |
d59431bf | 1311 | |
b5e89ed5 DA |
1312 | if (ret == 0) { |
1313 | if (copy_to_user((void __user *)arg, &request, sizeof(request))) { | |
d59431bf DA |
1314 | ret = -EFAULT; |
1315 | } | |
1316 | } | |
1317 | return ret; | |
1da177e4 LT |
1318 | } |
1319 | ||
1da177e4 LT |
1320 | /** |
1321 | * Get information about the buffer mappings. | |
1322 | * | |
1323 | * This was originally mean for debugging purposes, or by a sophisticated | |
1324 | * client library to determine how best to use the available buffers (e.g., | |
1325 | * large buffers can be used for image transfer). | |
1326 | * | |
1327 | * \param inode device inode. | |
1328 | * \param filp file pointer. | |
1329 | * \param cmd command. | |
1330 | * \param arg pointer to a drm_buf_info structure. | |
1331 | * \return zero on success or a negative number on failure. | |
1332 | * | |
1333 | * Increments drm_device::buf_use while holding the drm_device::count_lock | |
1334 | * lock, preventing of allocating more buffers after this call. Information | |
1335 | * about each requested buffer is then copied into user space. | |
1336 | */ | |
b5e89ed5 DA |
1337 | int drm_infobufs(struct inode *inode, struct file *filp, |
1338 | unsigned int cmd, unsigned long arg) | |
1da177e4 LT |
1339 | { |
1340 | drm_file_t *priv = filp->private_data; | |
1341 | drm_device_t *dev = priv->head->dev; | |
1342 | drm_device_dma_t *dma = dev->dma; | |
1343 | drm_buf_info_t request; | |
1344 | drm_buf_info_t __user *argp = (void __user *)arg; | |
1345 | int i; | |
1346 | int count; | |
1347 | ||
1348 | if (!drm_core_check_feature(dev, DRIVER_HAVE_DMA)) | |
1349 | return -EINVAL; | |
1350 | ||
b5e89ed5 DA |
1351 | if (!dma) |
1352 | return -EINVAL; | |
1da177e4 | 1353 | |
b5e89ed5 DA |
1354 | spin_lock(&dev->count_lock); |
1355 | if (atomic_read(&dev->buf_alloc)) { | |
1356 | spin_unlock(&dev->count_lock); | |
1da177e4 LT |
1357 | return -EBUSY; |
1358 | } | |
1359 | ++dev->buf_use; /* Can't allocate more after this call */ | |
b5e89ed5 | 1360 | spin_unlock(&dev->count_lock); |
1da177e4 | 1361 | |
b5e89ed5 | 1362 | if (copy_from_user(&request, argp, sizeof(request))) |
1da177e4 LT |
1363 | return -EFAULT; |
1364 | ||
b5e89ed5 DA |
1365 | for (i = 0, count = 0; i < DRM_MAX_ORDER + 1; i++) { |
1366 | if (dma->bufs[i].buf_count) | |
1367 | ++count; | |
1da177e4 LT |
1368 | } |
1369 | ||
b5e89ed5 | 1370 | DRM_DEBUG("count = %d\n", count); |
1da177e4 | 1371 | |
b5e89ed5 DA |
1372 | if (request.count >= count) { |
1373 | for (i = 0, count = 0; i < DRM_MAX_ORDER + 1; i++) { | |
1374 | if (dma->bufs[i].buf_count) { | |
1375 | drm_buf_desc_t __user *to = | |
1376 | &request.list[count]; | |
1da177e4 LT |
1377 | drm_buf_entry_t *from = &dma->bufs[i]; |
1378 | drm_freelist_t *list = &dma->bufs[i].freelist; | |
b5e89ed5 DA |
1379 | if (copy_to_user(&to->count, |
1380 | &from->buf_count, | |
1381 | sizeof(from->buf_count)) || | |
1382 | copy_to_user(&to->size, | |
1383 | &from->buf_size, | |
1384 | sizeof(from->buf_size)) || | |
1385 | copy_to_user(&to->low_mark, | |
1386 | &list->low_mark, | |
1387 | sizeof(list->low_mark)) || | |
1388 | copy_to_user(&to->high_mark, | |
1389 | &list->high_mark, | |
1390 | sizeof(list->high_mark))) | |
1da177e4 LT |
1391 | return -EFAULT; |
1392 | ||
b5e89ed5 DA |
1393 | DRM_DEBUG("%d %d %d %d %d\n", |
1394 | i, | |
1395 | dma->bufs[i].buf_count, | |
1396 | dma->bufs[i].buf_size, | |
1397 | dma->bufs[i].freelist.low_mark, | |
1398 | dma->bufs[i].freelist.high_mark); | |
1da177e4 LT |
1399 | ++count; |
1400 | } | |
1401 | } | |
1402 | } | |
1403 | request.count = count; | |
1404 | ||
b5e89ed5 | 1405 | if (copy_to_user(argp, &request, sizeof(request))) |
1da177e4 LT |
1406 | return -EFAULT; |
1407 | ||
1408 | return 0; | |
1409 | } | |
1410 | ||
1411 | /** | |
1412 | * Specifies a low and high water mark for buffer allocation | |
1413 | * | |
1414 | * \param inode device inode. | |
1415 | * \param filp file pointer. | |
1416 | * \param cmd command. | |
1417 | * \param arg a pointer to a drm_buf_desc structure. | |
1418 | * \return zero on success or a negative number on failure. | |
1419 | * | |
1420 | * Verifies that the size order is bounded between the admissible orders and | |
1421 | * updates the respective drm_device_dma::bufs entry low and high water mark. | |
1422 | * | |
1423 | * \note This ioctl is deprecated and mostly never used. | |
1424 | */ | |
b5e89ed5 DA |
1425 | int drm_markbufs(struct inode *inode, struct file *filp, |
1426 | unsigned int cmd, unsigned long arg) | |
1da177e4 LT |
1427 | { |
1428 | drm_file_t *priv = filp->private_data; | |
1429 | drm_device_t *dev = priv->head->dev; | |
1430 | drm_device_dma_t *dma = dev->dma; | |
1431 | drm_buf_desc_t request; | |
1432 | int order; | |
1433 | drm_buf_entry_t *entry; | |
1434 | ||
1435 | if (!drm_core_check_feature(dev, DRIVER_HAVE_DMA)) | |
1436 | return -EINVAL; | |
1437 | ||
b5e89ed5 DA |
1438 | if (!dma) |
1439 | return -EINVAL; | |
1da177e4 | 1440 | |
b5e89ed5 DA |
1441 | if (copy_from_user(&request, |
1442 | (drm_buf_desc_t __user *) arg, sizeof(request))) | |
1da177e4 LT |
1443 | return -EFAULT; |
1444 | ||
b5e89ed5 DA |
1445 | DRM_DEBUG("%d, %d, %d\n", |
1446 | request.size, request.low_mark, request.high_mark); | |
1447 | order = drm_order(request.size); | |
1448 | if (order < DRM_MIN_ORDER || order > DRM_MAX_ORDER) | |
1449 | return -EINVAL; | |
1da177e4 LT |
1450 | entry = &dma->bufs[order]; |
1451 | ||
b5e89ed5 | 1452 | if (request.low_mark < 0 || request.low_mark > entry->buf_count) |
1da177e4 | 1453 | return -EINVAL; |
b5e89ed5 | 1454 | if (request.high_mark < 0 || request.high_mark > entry->buf_count) |
1da177e4 LT |
1455 | return -EINVAL; |
1456 | ||
b5e89ed5 | 1457 | entry->freelist.low_mark = request.low_mark; |
1da177e4 LT |
1458 | entry->freelist.high_mark = request.high_mark; |
1459 | ||
1460 | return 0; | |
1461 | } | |
1462 | ||
1463 | /** | |
b5e89ed5 | 1464 | * Unreserve the buffers in list, previously reserved using drmDMA. |
1da177e4 LT |
1465 | * |
1466 | * \param inode device inode. | |
1467 | * \param filp file pointer. | |
1468 | * \param cmd command. | |
1469 | * \param arg pointer to a drm_buf_free structure. | |
1470 | * \return zero on success or a negative number on failure. | |
b5e89ed5 | 1471 | * |
1da177e4 LT |
1472 | * Calls free_buffer() for each used buffer. |
1473 | * This function is primarily used for debugging. | |
1474 | */ | |
b5e89ed5 DA |
1475 | int drm_freebufs(struct inode *inode, struct file *filp, |
1476 | unsigned int cmd, unsigned long arg) | |
1da177e4 LT |
1477 | { |
1478 | drm_file_t *priv = filp->private_data; | |
1479 | drm_device_t *dev = priv->head->dev; | |
1480 | drm_device_dma_t *dma = dev->dma; | |
1481 | drm_buf_free_t request; | |
1482 | int i; | |
1483 | int idx; | |
1484 | drm_buf_t *buf; | |
1485 | ||
1486 | if (!drm_core_check_feature(dev, DRIVER_HAVE_DMA)) | |
1487 | return -EINVAL; | |
1488 | ||
b5e89ed5 DA |
1489 | if (!dma) |
1490 | return -EINVAL; | |
1da177e4 | 1491 | |
b5e89ed5 DA |
1492 | if (copy_from_user(&request, |
1493 | (drm_buf_free_t __user *) arg, sizeof(request))) | |
1da177e4 LT |
1494 | return -EFAULT; |
1495 | ||
b5e89ed5 DA |
1496 | DRM_DEBUG("%d\n", request.count); |
1497 | for (i = 0; i < request.count; i++) { | |
1498 | if (copy_from_user(&idx, &request.list[i], sizeof(idx))) | |
1da177e4 | 1499 | return -EFAULT; |
b5e89ed5 DA |
1500 | if (idx < 0 || idx >= dma->buf_count) { |
1501 | DRM_ERROR("Index %d (of %d max)\n", | |
1502 | idx, dma->buf_count - 1); | |
1da177e4 LT |
1503 | return -EINVAL; |
1504 | } | |
1505 | buf = dma->buflist[idx]; | |
b5e89ed5 DA |
1506 | if (buf->filp != filp) { |
1507 | DRM_ERROR("Process %d freeing buffer not owned\n", | |
1508 | current->pid); | |
1da177e4 LT |
1509 | return -EINVAL; |
1510 | } | |
b5e89ed5 | 1511 | drm_free_buffer(dev, buf); |
1da177e4 LT |
1512 | } |
1513 | ||
1514 | return 0; | |
1515 | } | |
1516 | ||
1517 | /** | |
1518 | * Maps all of the DMA buffers into client-virtual space (ioctl). | |
1519 | * | |
1520 | * \param inode device inode. | |
1521 | * \param filp file pointer. | |
1522 | * \param cmd command. | |
1523 | * \param arg pointer to a drm_buf_map structure. | |
1524 | * \return zero on success or a negative number on failure. | |
1525 | * | |
3417f33e GS |
1526 | * Maps the AGP, SG or PCI buffer region with do_mmap(), and copies information |
1527 | * about each buffer into user space. For PCI buffers, it calls do_mmap() with | |
1528 | * offset equal to 0, which drm_mmap() interpretes as PCI buffers and calls | |
1529 | * drm_mmap_dma(). | |
1da177e4 | 1530 | */ |
b5e89ed5 DA |
1531 | int drm_mapbufs(struct inode *inode, struct file *filp, |
1532 | unsigned int cmd, unsigned long arg) | |
1da177e4 LT |
1533 | { |
1534 | drm_file_t *priv = filp->private_data; | |
1535 | drm_device_t *dev = priv->head->dev; | |
1536 | drm_device_dma_t *dma = dev->dma; | |
1537 | drm_buf_map_t __user *argp = (void __user *)arg; | |
1538 | int retcode = 0; | |
1539 | const int zero = 0; | |
1540 | unsigned long virtual; | |
1541 | unsigned long address; | |
1542 | drm_buf_map_t request; | |
1543 | int i; | |
1544 | ||
1545 | if (!drm_core_check_feature(dev, DRIVER_HAVE_DMA)) | |
1546 | return -EINVAL; | |
1547 | ||
b5e89ed5 DA |
1548 | if (!dma) |
1549 | return -EINVAL; | |
1da177e4 | 1550 | |
b5e89ed5 DA |
1551 | spin_lock(&dev->count_lock); |
1552 | if (atomic_read(&dev->buf_alloc)) { | |
1553 | spin_unlock(&dev->count_lock); | |
1da177e4 LT |
1554 | return -EBUSY; |
1555 | } | |
1556 | dev->buf_use++; /* Can't allocate more after this call */ | |
b5e89ed5 | 1557 | spin_unlock(&dev->count_lock); |
1da177e4 | 1558 | |
b5e89ed5 | 1559 | if (copy_from_user(&request, argp, sizeof(request))) |
1da177e4 LT |
1560 | return -EFAULT; |
1561 | ||
b5e89ed5 | 1562 | if (request.count >= dma->buf_count) { |
b84397d6 | 1563 | if ((drm_core_has_AGP(dev) && (dma->flags & _DRM_DMA_USE_AGP)) |
b5e89ed5 | 1564 | || (drm_core_check_feature(dev, DRIVER_SG) |
b84397d6 DA |
1565 | && (dma->flags & _DRM_DMA_USE_SG)) |
1566 | || (drm_core_check_feature(dev, DRIVER_FB_DMA) | |
1567 | && (dma->flags & _DRM_DMA_USE_FB))) { | |
1da177e4 | 1568 | drm_map_t *map = dev->agp_buffer_map; |
d1f2b55a | 1569 | unsigned long token = dev->agp_buffer_token; |
1da177e4 | 1570 | |
b5e89ed5 | 1571 | if (!map) { |
1da177e4 LT |
1572 | retcode = -EINVAL; |
1573 | goto done; | |
1574 | } | |
1575 | ||
b5e89ed5 DA |
1576 | down_write(¤t->mm->mmap_sem); |
1577 | virtual = do_mmap(filp, 0, map->size, | |
1578 | PROT_READ | PROT_WRITE, | |
1579 | MAP_SHARED, token); | |
1580 | up_write(¤t->mm->mmap_sem); | |
1da177e4 | 1581 | } else { |
b5e89ed5 DA |
1582 | down_write(¤t->mm->mmap_sem); |
1583 | virtual = do_mmap(filp, 0, dma->byte_count, | |
1584 | PROT_READ | PROT_WRITE, | |
1585 | MAP_SHARED, 0); | |
1586 | up_write(¤t->mm->mmap_sem); | |
1da177e4 | 1587 | } |
b5e89ed5 | 1588 | if (virtual > -1024UL) { |
1da177e4 LT |
1589 | /* Real error */ |
1590 | retcode = (signed long)virtual; | |
1591 | goto done; | |
1592 | } | |
1593 | request.virtual = (void __user *)virtual; | |
1594 | ||
b5e89ed5 DA |
1595 | for (i = 0; i < dma->buf_count; i++) { |
1596 | if (copy_to_user(&request.list[i].idx, | |
1597 | &dma->buflist[i]->idx, | |
1598 | sizeof(request.list[0].idx))) { | |
1da177e4 LT |
1599 | retcode = -EFAULT; |
1600 | goto done; | |
1601 | } | |
b5e89ed5 DA |
1602 | if (copy_to_user(&request.list[i].total, |
1603 | &dma->buflist[i]->total, | |
1604 | sizeof(request.list[0].total))) { | |
1da177e4 LT |
1605 | retcode = -EFAULT; |
1606 | goto done; | |
1607 | } | |
b5e89ed5 DA |
1608 | if (copy_to_user(&request.list[i].used, |
1609 | &zero, sizeof(zero))) { | |
1da177e4 LT |
1610 | retcode = -EFAULT; |
1611 | goto done; | |
1612 | } | |
b5e89ed5 DA |
1613 | address = virtual + dma->buflist[i]->offset; /* *** */ |
1614 | if (copy_to_user(&request.list[i].address, | |
1615 | &address, sizeof(address))) { | |
1da177e4 LT |
1616 | retcode = -EFAULT; |
1617 | goto done; | |
1618 | } | |
1619 | } | |
1620 | } | |
b5e89ed5 | 1621 | done: |
1da177e4 | 1622 | request.count = dma->buf_count; |
b5e89ed5 | 1623 | DRM_DEBUG("%d buffers, retcode = %d\n", request.count, retcode); |
1da177e4 | 1624 | |
b5e89ed5 | 1625 | if (copy_to_user(argp, &request, sizeof(request))) |
1da177e4 LT |
1626 | return -EFAULT; |
1627 | ||
1628 | return retcode; | |
1629 | } | |
1630 | ||
836cf046 DA |
1631 | /** |
1632 | * Compute size order. Returns the exponent of the smaller power of two which | |
1633 | * is greater or equal to given number. | |
b5e89ed5 | 1634 | * |
836cf046 DA |
1635 | * \param size size. |
1636 | * \return order. | |
1637 | * | |
1638 | * \todo Can be made faster. | |
1639 | */ | |
b5e89ed5 | 1640 | int drm_order(unsigned long size) |
836cf046 DA |
1641 | { |
1642 | int order; | |
1643 | unsigned long tmp; | |
1644 | ||
b5e89ed5 | 1645 | for (order = 0, tmp = size >> 1; tmp; tmp >>= 1, order++) ; |
836cf046 DA |
1646 | |
1647 | if (size & (size - 1)) | |
1648 | ++order; | |
1649 | ||
1650 | return order; | |
1651 | } | |
1652 | EXPORT_SYMBOL(drm_order); | |
d985c108 DA |
1653 | |
1654 |