Commit | Line | Data |
---|---|---|
7a7d9a89 MCC |
1 | /* |
2 | * helper functions for PCI DMA video4linux capture buffers | |
3 | * | |
4 | * The functions expect the hardware being able to scatter gatter | |
5 | * (i.e. the buffers are not linear in physical memory, but fragmented | |
6 | * into PAGE_SIZE chunks). They also assume the driver does not need | |
7 | * to touch the video data. | |
8 | * | |
9 | * (c) 2007 Mauro Carvalho Chehab, <mchehab@infradead.org> | |
10 | * | |
11 | * Highly based on video-buf written originally by: | |
12 | * (c) 2001,02 Gerd Knorr <kraxel@bytesex.org> | |
13 | * (c) 2006 Mauro Carvalho Chehab, <mchehab@infradead.org> | |
14 | * (c) 2006 Ted Walther and John Sokol | |
15 | * | |
16 | * This program is free software; you can redistribute it and/or modify | |
17 | * it under the terms of the GNU General Public License as published by | |
18 | * the Free Software Foundation; either version 2 | |
19 | */ | |
20 | ||
21 | #include <linux/init.h> | |
22 | #include <linux/module.h> | |
23 | #include <linux/moduleparam.h> | |
24 | #include <linux/slab.h> | |
25 | #include <linux/interrupt.h> | |
26 | ||
27 | #include <linux/pci.h> | |
28 | #include <linux/vmalloc.h> | |
29 | #include <linux/pagemap.h> | |
30 | #include <asm/page.h> | |
31 | #include <asm/pgtable.h> | |
32 | ||
33 | #include <media/videobuf-dma-sg.h> | |
34 | ||
35 | #define MAGIC_DMABUF 0x19721112 | |
36 | #define MAGIC_SG_MEM 0x17890714 | |
37 | ||
38 | #define MAGIC_CHECK(is,should) if (unlikely((is) != (should))) \ | |
39 | { printk(KERN_ERR "magic mismatch: %x (expected %x)\n",is,should); BUG(); } | |
40 | ||
41 | static int debug = 0; | |
42 | module_param(debug, int, 0644); | |
43 | ||
44 | MODULE_DESCRIPTION("helper module to manage video4linux pci dma sg buffers"); | |
45 | MODULE_AUTHOR("Mauro Carvalho Chehab <mchehab@infradead.org>"); | |
46 | MODULE_LICENSE("GPL"); | |
47 | ||
48 | #define dprintk(level, fmt, arg...) if (debug >= level) \ | |
49 | printk(KERN_DEBUG "vbuf-sg: " fmt , ## arg) | |
50 | ||
51 | /* --------------------------------------------------------------------- */ | |
52 | ||
53 | struct scatterlist* | |
54 | videobuf_vmalloc_to_sg(unsigned char *virt, int nr_pages) | |
55 | { | |
56 | struct scatterlist *sglist; | |
57 | struct page *pg; | |
58 | int i; | |
59 | ||
60 | sglist = kcalloc(nr_pages, sizeof(struct scatterlist), GFP_KERNEL); | |
61 | if (NULL == sglist) | |
62 | return NULL; | |
63 | for (i = 0; i < nr_pages; i++, virt += PAGE_SIZE) { | |
64 | pg = vmalloc_to_page(virt); | |
65 | if (NULL == pg) | |
66 | goto err; | |
67 | BUG_ON(PageHighMem(pg)); | |
68 | sglist[i].page = pg; | |
69 | sglist[i].length = PAGE_SIZE; | |
70 | } | |
71 | return sglist; | |
72 | ||
73 | err: | |
74 | kfree(sglist); | |
75 | return NULL; | |
76 | } | |
77 | ||
78 | struct scatterlist* | |
79 | videobuf_pages_to_sg(struct page **pages, int nr_pages, int offset) | |
80 | { | |
81 | struct scatterlist *sglist; | |
82 | int i = 0; | |
83 | ||
84 | if (NULL == pages[0]) | |
85 | return NULL; | |
86 | sglist = kcalloc(nr_pages, sizeof(*sglist), GFP_KERNEL); | |
87 | if (NULL == sglist) | |
88 | return NULL; | |
89 | ||
90 | if (NULL == pages[0]) | |
91 | goto nopage; | |
92 | if (PageHighMem(pages[0])) | |
93 | /* DMA to highmem pages might not work */ | |
94 | goto highmem; | |
95 | sglist[0].page = pages[0]; | |
96 | sglist[0].offset = offset; | |
97 | sglist[0].length = PAGE_SIZE - offset; | |
98 | for (i = 1; i < nr_pages; i++) { | |
99 | if (NULL == pages[i]) | |
100 | goto nopage; | |
101 | if (PageHighMem(pages[i])) | |
102 | goto highmem; | |
103 | sglist[i].page = pages[i]; | |
104 | sglist[i].length = PAGE_SIZE; | |
105 | } | |
106 | return sglist; | |
107 | ||
108 | nopage: | |
109 | dprintk(2,"sgl: oops - no page\n"); | |
110 | kfree(sglist); | |
111 | return NULL; | |
112 | ||
113 | highmem: | |
114 | dprintk(2,"sgl: oops - highmem page\n"); | |
115 | kfree(sglist); | |
116 | return NULL; | |
117 | } | |
118 | ||
119 | /* --------------------------------------------------------------------- */ | |
120 | ||
121 | struct videobuf_dmabuf *videobuf_to_dma (struct videobuf_buffer *buf) | |
122 | { | |
123 | struct videbuf_pci_sg_memory *mem=buf->priv; | |
124 | BUG_ON (!mem); | |
125 | ||
126 | MAGIC_CHECK(mem->magic,MAGIC_SG_MEM); | |
127 | ||
128 | return &mem->dma; | |
129 | } | |
130 | ||
131 | void videobuf_dma_init(struct videobuf_dmabuf *dma) | |
132 | { | |
133 | memset(dma,0,sizeof(*dma)); | |
134 | dma->magic = MAGIC_DMABUF; | |
135 | } | |
136 | ||
9900132f ML |
137 | static int videobuf_dma_init_user_locked(struct videobuf_dmabuf *dma, |
138 | int direction, unsigned long data, unsigned long size) | |
7a7d9a89 MCC |
139 | { |
140 | unsigned long first,last; | |
141 | int err, rw = 0; | |
142 | ||
143 | dma->direction = direction; | |
144 | switch (dma->direction) { | |
145 | case PCI_DMA_FROMDEVICE: rw = READ; break; | |
146 | case PCI_DMA_TODEVICE: rw = WRITE; break; | |
147 | default: BUG(); | |
148 | } | |
149 | ||
150 | first = (data & PAGE_MASK) >> PAGE_SHIFT; | |
151 | last = ((data+size-1) & PAGE_MASK) >> PAGE_SHIFT; | |
152 | dma->offset = data & ~PAGE_MASK; | |
153 | dma->nr_pages = last-first+1; | |
154 | dma->pages = kmalloc(dma->nr_pages * sizeof(struct page*), | |
155 | GFP_KERNEL); | |
156 | if (NULL == dma->pages) | |
157 | return -ENOMEM; | |
158 | dprintk(1,"init user [0x%lx+0x%lx => %d pages]\n", | |
159 | data,size,dma->nr_pages); | |
160 | ||
161 | dma->varea = (void *) data; | |
162 | ||
9900132f | 163 | |
7a7d9a89 MCC |
164 | err = get_user_pages(current,current->mm, |
165 | data & PAGE_MASK, dma->nr_pages, | |
166 | rw == READ, 1, /* force */ | |
167 | dma->pages, NULL); | |
9900132f | 168 | |
7a7d9a89 MCC |
169 | if (err != dma->nr_pages) { |
170 | dma->nr_pages = (err >= 0) ? err : 0; | |
171 | dprintk(1,"get_user_pages: err=%d [%d]\n",err,dma->nr_pages); | |
172 | return err < 0 ? err : -EINVAL; | |
173 | } | |
174 | return 0; | |
175 | } | |
176 | ||
9900132f ML |
177 | int videobuf_dma_init_user(struct videobuf_dmabuf *dma, int direction, |
178 | unsigned long data, unsigned long size) | |
179 | { | |
180 | int ret; | |
181 | down_read(¤t->mm->mmap_sem); | |
182 | ret = videobuf_dma_init_user_locked(dma, direction, data, size); | |
183 | up_read(¤t->mm->mmap_sem); | |
184 | ||
185 | return ret; | |
186 | } | |
187 | ||
7a7d9a89 MCC |
188 | int videobuf_dma_init_kernel(struct videobuf_dmabuf *dma, int direction, |
189 | int nr_pages) | |
190 | { | |
191 | dprintk(1,"init kernel [%d pages]\n",nr_pages); | |
192 | dma->direction = direction; | |
193 | dma->vmalloc = vmalloc_32(nr_pages << PAGE_SHIFT); | |
194 | if (NULL == dma->vmalloc) { | |
195 | dprintk(1,"vmalloc_32(%d pages) failed\n",nr_pages); | |
196 | return -ENOMEM; | |
197 | } | |
198 | dprintk(1,"vmalloc is at addr 0x%08lx, size=%d\n", | |
199 | (unsigned long)dma->vmalloc, | |
200 | nr_pages << PAGE_SHIFT); | |
201 | memset(dma->vmalloc,0,nr_pages << PAGE_SHIFT); | |
202 | dma->nr_pages = nr_pages; | |
203 | return 0; | |
204 | } | |
205 | ||
206 | int videobuf_dma_init_overlay(struct videobuf_dmabuf *dma, int direction, | |
207 | dma_addr_t addr, int nr_pages) | |
208 | { | |
209 | dprintk(1,"init overlay [%d pages @ bus 0x%lx]\n", | |
210 | nr_pages,(unsigned long)addr); | |
211 | dma->direction = direction; | |
212 | if (0 == addr) | |
213 | return -EINVAL; | |
214 | ||
215 | dma->bus_addr = addr; | |
216 | dma->nr_pages = nr_pages; | |
217 | return 0; | |
218 | } | |
219 | ||
220 | int videobuf_dma_map(struct videobuf_queue* q,struct videobuf_dmabuf *dma) | |
221 | { | |
222 | void *dev=q->dev; | |
223 | struct videobuf_dma_sg_ops *ops=q->priv_ops; | |
224 | ||
225 | MAGIC_CHECK(dma->magic,MAGIC_DMABUF); | |
226 | BUG_ON(0 == dma->nr_pages); | |
227 | ||
228 | if (dma->pages) { | |
229 | dma->sglist = videobuf_pages_to_sg(dma->pages, dma->nr_pages, | |
230 | dma->offset); | |
231 | } | |
232 | if (dma->vmalloc) { | |
233 | dma->sglist = videobuf_vmalloc_to_sg | |
234 | (dma->vmalloc,dma->nr_pages); | |
235 | } | |
236 | if (dma->bus_addr) { | |
237 | dma->sglist = kmalloc(sizeof(struct scatterlist), GFP_KERNEL); | |
238 | if (NULL != dma->sglist) { | |
239 | dma->sglen = 1; | |
240 | sg_dma_address(&dma->sglist[0]) = dma->bus_addr & PAGE_MASK; | |
241 | dma->sglist[0].offset = dma->bus_addr & ~PAGE_MASK; | |
242 | sg_dma_len(&dma->sglist[0]) = dma->nr_pages * PAGE_SIZE; | |
243 | } | |
244 | } | |
245 | if (NULL == dma->sglist) { | |
246 | dprintk(1,"scatterlist is NULL\n"); | |
247 | return -ENOMEM; | |
248 | } | |
249 | if (!dma->bus_addr) { | |
250 | if (ops && ops->vb_map_sg) { | |
251 | dma->sglen = ops->vb_map_sg(dev,dma->sglist, | |
252 | dma->nr_pages, dma->direction); | |
253 | } | |
254 | if (0 == dma->sglen) { | |
255 | printk(KERN_WARNING | |
256 | "%s: videobuf_map_sg failed\n",__FUNCTION__); | |
257 | kfree(dma->sglist); | |
258 | dma->sglist = NULL; | |
259 | dma->sglen = 0; | |
260 | return -EIO; | |
261 | } | |
262 | } | |
263 | return 0; | |
264 | } | |
265 | ||
266 | int videobuf_dma_sync(struct videobuf_queue *q,struct videobuf_dmabuf *dma) | |
267 | { | |
268 | void *dev=q->dev; | |
269 | struct videobuf_dma_sg_ops *ops=q->priv_ops; | |
270 | ||
271 | MAGIC_CHECK(dma->magic,MAGIC_DMABUF); | |
272 | BUG_ON(!dma->sglen); | |
273 | ||
274 | if (!dma->bus_addr && ops && ops->vb_dma_sync_sg) | |
275 | ops->vb_dma_sync_sg(dev,dma->sglist,dma->nr_pages, | |
276 | dma->direction); | |
277 | ||
278 | return 0; | |
279 | } | |
280 | ||
281 | int videobuf_dma_unmap(struct videobuf_queue* q,struct videobuf_dmabuf *dma) | |
282 | { | |
283 | void *dev=q->dev; | |
284 | struct videobuf_dma_sg_ops *ops=q->priv_ops; | |
285 | ||
286 | MAGIC_CHECK(dma->magic,MAGIC_DMABUF); | |
287 | if (!dma->sglen) | |
288 | return 0; | |
289 | ||
290 | if (!dma->bus_addr && ops && ops->vb_unmap_sg) | |
291 | ops->vb_unmap_sg(dev,dma->sglist,dma->nr_pages, | |
292 | dma->direction); | |
293 | kfree(dma->sglist); | |
294 | dma->sglist = NULL; | |
295 | dma->sglen = 0; | |
296 | return 0; | |
297 | } | |
298 | ||
299 | int videobuf_dma_free(struct videobuf_dmabuf *dma) | |
300 | { | |
301 | MAGIC_CHECK(dma->magic,MAGIC_DMABUF); | |
302 | BUG_ON(dma->sglen); | |
303 | ||
304 | if (dma->pages) { | |
305 | int i; | |
306 | for (i=0; i < dma->nr_pages; i++) | |
307 | page_cache_release(dma->pages[i]); | |
308 | kfree(dma->pages); | |
309 | dma->pages = NULL; | |
310 | } | |
311 | ||
312 | vfree(dma->vmalloc); | |
313 | dma->vmalloc = NULL; | |
314 | dma->varea = NULL; | |
315 | ||
316 | if (dma->bus_addr) { | |
317 | dma->bus_addr = 0; | |
318 | } | |
319 | dma->direction = PCI_DMA_NONE; | |
320 | return 0; | |
321 | } | |
322 | ||
323 | /* --------------------------------------------------------------------- */ | |
324 | ||
325 | int videobuf_pci_dma_map(struct pci_dev *pci,struct videobuf_dmabuf *dma) | |
326 | { | |
327 | struct videobuf_queue q; | |
328 | struct videobuf_dma_sg_ops qops; | |
329 | ||
330 | q.dev=pci; | |
331 | qops.vb_map_sg=(vb_map_sg_t *)pci_map_sg; | |
332 | qops.vb_unmap_sg=(vb_map_sg_t *)pci_unmap_sg; | |
333 | q.priv_ops = &qops; | |
334 | ||
335 | return (videobuf_dma_map(&q,dma)); | |
336 | } | |
337 | ||
338 | int videobuf_pci_dma_unmap(struct pci_dev *pci,struct videobuf_dmabuf *dma) | |
339 | { | |
340 | struct videobuf_queue q; | |
341 | struct videobuf_dma_sg_ops qops; | |
342 | ||
343 | q.dev=pci; | |
344 | qops.vb_map_sg=(vb_map_sg_t *)pci_map_sg; | |
345 | qops.vb_unmap_sg=(vb_map_sg_t *)pci_unmap_sg; | |
346 | q.priv_ops = &qops; | |
347 | ||
348 | return (videobuf_dma_unmap(&q,dma)); | |
349 | } | |
350 | ||
351 | /* --------------------------------------------------------------------- */ | |
352 | ||
353 | static void | |
354 | videobuf_vm_open(struct vm_area_struct *vma) | |
355 | { | |
356 | struct videobuf_mapping *map = vma->vm_private_data; | |
357 | ||
358 | dprintk(2,"vm_open %p [count=%d,vma=%08lx-%08lx]\n",map, | |
359 | map->count,vma->vm_start,vma->vm_end); | |
360 | map->count++; | |
361 | } | |
362 | ||
363 | static void | |
364 | videobuf_vm_close(struct vm_area_struct *vma) | |
365 | { | |
366 | struct videobuf_mapping *map = vma->vm_private_data; | |
367 | struct videobuf_queue *q = map->q; | |
368 | struct videbuf_pci_sg_memory *mem; | |
369 | int i; | |
370 | ||
371 | dprintk(2,"vm_close %p [count=%d,vma=%08lx-%08lx]\n",map, | |
372 | map->count,vma->vm_start,vma->vm_end); | |
373 | ||
374 | map->count--; | |
375 | if (0 == map->count) { | |
376 | dprintk(1,"munmap %p q=%p\n",map,q); | |
377 | mutex_lock(&q->lock); | |
378 | for (i = 0; i < VIDEO_MAX_FRAME; i++) { | |
379 | if (NULL == q->bufs[i]) | |
380 | continue; | |
381 | mem=q->bufs[i]->priv; | |
382 | ||
383 | if (!mem) | |
384 | continue; | |
385 | ||
386 | MAGIC_CHECK(mem->magic,MAGIC_SG_MEM); | |
387 | ||
851c0c96 | 388 | if (q->bufs[i]->map != map) |
7a7d9a89 | 389 | continue; |
851c0c96 | 390 | q->bufs[i]->map = NULL; |
7a7d9a89 MCC |
391 | q->bufs[i]->baddr = 0; |
392 | q->ops->buf_release(q,q->bufs[i]); | |
393 | } | |
394 | mutex_unlock(&q->lock); | |
395 | kfree(map); | |
396 | } | |
397 | return; | |
398 | } | |
399 | ||
400 | /* | |
401 | * Get a anonymous page for the mapping. Make sure we can DMA to that | |
402 | * memory location with 32bit PCI devices (i.e. don't use highmem for | |
403 | * now ...). Bounce buffers don't work very well for the data rates | |
404 | * video capture has. | |
405 | */ | |
406 | static struct page* | |
407 | videobuf_vm_nopage(struct vm_area_struct *vma, unsigned long vaddr, | |
408 | int *type) | |
409 | { | |
410 | struct page *page; | |
411 | ||
412 | dprintk(3,"nopage: fault @ %08lx [vma %08lx-%08lx]\n", | |
413 | vaddr,vma->vm_start,vma->vm_end); | |
414 | if (vaddr > vma->vm_end) | |
415 | return NOPAGE_SIGBUS; | |
416 | page = alloc_page(GFP_USER | __GFP_DMA32); | |
417 | if (!page) | |
418 | return NOPAGE_OOM; | |
419 | clear_user_page(page_address(page), vaddr, page); | |
420 | if (type) | |
421 | *type = VM_FAULT_MINOR; | |
422 | return page; | |
423 | } | |
424 | ||
425 | static struct vm_operations_struct videobuf_vm_ops = | |
426 | { | |
427 | .open = videobuf_vm_open, | |
428 | .close = videobuf_vm_close, | |
429 | .nopage = videobuf_vm_nopage, | |
430 | }; | |
431 | ||
432 | /* --------------------------------------------------------------------- | |
433 | * PCI handlers for the generic methods | |
434 | */ | |
435 | ||
436 | /* Allocated area consists on 3 parts: | |
437 | struct video_buffer | |
438 | struct <driver>_buffer (cx88_buffer, saa7134_buf, ...) | |
439 | struct videobuf_pci_sg_memory | |
440 | */ | |
441 | ||
442 | static void *__videobuf_alloc(size_t size) | |
443 | { | |
444 | struct videbuf_pci_sg_memory *mem; | |
445 | struct videobuf_buffer *vb; | |
446 | ||
447 | vb = kzalloc(size+sizeof(*mem),GFP_KERNEL); | |
448 | ||
449 | mem = vb->priv = ((char *)vb)+size; | |
450 | mem->magic=MAGIC_SG_MEM; | |
451 | ||
452 | videobuf_dma_init(&mem->dma); | |
453 | ||
454 | dprintk(1,"%s: allocated at %p(%ld+%ld) & %p(%ld)\n", | |
455 | __FUNCTION__,vb,(long)sizeof(*vb),(long)size-sizeof(*vb), | |
456 | mem,(long)sizeof(*mem)); | |
457 | ||
458 | return vb; | |
459 | } | |
460 | ||
461 | static int __videobuf_iolock (struct videobuf_queue* q, | |
462 | struct videobuf_buffer *vb, | |
463 | struct v4l2_framebuffer *fbuf) | |
464 | { | |
465 | int err,pages; | |
466 | dma_addr_t bus; | |
467 | struct videbuf_pci_sg_memory *mem=vb->priv; | |
468 | BUG_ON(!mem); | |
469 | ||
470 | MAGIC_CHECK(mem->magic,MAGIC_SG_MEM); | |
471 | ||
472 | switch (vb->memory) { | |
473 | case V4L2_MEMORY_MMAP: | |
474 | case V4L2_MEMORY_USERPTR: | |
475 | if (0 == vb->baddr) { | |
476 | /* no userspace addr -- kernel bounce buffer */ | |
477 | pages = PAGE_ALIGN(vb->size) >> PAGE_SHIFT; | |
478 | err = videobuf_dma_init_kernel( &mem->dma, | |
479 | PCI_DMA_FROMDEVICE, | |
480 | pages ); | |
481 | if (0 != err) | |
482 | return err; | |
9900132f | 483 | } else if (vb->memory == V4L2_MEMORY_USERPTR) { |
7a7d9a89 MCC |
484 | /* dma directly to userspace */ |
485 | err = videobuf_dma_init_user( &mem->dma, | |
486 | PCI_DMA_FROMDEVICE, | |
487 | vb->baddr,vb->bsize ); | |
488 | if (0 != err) | |
489 | return err; | |
9900132f ML |
490 | } else { |
491 | /* NOTE: HACK: videobuf_iolock on V4L2_MEMORY_MMAP | |
492 | buffers can only be called from videobuf_qbuf | |
493 | we take current->mm->mmap_sem there, to prevent | |
494 | locking inversion, so don't take it here */ | |
495 | ||
496 | err = videobuf_dma_init_user_locked(&mem->dma, | |
497 | PCI_DMA_FROMDEVICE, | |
498 | vb->baddr, vb->bsize); | |
499 | if (0 != err) | |
500 | return err; | |
7a7d9a89 MCC |
501 | } |
502 | break; | |
503 | case V4L2_MEMORY_OVERLAY: | |
504 | if (NULL == fbuf) | |
505 | return -EINVAL; | |
506 | /* FIXME: need sanity checks for vb->boff */ | |
507 | /* | |
508 | * Using a double cast to avoid compiler warnings when | |
509 | * building for PAE. Compiler doesn't like direct casting | |
510 | * of a 32 bit ptr to 64 bit integer. | |
511 | */ | |
512 | bus = (dma_addr_t)(unsigned long)fbuf->base + vb->boff; | |
513 | pages = PAGE_ALIGN(vb->size) >> PAGE_SHIFT; | |
514 | err = videobuf_dma_init_overlay(&mem->dma,PCI_DMA_FROMDEVICE, | |
515 | bus, pages); | |
516 | if (0 != err) | |
517 | return err; | |
518 | break; | |
519 | default: | |
520 | BUG(); | |
521 | } | |
522 | err = videobuf_dma_map(q,&mem->dma); | |
523 | if (0 != err) | |
524 | return err; | |
525 | ||
526 | return 0; | |
527 | } | |
528 | ||
529 | static int __videobuf_sync(struct videobuf_queue *q, | |
530 | struct videobuf_buffer *buf) | |
531 | { | |
532 | struct videbuf_pci_sg_memory *mem=buf->priv; | |
533 | BUG_ON (!mem); | |
534 | MAGIC_CHECK(mem->magic,MAGIC_SG_MEM); | |
535 | ||
536 | return videobuf_dma_sync(q,&mem->dma); | |
537 | } | |
538 | ||
539 | static int __videobuf_mmap_free(struct videobuf_queue *q) | |
540 | { | |
541 | int i; | |
542 | ||
543 | for (i = 0; i < VIDEO_MAX_FRAME; i++) { | |
544 | if (q->bufs[i]) { | |
851c0c96 | 545 | if (q->bufs[i]->map) |
7a7d9a89 MCC |
546 | return -EBUSY; |
547 | } | |
548 | } | |
549 | ||
550 | return 0; | |
551 | } | |
552 | ||
553 | static int __videobuf_mmap_mapper(struct videobuf_queue *q, | |
554 | struct vm_area_struct *vma) | |
555 | { | |
556 | struct videbuf_pci_sg_memory *mem; | |
557 | struct videobuf_mapping *map; | |
558 | unsigned int first,last,size,i; | |
559 | int retval; | |
560 | ||
561 | retval = -EINVAL; | |
562 | if (!(vma->vm_flags & VM_WRITE)) { | |
563 | dprintk(1,"mmap app bug: PROT_WRITE please\n"); | |
564 | goto done; | |
565 | } | |
566 | if (!(vma->vm_flags & VM_SHARED)) { | |
567 | dprintk(1,"mmap app bug: MAP_SHARED please\n"); | |
568 | goto done; | |
569 | } | |
570 | ||
571 | /* look for first buffer to map */ | |
572 | for (first = 0; first < VIDEO_MAX_FRAME; first++) { | |
573 | if (NULL == q->bufs[first]) | |
574 | continue; | |
575 | mem=q->bufs[first]->priv; | |
576 | BUG_ON (!mem); | |
577 | MAGIC_CHECK(mem->magic,MAGIC_SG_MEM); | |
578 | ||
579 | if (V4L2_MEMORY_MMAP != q->bufs[first]->memory) | |
580 | continue; | |
581 | if (q->bufs[first]->boff == (vma->vm_pgoff << PAGE_SHIFT)) | |
582 | break; | |
583 | } | |
584 | if (VIDEO_MAX_FRAME == first) { | |
585 | dprintk(1,"mmap app bug: offset invalid [offset=0x%lx]\n", | |
586 | (vma->vm_pgoff << PAGE_SHIFT)); | |
587 | goto done; | |
588 | } | |
589 | ||
590 | /* look for last buffer to map */ | |
591 | for (size = 0, last = first; last < VIDEO_MAX_FRAME; last++) { | |
592 | if (NULL == q->bufs[last]) | |
593 | continue; | |
594 | if (V4L2_MEMORY_MMAP != q->bufs[last]->memory) | |
595 | continue; | |
851c0c96 | 596 | if (q->bufs[last]->map) { |
7a7d9a89 MCC |
597 | retval = -EBUSY; |
598 | goto done; | |
599 | } | |
600 | size += q->bufs[last]->bsize; | |
601 | if (size == (vma->vm_end - vma->vm_start)) | |
602 | break; | |
603 | } | |
604 | if (VIDEO_MAX_FRAME == last) { | |
605 | dprintk(1,"mmap app bug: size invalid [size=0x%lx]\n", | |
606 | (vma->vm_end - vma->vm_start)); | |
607 | goto done; | |
608 | } | |
609 | ||
610 | /* create mapping + update buffer list */ | |
611 | retval = -ENOMEM; | |
612 | map = kmalloc(sizeof(struct videobuf_mapping),GFP_KERNEL); | |
613 | if (NULL == map) | |
614 | goto done; | |
615 | for (size = 0, i = first; i <= last; size += q->bufs[i++]->bsize) { | |
851c0c96 | 616 | q->bufs[i]->map = map; |
7a7d9a89 MCC |
617 | q->bufs[i]->baddr = vma->vm_start + size; |
618 | } | |
619 | map->count = 1; | |
620 | map->start = vma->vm_start; | |
621 | map->end = vma->vm_end; | |
622 | map->q = q; | |
623 | vma->vm_ops = &videobuf_vm_ops; | |
624 | vma->vm_flags |= VM_DONTEXPAND | VM_RESERVED; | |
625 | vma->vm_flags &= ~VM_IO; /* using shared anonymous pages */ | |
626 | vma->vm_private_data = map; | |
627 | dprintk(1,"mmap %p: q=%p %08lx-%08lx pgoff %08lx bufs %d-%d\n", | |
628 | map,q,vma->vm_start,vma->vm_end,vma->vm_pgoff,first,last); | |
629 | retval = 0; | |
630 | ||
631 | done: | |
632 | return retval; | |
633 | } | |
634 | ||
7a7d9a89 MCC |
635 | static int __videobuf_copy_to_user ( struct videobuf_queue *q, |
636 | char __user *data, size_t count, | |
637 | int nonblocking ) | |
638 | { | |
639 | struct videbuf_pci_sg_memory *mem=q->read_buf->priv; | |
640 | BUG_ON (!mem); | |
641 | MAGIC_CHECK(mem->magic,MAGIC_SG_MEM); | |
642 | ||
643 | /* copy to userspace */ | |
644 | if (count > q->read_buf->size - q->read_off) | |
645 | count = q->read_buf->size - q->read_off; | |
646 | ||
647 | if (copy_to_user(data, mem->dma.vmalloc+q->read_off, count)) | |
648 | return -EFAULT; | |
649 | ||
650 | return count; | |
651 | } | |
652 | ||
653 | static int __videobuf_copy_stream ( struct videobuf_queue *q, | |
654 | char __user *data, size_t count, size_t pos, | |
655 | int vbihack, int nonblocking ) | |
656 | { | |
657 | unsigned int *fc; | |
658 | struct videbuf_pci_sg_memory *mem=q->read_buf->priv; | |
659 | BUG_ON (!mem); | |
660 | MAGIC_CHECK(mem->magic,MAGIC_SG_MEM); | |
661 | ||
662 | if (vbihack) { | |
663 | /* dirty, undocumented hack -- pass the frame counter | |
664 | * within the last four bytes of each vbi data block. | |
665 | * We need that one to maintain backward compatibility | |
666 | * to all vbi decoding software out there ... */ | |
667 | fc = (unsigned int*)mem->dma.vmalloc; | |
668 | fc += (q->read_buf->size>>2) -1; | |
669 | *fc = q->read_buf->field_count >> 1; | |
670 | dprintk(1,"vbihack: %d\n",*fc); | |
671 | } | |
672 | ||
673 | /* copy stuff using the common method */ | |
674 | count = __videobuf_copy_to_user (q,data,count,nonblocking); | |
675 | ||
676 | if ( (count==-EFAULT) && (0 == pos) ) | |
677 | return -EFAULT; | |
678 | ||
679 | return count; | |
680 | } | |
681 | ||
682 | static struct videobuf_qtype_ops pci_ops = { | |
683 | .magic = MAGIC_QTYPE_OPS, | |
684 | ||
685 | .alloc = __videobuf_alloc, | |
686 | .iolock = __videobuf_iolock, | |
687 | .sync = __videobuf_sync, | |
688 | .mmap_free = __videobuf_mmap_free, | |
689 | .mmap_mapper = __videobuf_mmap_mapper, | |
7a7d9a89 MCC |
690 | .copy_to_user = __videobuf_copy_to_user, |
691 | .copy_stream = __videobuf_copy_stream, | |
692 | }; | |
693 | ||
694 | void *videobuf_pci_alloc (size_t size) | |
695 | { | |
696 | struct videobuf_queue q; | |
697 | ||
698 | /* Required to make generic handler to call __videobuf_alloc */ | |
699 | q.int_ops=&pci_ops; | |
700 | ||
701 | q.msize=size; | |
702 | ||
703 | return videobuf_alloc (&q); | |
704 | } | |
705 | ||
706 | void videobuf_queue_pci_init(struct videobuf_queue* q, | |
707 | struct videobuf_queue_ops *ops, | |
708 | void *dev, | |
709 | spinlock_t *irqlock, | |
710 | enum v4l2_buf_type type, | |
711 | enum v4l2_field field, | |
712 | unsigned int msize, | |
713 | void *priv) | |
714 | { | |
715 | struct videobuf_dma_sg_ops *priv_ops; | |
716 | ||
717 | videobuf_queue_init(q, ops, dev, irqlock, type, field, msize, priv); | |
718 | q->int_ops=&pci_ops; | |
719 | ||
720 | /* FIXME: the code bellow should be removed after having a proper | |
721 | * memory allocation method for vivi and tm6000 | |
722 | */ | |
723 | q->priv_ops= kzalloc(sizeof(struct videobuf_dma_sg_ops), GFP_KERNEL); | |
724 | BUG_ON (!q->priv_ops); | |
725 | ||
726 | priv_ops=q->priv_ops; | |
727 | ||
728 | /* Sets default methods for handling Scatter Gather mapping */ | |
729 | priv_ops->vb_map_sg=(vb_map_sg_t *)pci_map_sg; | |
730 | priv_ops->vb_unmap_sg=(vb_map_sg_t *)pci_unmap_sg; | |
731 | priv_ops->vb_dma_sync_sg=(vb_map_sg_t *)pci_dma_sync_sg_for_cpu; | |
732 | } | |
733 | ||
734 | void videobuf_set_pci_ops (struct videobuf_queue* q, | |
735 | struct videobuf_dma_sg_ops *ops) | |
736 | { | |
737 | kfree (q->priv_ops); | |
738 | ||
739 | q->priv_ops=ops; | |
740 | ||
741 | if (!ops) | |
742 | return; | |
743 | ||
744 | /* If not specified, defaults to PCI map sg */ | |
745 | if (!ops->vb_map_sg) | |
746 | ops->vb_map_sg=(vb_map_sg_t *)pci_map_sg; | |
747 | ||
748 | if (!ops->vb_dma_sync_sg) | |
749 | ops->vb_dma_sync_sg=(vb_map_sg_t *)pci_dma_sync_sg_for_cpu; | |
750 | if (!ops->vb_unmap_sg) | |
751 | ops->vb_unmap_sg=(vb_map_sg_t *)pci_unmap_sg; | |
752 | } | |
753 | ||
754 | ||
755 | /* --------------------------------------------------------------------- */ | |
756 | ||
757 | EXPORT_SYMBOL_GPL(videobuf_vmalloc_to_sg); | |
758 | ||
759 | EXPORT_SYMBOL_GPL(videobuf_to_dma); | |
760 | EXPORT_SYMBOL_GPL(videobuf_dma_init); | |
761 | EXPORT_SYMBOL_GPL(videobuf_dma_init_user); | |
762 | EXPORT_SYMBOL_GPL(videobuf_dma_init_kernel); | |
763 | EXPORT_SYMBOL_GPL(videobuf_dma_init_overlay); | |
764 | EXPORT_SYMBOL_GPL(videobuf_dma_map); | |
765 | EXPORT_SYMBOL_GPL(videobuf_dma_sync); | |
766 | EXPORT_SYMBOL_GPL(videobuf_dma_unmap); | |
767 | EXPORT_SYMBOL_GPL(videobuf_dma_free); | |
768 | ||
769 | EXPORT_SYMBOL_GPL(videobuf_pci_dma_map); | |
770 | EXPORT_SYMBOL_GPL(videobuf_pci_dma_unmap); | |
771 | EXPORT_SYMBOL_GPL(videobuf_pci_alloc); | |
772 | ||
773 | EXPORT_SYMBOL_GPL(videobuf_queue_pci_init); | |
774 | EXPORT_SYMBOL_GPL(videobuf_set_pci_ops); | |
775 | ||
776 | /* | |
777 | * Local variables: | |
778 | * c-basic-offset: 8 | |
779 | * End: | |
780 | */ |