videobuf2-memops.c 6.0 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228
  1. /*
  2. * videobuf2-memops.c - generic memory handling routines for videobuf2
  3. *
  4. * Copyright (C) 2010 Samsung Electronics
  5. *
  6. * Author: Pawel Osciak <pawel@osciak.com>
  7. * Marek Szyprowski <m.szyprowski@samsung.com>
  8. *
  9. * This program is free software; you can redistribute it and/or modify
  10. * it under the terms of the GNU General Public License as published by
  11. * the Free Software Foundation.
  12. */
  13. #include <linux/slab.h>
  14. #include <linux/module.h>
  15. #include <linux/dma-mapping.h>
  16. #include <linux/vmalloc.h>
  17. #include <linux/mm.h>
  18. #include <linux/sched.h>
  19. #include <linux/file.h>
  20. #include <media/videobuf2-core.h>
  21. #include <media/videobuf2-memops.h>
  22. /**
  23. * vb2_get_vma() - acquire and lock the virtual memory area
  24. * @vma: given virtual memory area
  25. *
  26. * This function attempts to acquire an area mapped in the userspace for
  27. * the duration of a hardware operation. The area is "locked" by performing
  28. * the same set of operation that are done when process calls fork() and
  29. * memory areas are duplicated.
  30. *
  31. * Returns a copy of a virtual memory region on success or NULL.
  32. */
  33. struct vm_area_struct *vb2_get_vma(struct vm_area_struct *vma)
  34. {
  35. struct vm_area_struct *vma_copy;
  36. vma_copy = kmalloc(sizeof(*vma_copy), GFP_KERNEL);
  37. if (vma_copy == NULL)
  38. return NULL;
  39. if (vma->vm_ops && vma->vm_ops->open)
  40. vma->vm_ops->open(vma);
  41. if (vma->vm_file)
  42. get_file(vma->vm_file);
  43. memcpy(vma_copy, vma, sizeof(*vma));
  44. vma_copy->vm_mm = NULL;
  45. vma_copy->vm_next = NULL;
  46. vma_copy->vm_prev = NULL;
  47. return vma_copy;
  48. }
  49. EXPORT_SYMBOL_GPL(vb2_get_vma);
  50. /**
  51. * vb2_put_userptr() - release a userspace virtual memory area
  52. * @vma: virtual memory region associated with the area to be released
  53. *
  54. * This function releases the previously acquired memory area after a hardware
  55. * operation.
  56. */
  57. void vb2_put_vma(struct vm_area_struct *vma)
  58. {
  59. if (!vma)
  60. return;
  61. if (vma->vm_ops && vma->vm_ops->close)
  62. vma->vm_ops->close(vma);
  63. if (vma->vm_file)
  64. fput(vma->vm_file);
  65. kfree(vma);
  66. }
  67. EXPORT_SYMBOL_GPL(vb2_put_vma);
  68. /**
  69. * vb2_get_contig_userptr() - lock physically contiguous userspace mapped memory
  70. * @vaddr: starting virtual address of the area to be verified
  71. * @size: size of the area
  72. * @res_paddr: will return physical address for the given vaddr
  73. * @res_vma: will return locked copy of struct vm_area for the given area
  74. *
  75. * This function will go through memory area of size @size mapped at @vaddr and
  76. * verify that the underlying physical pages are contiguous. If they are
  77. * contiguous the virtual memory area is locked and a @res_vma is filled with
  78. * the copy and @res_pa set to the physical address of the buffer.
  79. *
  80. * Returns 0 on success.
  81. */
  82. int vb2_get_contig_userptr(unsigned long vaddr, unsigned long size,
  83. struct vm_area_struct **res_vma, dma_addr_t *res_pa)
  84. {
  85. struct mm_struct *mm = current->mm;
  86. struct vm_area_struct *vma;
  87. unsigned long offset, start, end;
  88. unsigned long this_pfn, prev_pfn;
  89. dma_addr_t pa = 0;
  90. start = vaddr;
  91. offset = start & ~PAGE_MASK;
  92. end = start + size;
  93. vma = find_vma(mm, start);
  94. if (vma == NULL || vma->vm_end < end)
  95. return -EFAULT;
  96. for (prev_pfn = 0; start < end; start += PAGE_SIZE) {
  97. int ret = follow_pfn(vma, start, &this_pfn);
  98. if (ret)
  99. return ret;
  100. if (prev_pfn == 0)
  101. pa = this_pfn << PAGE_SHIFT;
  102. else if (this_pfn != prev_pfn + 1)
  103. return -EFAULT;
  104. prev_pfn = this_pfn;
  105. }
  106. /*
  107. * Memory is contigous, lock vma and return to the caller
  108. */
  109. *res_vma = vb2_get_vma(vma);
  110. if (*res_vma == NULL)
  111. return -ENOMEM;
  112. *res_pa = pa + offset;
  113. return 0;
  114. }
  115. EXPORT_SYMBOL_GPL(vb2_get_contig_userptr);
  116. /**
  117. * vb2_mmap_pfn_range() - map physical pages to userspace
  118. * @vma: virtual memory region for the mapping
  119. * @paddr: starting physical address of the memory to be mapped
  120. * @size: size of the memory to be mapped
  121. * @vm_ops: vm operations to be assigned to the created area
  122. * @priv: private data to be associated with the area
  123. *
  124. * Returns 0 on success.
  125. */
  126. int vb2_mmap_pfn_range(struct vm_area_struct *vma, unsigned long paddr,
  127. unsigned long size,
  128. const struct vm_operations_struct *vm_ops,
  129. void *priv)
  130. {
  131. int ret;
  132. size = min_t(unsigned long, vma->vm_end - vma->vm_start, size);
  133. vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
  134. ret = remap_pfn_range(vma, vma->vm_start, paddr >> PAGE_SHIFT,
  135. size, vma->vm_page_prot);
  136. if (ret) {
  137. printk(KERN_ERR "Remapping memory failed, error: %d\n", ret);
  138. return ret;
  139. }
  140. vma->vm_flags |= VM_DONTEXPAND | VM_RESERVED;
  141. vma->vm_private_data = priv;
  142. vma->vm_ops = vm_ops;
  143. vma->vm_ops->open(vma);
  144. pr_debug("%s: mapped paddr 0x%08lx at 0x%08lx, size %ld\n",
  145. __func__, paddr, vma->vm_start, size);
  146. return 0;
  147. }
  148. EXPORT_SYMBOL_GPL(vb2_mmap_pfn_range);
  149. /**
  150. * vb2_common_vm_open() - increase refcount of the vma
  151. * @vma: virtual memory region for the mapping
  152. *
  153. * This function adds another user to the provided vma. It expects
  154. * struct vb2_vmarea_handler pointer in vma->vm_private_data.
  155. */
  156. static void vb2_common_vm_open(struct vm_area_struct *vma)
  157. {
  158. struct vb2_vmarea_handler *h = vma->vm_private_data;
  159. pr_debug("%s: %p, refcount: %d, vma: %08lx-%08lx\n",
  160. __func__, h, atomic_read(h->refcount), vma->vm_start,
  161. vma->vm_end);
  162. atomic_inc(h->refcount);
  163. }
  164. /**
  165. * vb2_common_vm_close() - decrease refcount of the vma
  166. * @vma: virtual memory region for the mapping
  167. *
  168. * This function releases the user from the provided vma. It expects
  169. * struct vb2_vmarea_handler pointer in vma->vm_private_data.
  170. */
  171. static void vb2_common_vm_close(struct vm_area_struct *vma)
  172. {
  173. struct vb2_vmarea_handler *h = vma->vm_private_data;
  174. pr_debug("%s: %p, refcount: %d, vma: %08lx-%08lx\n",
  175. __func__, h, atomic_read(h->refcount), vma->vm_start,
  176. vma->vm_end);
  177. h->put(h->arg);
  178. }
  179. /**
  180. * vb2_common_vm_ops - common vm_ops used for tracking refcount of mmaped
  181. * video buffers
  182. */
  183. const struct vm_operations_struct vb2_common_vm_ops = {
  184. .open = vb2_common_vm_open,
  185. .close = vb2_common_vm_close,
  186. };
  187. EXPORT_SYMBOL_GPL(vb2_common_vm_ops);
  188. MODULE_DESCRIPTION("common memory handling routines for videobuf2");
  189. MODULE_AUTHOR("Pawel Osciak <pawel@osciak.com>");
  190. MODULE_LICENSE("GPL");