diff options
author | Jonathan Herman <hermanjl@cs.unc.edu> | 2013-01-22 10:38:37 -0500 |
---|---|---|
committer | Jonathan Herman <hermanjl@cs.unc.edu> | 2013-01-22 10:38:37 -0500 |
commit | fcc9d2e5a6c89d22b8b773a64fb4ad21ac318446 (patch) | |
tree | a57612d1888735a2ec7972891b68c1ac5ec8faea /drivers/gpu/ion/tegra | |
parent | 8dea78da5cee153b8af9c07a2745f6c55057fe12 (diff) |
Diffstat (limited to 'drivers/gpu/ion/tegra')
-rw-r--r-- | drivers/gpu/ion/tegra/Makefile | 1 | ||||
-rw-r--r-- | drivers/gpu/ion/tegra/tegra_ion.c | 599 |
2 files changed, 600 insertions, 0 deletions
diff --git a/drivers/gpu/ion/tegra/Makefile b/drivers/gpu/ion/tegra/Makefile new file mode 100644 index 00000000000..11cd003fb08 --- /dev/null +++ b/drivers/gpu/ion/tegra/Makefile | |||
@@ -0,0 +1 @@ | |||
obj-y += tegra_ion.o | |||
diff --git a/drivers/gpu/ion/tegra/tegra_ion.c b/drivers/gpu/ion/tegra/tegra_ion.c new file mode 100644 index 00000000000..2252079279e --- /dev/null +++ b/drivers/gpu/ion/tegra/tegra_ion.c | |||
@@ -0,0 +1,599 @@ | |||
1 | /* | ||
2 | * drivers/gpu/tegra/tegra_ion.c | ||
3 | * | ||
4 | * Copyright (C) 2011 Google, Inc. | ||
5 | * Copyright (C) 2011, NVIDIA Corporation. | ||
6 | * | ||
7 | * This software is licensed under the terms of the GNU General Public | ||
8 | * License version 2, as published by the Free Software Foundation, and | ||
9 | * may be copied, distributed, and modified under those terms. | ||
10 | * | ||
11 | * This program is distributed in the hope that it will be useful, | ||
12 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
13 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
14 | * GNU General Public License for more details. | ||
15 | * | ||
16 | */ | ||
17 | |||
18 | #define pr_fmt(fmt) "%s():%d: " fmt, __func__, __LINE__ | ||
19 | |||
20 | #include <linux/err.h> | ||
21 | #include <linux/ion.h> | ||
22 | #include <linux/tegra_ion.h> | ||
23 | #include <linux/platform_device.h> | ||
24 | #include <linux/slab.h> | ||
25 | #include <linux/uaccess.h> | ||
26 | #include <linux/syscalls.h> | ||
27 | #include <linux/io.h> | ||
28 | #include "../ion_priv.h" | ||
29 | |||
30 | #define CLIENT_HEAP_MASK 0xFFFFFFFF | ||
31 | #define HEAP_FLAGS 0xFF | ||
32 | |||
33 | #if !defined(CONFIG_TEGRA_NVMAP) | ||
34 | #include "mach/nvmap.h" | ||
35 | struct nvmap_device *nvmap_dev; | ||
36 | #endif | ||
37 | |||
38 | static struct ion_device *idev; | ||
39 | static int num_heaps; | ||
40 | static struct ion_heap **heaps; | ||
41 | |||
42 | static int tegra_ion_pin(struct ion_client *client, | ||
43 | unsigned int cmd, | ||
44 | unsigned long arg) | ||
45 | { | ||
46 | struct tegra_ion_pin_data data; | ||
47 | int ret; | ||
48 | struct ion_handle *on_stack[16]; | ||
49 | struct ion_handle **refs = on_stack; | ||
50 | int i; | ||
51 | bool valid_handle; | ||
52 | |||
53 | if (copy_from_user(&data, (void __user *)arg, sizeof(data))) | ||
54 | return -EFAULT; | ||
55 | if (data.count) { | ||
56 | size_t bytes = data.count * sizeof(struct ion_handle *); | ||
57 | |||
58 | if (data.count > ARRAY_SIZE(on_stack)) | ||
59 | refs = kmalloc(data.count * sizeof(*refs), GFP_KERNEL); | ||
60 | else | ||
61 | refs = on_stack; | ||
62 | if (!refs) | ||
63 | return -ENOMEM; | ||
64 | if (copy_from_user(refs, (void *)data.handles, bytes)) { | ||
65 | ret = -EFAULT; | ||
66 | goto err; | ||
67 | } | ||
68 | } else | ||
69 | return -EINVAL; | ||
70 | |||
71 | mutex_lock(&client->lock); | ||
72 | for (i = 0; i < data.count; i++) { | ||
73 | /* Ignore NULL pointers during unpin operation. */ | ||
74 | if (!refs[i] && cmd == TEGRA_ION_UNPIN) | ||
75 | continue; | ||
76 | valid_handle = ion_handle_validate(client, refs[i]); | ||
77 | if (!valid_handle) { | ||
78 | WARN(1, "invalid handle passed h=0x%x", (u32)refs[i]); | ||
79 | mutex_unlock(&client->lock); | ||
80 | ret = -EINVAL; | ||
81 | goto err; | ||
82 | } | ||
83 | } | ||
84 | mutex_unlock(&client->lock); | ||
85 | |||
86 | if (cmd == TEGRA_ION_PIN) { | ||
87 | ion_phys_addr_t addr; | ||
88 | size_t len; | ||
89 | |||
90 | for (i = 0; i < data.count; i++) { | ||
91 | ret = ion_phys(client, refs[i], &addr, &len); | ||
92 | if (ret) | ||
93 | goto err; | ||
94 | ion_handle_get(refs[i]); | ||
95 | ret = put_user(addr, &data.addr[i]); | ||
96 | if (ret) | ||
97 | return ret; | ||
98 | } | ||
99 | } else if (cmd == TEGRA_ION_UNPIN) { | ||
100 | for (i = 0; i < data.count; i++) { | ||
101 | if (refs[i]) | ||
102 | ion_handle_put(refs[i]); | ||
103 | } | ||
104 | } | ||
105 | |||
106 | err: | ||
107 | if (ret) { | ||
108 | pr_err("error, ret=0x%x", ret); | ||
109 | /* FIXME: undo pinning. */ | ||
110 | } | ||
111 | if (refs != on_stack) | ||
112 | kfree(refs); | ||
113 | return ret; | ||
114 | } | ||
115 | |||
116 | static int tegra_ion_alloc_from_id(struct ion_client *client, | ||
117 | unsigned int cmd, | ||
118 | unsigned long arg) | ||
119 | { | ||
120 | struct tegra_ion_id_data data; | ||
121 | struct ion_buffer *buffer; | ||
122 | struct tegra_ion_id_data *user_data = (struct tegra_ion_id_data *)arg; | ||
123 | |||
124 | if (copy_from_user(&data, (void __user *)arg, sizeof(data))) | ||
125 | return -EFAULT; | ||
126 | buffer = (struct ion_buffer *)data.id; | ||
127 | data.handle = ion_import(client, buffer); | ||
128 | data.size = buffer->size; | ||
129 | if (put_user(data.handle, &user_data->handle)) | ||
130 | return -EFAULT; | ||
131 | if (put_user(data.size, &user_data->size)) | ||
132 | return -EFAULT; | ||
133 | return 0; | ||
134 | } | ||
135 | |||
136 | static int tegra_ion_get_id(struct ion_client *client, | ||
137 | unsigned int cmd, | ||
138 | unsigned long arg) | ||
139 | { | ||
140 | bool valid_handle; | ||
141 | struct tegra_ion_id_data data; | ||
142 | struct tegra_ion_id_data *user_data = (struct tegra_ion_id_data *)arg; | ||
143 | |||
144 | if (copy_from_user(&data, (void __user *)arg, sizeof(data))) | ||
145 | return -EFAULT; | ||
146 | |||
147 | mutex_lock(&client->lock); | ||
148 | valid_handle = ion_handle_validate(client, data.handle); | ||
149 | mutex_unlock(&client->lock); | ||
150 | |||
151 | if (!valid_handle) { | ||
152 | WARN(1, "invalid handle passed\n"); | ||
153 | return -EINVAL; | ||
154 | } | ||
155 | |||
156 | pr_debug("h=0x%x, b=0x%x, bref=%d", | ||
157 | (u32)data.handle, (u32)data.handle->buffer, | ||
158 | atomic_read(&data.handle->buffer->ref.refcount)); | ||
159 | if (put_user((unsigned long)ion_handle_buffer(data.handle), | ||
160 | &user_data->id)) | ||
161 | return -EFAULT; | ||
162 | return 0; | ||
163 | } | ||
164 | |||
165 | static int tegra_ion_cache_maint(struct ion_client *client, | ||
166 | unsigned int cmd, | ||
167 | unsigned long arg) | ||
168 | { | ||
169 | wmb(); | ||
170 | return 0; | ||
171 | } | ||
172 | |||
173 | static int tegra_ion_rw(struct ion_client *client, | ||
174 | unsigned int cmd, | ||
175 | unsigned long arg) | ||
176 | { | ||
177 | bool valid_handle; | ||
178 | struct tegra_ion_rw_data data; | ||
179 | char *kern_addr, *src; | ||
180 | int ret = 0; | ||
181 | size_t copied = 0; | ||
182 | |||
183 | if (copy_from_user(&data, (void __user *)arg, sizeof(data))) | ||
184 | return -EFAULT; | ||
185 | |||
186 | if (!data.handle || !data.addr || !data.count || !data.elem_size) | ||
187 | return -EINVAL; | ||
188 | |||
189 | mutex_lock(&client->lock); | ||
190 | valid_handle = ion_handle_validate(client, data.handle); | ||
191 | mutex_unlock(&client->lock); | ||
192 | |||
193 | if (!valid_handle) { | ||
194 | WARN(1, "%s: invalid handle passed to get id.\n", __func__); | ||
195 | return -EINVAL; | ||
196 | } | ||
197 | |||
198 | if (data.elem_size == data.mem_stride && | ||
199 | data.elem_size == data.user_stride) { | ||
200 | data.elem_size *= data.count; | ||
201 | data.mem_stride = data.elem_size; | ||
202 | data.user_stride = data.elem_size; | ||
203 | data.count = 1; | ||
204 | } | ||
205 | |||
206 | kern_addr = ion_map_kernel(client, data.handle); | ||
207 | |||
208 | while (data.count--) { | ||
209 | if (data.offset + data.elem_size > data.handle->buffer->size) { | ||
210 | WARN(1, "read/write outside of handle\n"); | ||
211 | ret = -EFAULT; | ||
212 | break; | ||
213 | } | ||
214 | |||
215 | src = kern_addr + data.offset; | ||
216 | if (cmd == TEGRA_ION_READ) | ||
217 | ret = copy_to_user((void *)data.addr, | ||
218 | src, data.elem_size); | ||
219 | else | ||
220 | ret = copy_from_user(src, | ||
221 | (void *)data.addr, data.elem_size); | ||
222 | |||
223 | if (ret) | ||
224 | break; | ||
225 | |||
226 | copied += data.elem_size; | ||
227 | data.addr += data.user_stride; | ||
228 | data.offset += data.mem_stride; | ||
229 | } | ||
230 | |||
231 | ion_unmap_kernel(client, data.handle); | ||
232 | return ret; | ||
233 | } | ||
234 | |||
235 | static int tegra_ion_get_param(struct ion_client *client, | ||
236 | unsigned int cmd, | ||
237 | unsigned long arg) | ||
238 | { | ||
239 | bool valid_handle; | ||
240 | struct tegra_ion_get_params_data data; | ||
241 | struct tegra_ion_get_params_data *user_data = | ||
242 | (struct tegra_ion_get_params_data *)arg; | ||
243 | struct ion_buffer *buffer; | ||
244 | |||
245 | if (copy_from_user(&data, (void __user *)arg, sizeof(data))) | ||
246 | return -EFAULT; | ||
247 | |||
248 | mutex_lock(&client->lock); | ||
249 | valid_handle = ion_handle_validate(client, data.handle); | ||
250 | mutex_unlock(&client->lock); | ||
251 | |||
252 | if (!valid_handle) { | ||
253 | WARN(1, "%s: invalid handle passed to get id.\n", __func__); | ||
254 | return -EINVAL; | ||
255 | } | ||
256 | |||
257 | buffer = ion_handle_buffer(data.handle); | ||
258 | data.align = 4096; | ||
259 | data.heap = 1; | ||
260 | ion_phys(client, data.handle, &data.addr, &data.size); | ||
261 | |||
262 | if (copy_to_user(user_data, &data, sizeof(data))) | ||
263 | return -EFAULT; | ||
264 | |||
265 | return 0; | ||
266 | } | ||
267 | |||
268 | static long tegra_ion_ioctl(struct ion_client *client, | ||
269 | unsigned int cmd, | ||
270 | unsigned long arg) | ||
271 | { | ||
272 | int ret = -ENOTTY; | ||
273 | |||
274 | switch (cmd) { | ||
275 | case TEGRA_ION_ALLOC_FROM_ID: | ||
276 | ret = tegra_ion_alloc_from_id(client, cmd, arg); | ||
277 | break; | ||
278 | case TEGRA_ION_GET_ID: | ||
279 | ret = tegra_ion_get_id(client, cmd, arg); | ||
280 | break; | ||
281 | case TEGRA_ION_PIN: | ||
282 | case TEGRA_ION_UNPIN: | ||
283 | ret = tegra_ion_pin(client, cmd, arg); | ||
284 | break; | ||
285 | case TEGRA_ION_CACHE_MAINT: | ||
286 | ret = tegra_ion_cache_maint(client, cmd, arg); | ||
287 | break; | ||
288 | case TEGRA_ION_READ: | ||
289 | case TEGRA_ION_WRITE: | ||
290 | ret = tegra_ion_rw(client, cmd, arg); | ||
291 | break; | ||
292 | case TEGRA_ION_GET_PARAM: | ||
293 | ret = tegra_ion_get_param(client, cmd, arg); | ||
294 | break; | ||
295 | default: | ||
296 | WARN(1, "Unknown custom ioctl\n"); | ||
297 | return -ENOTTY; | ||
298 | } | ||
299 | return ret; | ||
300 | } | ||
301 | |||
302 | int tegra_ion_probe(struct platform_device *pdev) | ||
303 | { | ||
304 | struct ion_platform_data *pdata = pdev->dev.platform_data; | ||
305 | int i; | ||
306 | |||
307 | num_heaps = pdata->nr; | ||
308 | |||
309 | heaps = kzalloc(sizeof(struct ion_heap *) * pdata->nr, GFP_KERNEL); | ||
310 | |||
311 | idev = ion_device_create(tegra_ion_ioctl); | ||
312 | if (IS_ERR_OR_NULL(idev)) { | ||
313 | kfree(heaps); | ||
314 | return PTR_ERR(idev); | ||
315 | } | ||
316 | |||
317 | /* create the heaps as specified in the board file */ | ||
318 | for (i = 0; i < num_heaps; i++) { | ||
319 | struct ion_platform_heap *heap_data = &pdata->heaps[i]; | ||
320 | |||
321 | heaps[i] = ion_heap_create(heap_data); | ||
322 | if (IS_ERR_OR_NULL(heaps[i])) { | ||
323 | pr_warn("%s(type:%d id:%d) isn't supported\n", | ||
324 | heap_data->name, | ||
325 | heap_data->type, heap_data->id); | ||
326 | continue; | ||
327 | } | ||
328 | ion_device_add_heap(idev, heaps[i]); | ||
329 | } | ||
330 | platform_set_drvdata(pdev, idev); | ||
331 | #if !defined(CONFIG_TEGRA_NVMAP) | ||
332 | nvmap_dev = (struct nvmap_device *)idev; | ||
333 | #endif | ||
334 | return 0; | ||
335 | } | ||
336 | |||
337 | int tegra_ion_remove(struct platform_device *pdev) | ||
338 | { | ||
339 | struct ion_device *idev = platform_get_drvdata(pdev); | ||
340 | int i; | ||
341 | |||
342 | ion_device_destroy(idev); | ||
343 | for (i = 0; i < num_heaps; i++) | ||
344 | ion_heap_destroy(heaps[i]); | ||
345 | kfree(heaps); | ||
346 | return 0; | ||
347 | } | ||
348 | |||
349 | static struct platform_driver ion_driver = { | ||
350 | .probe = tegra_ion_probe, | ||
351 | .remove = tegra_ion_remove, | ||
352 | .driver = { .name = "ion-tegra" } | ||
353 | }; | ||
354 | |||
355 | static int __init ion_init(void) | ||
356 | { | ||
357 | return platform_driver_register(&ion_driver); | ||
358 | } | ||
359 | |||
360 | static void __exit ion_exit(void) | ||
361 | { | ||
362 | platform_driver_unregister(&ion_driver); | ||
363 | } | ||
364 | |||
365 | fs_initcall(ion_init); | ||
366 | module_exit(ion_exit); | ||
367 | |||
368 | #if !defined(CONFIG_TEGRA_NVMAP) | ||
369 | struct nvmap_client *nvmap_create_client(struct nvmap_device *dev, | ||
370 | const char *name) | ||
371 | { | ||
372 | return ion_client_create(dev, CLIENT_HEAP_MASK, name); | ||
373 | } | ||
374 | |||
375 | struct nvmap_handle_ref *nvmap_alloc(struct nvmap_client *client, size_t size, | ||
376 | size_t align, unsigned int flags, | ||
377 | unsigned int heap_mask) | ||
378 | { | ||
379 | return ion_alloc(client, size, align, HEAP_FLAGS); | ||
380 | } | ||
381 | |||
382 | void nvmap_free(struct nvmap_client *client, struct nvmap_handle_ref *r) | ||
383 | { | ||
384 | ion_free(client, r); | ||
385 | } | ||
386 | |||
387 | void *nvmap_mmap(struct nvmap_handle_ref *r) | ||
388 | { | ||
389 | return ion_map_kernel(r->client, r); | ||
390 | } | ||
391 | |||
392 | void nvmap_munmap(struct nvmap_handle_ref *r, void *addr) | ||
393 | { | ||
394 | ion_unmap_kernel(r->client, r); | ||
395 | } | ||
396 | |||
397 | struct nvmap_client *nvmap_client_get_file(int fd) | ||
398 | { | ||
399 | return ion_client_get_file(fd); | ||
400 | } | ||
401 | |||
402 | struct nvmap_client *nvmap_client_get(struct nvmap_client *client) | ||
403 | { | ||
404 | ion_client_get(client); | ||
405 | return client; | ||
406 | } | ||
407 | |||
408 | void nvmap_client_put(struct nvmap_client *c) | ||
409 | { | ||
410 | ion_client_put(c); | ||
411 | } | ||
412 | |||
413 | phys_addr_t nvmap_pin(struct nvmap_client *c, struct nvmap_handle_ref *r) | ||
414 | { | ||
415 | ion_phys_addr_t addr; | ||
416 | size_t len; | ||
417 | |||
418 | ion_handle_get(r); | ||
419 | ion_phys(c, r, &addr, &len); | ||
420 | wmb(); | ||
421 | return addr; | ||
422 | } | ||
423 | |||
424 | phys_addr_t nvmap_handle_address(struct nvmap_client *c, unsigned long id) | ||
425 | { | ||
426 | struct ion_handle *handle; | ||
427 | ion_phys_addr_t addr; | ||
428 | size_t len; | ||
429 | |||
430 | handle = nvmap_convert_handle_u2k(id); | ||
431 | ion_phys(c, handle, &addr, &len); | ||
432 | return addr; | ||
433 | } | ||
434 | |||
435 | void nvmap_unpin(struct nvmap_client *client, struct nvmap_handle_ref *r) | ||
436 | { | ||
437 | if (r) | ||
438 | ion_handle_put(r); | ||
439 | } | ||
440 | |||
441 | static int nvmap_reloc_pin_array(struct ion_client *client, | ||
442 | const struct nvmap_pinarray_elem *arr, | ||
443 | int nr, struct ion_handle *gather) | ||
444 | { | ||
445 | struct ion_handle *last_patch = NULL; | ||
446 | void *patch_addr; | ||
447 | ion_phys_addr_t pin_addr; | ||
448 | size_t len; | ||
449 | int i; | ||
450 | |||
451 | for (i = 0; i < nr; i++) { | ||
452 | struct ion_handle *patch; | ||
453 | struct ion_handle *pin; | ||
454 | ion_phys_addr_t reloc_addr; | ||
455 | |||
456 | /* all of the handles are validated and get'ted prior to | ||
457 | * calling this function, so casting is safe here */ | ||
458 | pin = (struct ion_handle *)arr[i].pin_mem; | ||
459 | |||
460 | if (arr[i].patch_mem == (unsigned long)last_patch) { | ||
461 | patch = last_patch; | ||
462 | } else if (arr[i].patch_mem == (unsigned long)gather) { | ||
463 | patch = gather; | ||
464 | } else { | ||
465 | if (last_patch) | ||
466 | ion_handle_put(last_patch); | ||
467 | |||
468 | ion_handle_get((struct ion_handle *)arr[i].patch_mem); | ||
469 | patch = (struct ion_handle *)arr[i].patch_mem; | ||
470 | if (!patch) | ||
471 | return -EPERM; | ||
472 | last_patch = patch; | ||
473 | } | ||
474 | |||
475 | patch_addr = ion_map_kernel(client, patch); | ||
476 | patch_addr = patch_addr + arr[i].patch_offset; | ||
477 | |||
478 | ion_phys(client, pin, &pin_addr, &len); | ||
479 | reloc_addr = pin_addr + arr[i].pin_offset; | ||
480 | __raw_writel(reloc_addr, patch_addr); | ||
481 | ion_unmap_kernel(client, patch); | ||
482 | } | ||
483 | |||
484 | if (last_patch) | ||
485 | ion_handle_put(last_patch); | ||
486 | |||
487 | wmb(); | ||
488 | return 0; | ||
489 | } | ||
490 | |||
491 | int nvmap_pin_array(struct nvmap_client *client, struct nvmap_handle *gather, | ||
492 | const struct nvmap_pinarray_elem *arr, int nr, | ||
493 | struct nvmap_handle **unique) | ||
494 | { | ||
495 | int i; | ||
496 | int count = 0; | ||
497 | |||
498 | /* FIXME: take care of duplicate ones & validation. */ | ||
499 | for (i = 0; i < nr; i++) { | ||
500 | unique[i] = (struct nvmap_handle *)arr[i].pin_mem; | ||
501 | nvmap_pin(client, (struct nvmap_handle_ref *)unique[i]); | ||
502 | count++; | ||
503 | } | ||
504 | nvmap_reloc_pin_array((struct ion_client *)client, | ||
505 | arr, nr, (struct ion_handle *)gather); | ||
506 | return nr; | ||
507 | } | ||
508 | |||
509 | void nvmap_unpin_handles(struct nvmap_client *client, | ||
510 | struct nvmap_handle **h, int nr) | ||
511 | { | ||
512 | int i; | ||
513 | |||
514 | for (i = 0; i < nr; i++) | ||
515 | nvmap_unpin(client, h[i]); | ||
516 | } | ||
517 | |||
518 | int nvmap_patch_word(struct nvmap_client *client, | ||
519 | struct nvmap_handle *patch, | ||
520 | u32 patch_offset, u32 patch_value) | ||
521 | { | ||
522 | void *vaddr; | ||
523 | u32 *patch_addr; | ||
524 | |||
525 | vaddr = ion_map_kernel(client, patch); | ||
526 | patch_addr = vaddr + patch_offset; | ||
527 | __raw_writel(patch_value, patch_addr); | ||
528 | wmb(); | ||
529 | ion_unmap_kernel(client, patch); | ||
530 | return 0; | ||
531 | } | ||
532 | |||
533 | struct nvmap_handle *nvmap_handle_get(struct nvmap_handle *h); | ||
534 | struct nvmap_handle *nvmap_get_handle_id(struct nvmap_client *client, | ||
535 | unsigned long id) | ||
536 | { | ||
537 | struct ion_handle *handle; | ||
538 | |||
539 | handle = (struct ion_handle *)nvmap_convert_handle_u2k(id); | ||
540 | pr_debug("id=0x%x, h=0x%x,c=0x%x", | ||
541 | (u32)id, (u32)handle, (u32)client); | ||
542 | nvmap_handle_get(handle); | ||
543 | return handle; | ||
544 | } | ||
545 | |||
546 | struct nvmap_handle_ref *nvmap_duplicate_handle_id(struct nvmap_client *client, | ||
547 | unsigned long id) | ||
548 | { | ||
549 | struct ion_buffer *buffer; | ||
550 | struct ion_handle *handle; | ||
551 | struct ion_client *ion_client = client; | ||
552 | |||
553 | handle = (struct ion_handle *)nvmap_convert_handle_u2k(id); | ||
554 | pr_debug("id=0x%x, h=0x%x,c=0x%x", | ||
555 | (u32)id, (u32)handle, (u32)client); | ||
556 | buffer = handle->buffer; | ||
557 | |||
558 | handle = ion_handle_create(client, buffer); | ||
559 | |||
560 | mutex_lock(&ion_client->lock); | ||
561 | ion_handle_add(ion_client, handle); | ||
562 | mutex_unlock(&ion_client->lock); | ||
563 | |||
564 | pr_debug("dup id=0x%x, h=0x%x", (u32)id, (u32)handle); | ||
565 | return handle; | ||
566 | } | ||
567 | |||
568 | void _nvmap_handle_free(struct nvmap_handle *h) | ||
569 | { | ||
570 | ion_handle_put(h); | ||
571 | } | ||
572 | |||
573 | struct nvmap_handle_ref *nvmap_alloc_iovm(struct nvmap_client *client, | ||
574 | size_t size, size_t align, unsigned int flags, unsigned int iova_start) | ||
575 | { | ||
576 | struct ion_handle *h; | ||
577 | |||
578 | h = ion_alloc(client, size, align, 0xFF); | ||
579 | ion_remap_dma(client, h, iova_start); | ||
580 | return h; | ||
581 | } | ||
582 | |||
583 | void nvmap_free_iovm(struct nvmap_client *client, struct nvmap_handle_ref *r) | ||
584 | { | ||
585 | ion_free(client, r); | ||
586 | } | ||
587 | |||
588 | struct nvmap_handle *nvmap_handle_get(struct nvmap_handle *h) | ||
589 | { | ||
590 | ion_handle_get(h); | ||
591 | return h; | ||
592 | } | ||
593 | |||
594 | void nvmap_handle_put(struct nvmap_handle *h) | ||
595 | { | ||
596 | ion_handle_put(h); | ||
597 | } | ||
598 | |||
599 | #endif | ||