2 * Copyright © 2014 Advanced Micro Devices, Inc.
5 * Permission is hereby granted, free of charge, to any person obtaining a
6 * copy of this software and associated documentation files (the "Software"),
7 * to deal in the Software without restriction, including without limitation
8 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
9 * and/or sell copies of the Software, and to permit persons to whom the
10 * Software is furnished to do so, subject to the following conditions:
12 * The above copyright notice and this permission notice shall be included in
13 * all copies or substantial portions of the Software.
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
19 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
20 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
21 * OTHER DEALINGS IN THE SOFTWARE.
36 #include <sys/ioctl.h>
40 #include "libdrm_macros.h"
42 #include "amdgpu_drm.h"
43 #include "amdgpu_internal.h"
44 #include "util_hash_table.h"
45 #include "util_math.h"
47 static void amdgpu_close_kms_handle(amdgpu_device_handle dev,
50 struct drm_gem_close args = {};
53 drmIoctl(dev->fd, DRM_IOCTL_GEM_CLOSE, &args);
56 drm_private void amdgpu_bo_free_internal(amdgpu_bo_handle bo)
58 /* Remove the buffer from the hash tables. */
59 pthread_mutex_lock(&bo->dev->bo_table_mutex);
60 util_hash_table_remove(bo->dev->bo_handles,
61 (void*)(uintptr_t)bo->handle);
63 util_hash_table_remove(bo->dev->bo_flink_names,
64 (void*)(uintptr_t)bo->flink_name);
66 pthread_mutex_unlock(&bo->dev->bo_table_mutex);
68 /* Release CPU access. */
69 if (bo->cpu_map_count > 0) {
70 bo->cpu_map_count = 1;
71 amdgpu_bo_cpu_unmap(bo);
74 amdgpu_close_kms_handle(bo->dev, bo->handle);
75 pthread_mutex_destroy(&bo->cpu_access_mutex);
79 int amdgpu_bo_alloc(amdgpu_device_handle dev,
80 struct amdgpu_bo_alloc_request *alloc_buffer,
81 amdgpu_bo_handle *buf_handle)
84 union drm_amdgpu_gem_create args;
85 unsigned heap = alloc_buffer->preferred_heap;
88 /* It's an error if the heap is not specified */
89 if (!(heap & (AMDGPU_GEM_DOMAIN_GTT | AMDGPU_GEM_DOMAIN_VRAM)))
92 bo = calloc(1, sizeof(struct amdgpu_bo));
96 atomic_set(&bo->refcount, 1);
98 bo->alloc_size = alloc_buffer->alloc_size;
100 memset(&args, 0, sizeof(args));
101 args.in.bo_size = alloc_buffer->alloc_size;
102 args.in.alignment = alloc_buffer->phys_alignment;
104 /* Set the placement. */
105 args.in.domains = heap;
106 args.in.domain_flags = alloc_buffer->flags;
108 /* Allocate the buffer with the preferred heap. */
109 r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_GEM_CREATE,
110 &args, sizeof(args));
116 bo->handle = args.out.handle;
118 pthread_mutex_init(&bo->cpu_access_mutex, NULL);
124 int amdgpu_bo_set_metadata(amdgpu_bo_handle bo,
125 struct amdgpu_bo_metadata *info)
127 struct drm_amdgpu_gem_metadata args = {};
129 args.handle = bo->handle;
130 args.op = AMDGPU_GEM_METADATA_OP_SET_METADATA;
131 args.data.flags = info->flags;
132 args.data.tiling_info = info->tiling_info;
134 if (info->size_metadata > sizeof(args.data.data))
137 if (info->size_metadata) {
138 args.data.data_size_bytes = info->size_metadata;
139 memcpy(args.data.data, info->umd_metadata, info->size_metadata);
142 return drmCommandWriteRead(bo->dev->fd,
143 DRM_AMDGPU_GEM_METADATA,
144 &args, sizeof(args));
147 int amdgpu_bo_query_info(amdgpu_bo_handle bo,
148 struct amdgpu_bo_info *info)
150 struct drm_amdgpu_gem_metadata metadata = {};
151 struct drm_amdgpu_gem_create_in bo_info = {};
152 struct drm_amdgpu_gem_op gem_op = {};
155 /* Validate the BO passed in */
159 /* Query metadata. */
160 metadata.handle = bo->handle;
161 metadata.op = AMDGPU_GEM_METADATA_OP_GET_METADATA;
163 r = drmCommandWriteRead(bo->dev->fd, DRM_AMDGPU_GEM_METADATA,
164 &metadata, sizeof(metadata));
168 if (metadata.data.data_size_bytes >
169 sizeof(info->metadata.umd_metadata))
172 /* Query buffer info. */
173 gem_op.handle = bo->handle;
174 gem_op.op = AMDGPU_GEM_OP_GET_GEM_CREATE_INFO;
175 gem_op.value = (uintptr_t)&bo_info;
177 r = drmCommandWriteRead(bo->dev->fd, DRM_AMDGPU_GEM_OP,
178 &gem_op, sizeof(gem_op));
182 memset(info, 0, sizeof(*info));
183 info->alloc_size = bo_info.bo_size;
184 info->phys_alignment = bo_info.alignment;
185 info->preferred_heap = bo_info.domains;
186 info->alloc_flags = bo_info.domain_flags;
187 info->metadata.flags = metadata.data.flags;
188 info->metadata.tiling_info = metadata.data.tiling_info;
190 info->metadata.size_metadata = metadata.data.data_size_bytes;
191 if (metadata.data.data_size_bytes > 0)
192 memcpy(info->metadata.umd_metadata, metadata.data.data,
193 metadata.data.data_size_bytes);
198 static void amdgpu_add_handle_to_table(amdgpu_bo_handle bo)
200 pthread_mutex_lock(&bo->dev->bo_table_mutex);
201 util_hash_table_set(bo->dev->bo_handles,
202 (void*)(uintptr_t)bo->handle, bo);
203 pthread_mutex_unlock(&bo->dev->bo_table_mutex);
206 static int amdgpu_bo_export_flink(amdgpu_bo_handle bo)
208 struct drm_gem_flink flink;
219 if (bo->dev->flink_fd != bo->dev->fd) {
220 r = drmPrimeHandleToFD(bo->dev->fd, bo->handle, DRM_CLOEXEC,
223 r = drmPrimeFDToHandle(bo->dev->flink_fd, dma_fd, &handle);
228 fd = bo->dev->flink_fd;
230 memset(&flink, 0, sizeof(flink));
231 flink.handle = handle;
233 r = drmIoctl(fd, DRM_IOCTL_GEM_FLINK, &flink);
237 bo->flink_name = flink.name;
239 if (bo->dev->flink_fd != bo->dev->fd) {
240 struct drm_gem_close args = {};
241 args.handle = handle;
242 drmIoctl(bo->dev->flink_fd, DRM_IOCTL_GEM_CLOSE, &args);
245 pthread_mutex_lock(&bo->dev->bo_table_mutex);
246 util_hash_table_set(bo->dev->bo_flink_names,
247 (void*)(uintptr_t)bo->flink_name,
249 pthread_mutex_unlock(&bo->dev->bo_table_mutex);
254 int amdgpu_bo_export(amdgpu_bo_handle bo,
255 enum amdgpu_bo_handle_type type,
256 uint32_t *shared_handle)
261 case amdgpu_bo_handle_type_gem_flink_name:
262 r = amdgpu_bo_export_flink(bo);
266 *shared_handle = bo->flink_name;
269 case amdgpu_bo_handle_type_kms:
270 amdgpu_add_handle_to_table(bo);
271 *shared_handle = bo->handle;
274 case amdgpu_bo_handle_type_dma_buf_fd:
275 amdgpu_add_handle_to_table(bo);
276 return drmPrimeHandleToFD(bo->dev->fd, bo->handle, DRM_CLOEXEC,
277 (int*)shared_handle);
282 int amdgpu_bo_import(amdgpu_device_handle dev,
283 enum amdgpu_bo_handle_type type,
284 uint32_t shared_handle,
285 struct amdgpu_bo_import_result *output)
287 struct drm_gem_open open_arg = {};
288 struct amdgpu_bo *bo = NULL;
291 uint64_t dma_buf_size = 0;
293 /* We must maintain a list of pairs <handle, bo>, so that we always
294 * return the same amdgpu_bo instance for the same handle. */
295 pthread_mutex_lock(&dev->bo_table_mutex);
297 /* Convert a DMA buf handle to a KMS handle now. */
298 if (type == amdgpu_bo_handle_type_dma_buf_fd) {
302 /* Get a KMS handle. */
303 r = drmPrimeFDToHandle(dev->fd, shared_handle, &handle);
305 pthread_mutex_unlock(&dev->bo_table_mutex);
309 /* Query the buffer size. */
310 size = lseek(shared_handle, 0, SEEK_END);
311 if (size == (off_t)-1) {
312 pthread_mutex_unlock(&dev->bo_table_mutex);
313 amdgpu_close_kms_handle(dev, handle);
316 lseek(shared_handle, 0, SEEK_SET);
319 shared_handle = handle;
322 /* If we have already created a buffer with this handle, find it. */
324 case amdgpu_bo_handle_type_gem_flink_name:
325 bo = util_hash_table_get(dev->bo_flink_names,
326 (void*)(uintptr_t)shared_handle);
329 case amdgpu_bo_handle_type_dma_buf_fd:
330 bo = util_hash_table_get(dev->bo_handles,
331 (void*)(uintptr_t)shared_handle);
334 case amdgpu_bo_handle_type_kms:
335 /* Importing a KMS handle in not allowed. */
336 pthread_mutex_unlock(&dev->bo_table_mutex);
340 pthread_mutex_unlock(&dev->bo_table_mutex);
345 pthread_mutex_unlock(&dev->bo_table_mutex);
347 /* The buffer already exists, just bump the refcount. */
348 atomic_inc(&bo->refcount);
350 output->buf_handle = bo;
351 output->alloc_size = bo->alloc_size;
355 bo = calloc(1, sizeof(struct amdgpu_bo));
357 pthread_mutex_unlock(&dev->bo_table_mutex);
358 if (type == amdgpu_bo_handle_type_dma_buf_fd) {
359 amdgpu_close_kms_handle(dev, shared_handle);
364 /* Open the handle. */
366 case amdgpu_bo_handle_type_gem_flink_name:
367 open_arg.name = shared_handle;
368 r = drmIoctl(dev->flink_fd, DRM_IOCTL_GEM_OPEN, &open_arg);
371 pthread_mutex_unlock(&dev->bo_table_mutex);
375 bo->handle = open_arg.handle;
376 if (dev->flink_fd != dev->fd) {
377 r = drmPrimeHandleToFD(dev->flink_fd, bo->handle, DRM_CLOEXEC, &dma_fd);
380 pthread_mutex_unlock(&dev->bo_table_mutex);
383 r = drmPrimeFDToHandle(dev->fd, dma_fd, &bo->handle );
389 pthread_mutex_unlock(&dev->bo_table_mutex);
393 bo->flink_name = shared_handle;
394 bo->alloc_size = open_arg.size;
395 util_hash_table_set(dev->bo_flink_names,
396 (void*)(uintptr_t)bo->flink_name, bo);
399 case amdgpu_bo_handle_type_dma_buf_fd:
400 bo->handle = shared_handle;
401 bo->alloc_size = dma_buf_size;
404 case amdgpu_bo_handle_type_kms:
405 assert(0); /* unreachable */
409 atomic_set(&bo->refcount, 1);
411 pthread_mutex_init(&bo->cpu_access_mutex, NULL);
413 util_hash_table_set(dev->bo_handles, (void*)(uintptr_t)bo->handle, bo);
414 pthread_mutex_unlock(&dev->bo_table_mutex);
416 output->buf_handle = bo;
417 output->alloc_size = bo->alloc_size;
421 int amdgpu_bo_free(amdgpu_bo_handle buf_handle)
423 /* Just drop the reference. */
424 amdgpu_bo_reference(&buf_handle, NULL);
428 int amdgpu_bo_cpu_map(amdgpu_bo_handle bo, void **cpu)
430 union drm_amdgpu_gem_mmap args;
434 pthread_mutex_lock(&bo->cpu_access_mutex);
438 assert(bo->cpu_map_count > 0);
441 pthread_mutex_unlock(&bo->cpu_access_mutex);
445 assert(bo->cpu_map_count == 0);
447 memset(&args, 0, sizeof(args));
449 /* Query the buffer address (args.addr_ptr).
450 * The kernel driver ignores the offset and size parameters. */
451 args.in.handle = bo->handle;
453 r = drmCommandWriteRead(bo->dev->fd, DRM_AMDGPU_GEM_MMAP, &args,
456 pthread_mutex_unlock(&bo->cpu_access_mutex);
460 /* Map the buffer. */
461 ptr = drm_mmap(NULL, bo->alloc_size, PROT_READ | PROT_WRITE, MAP_SHARED,
462 bo->dev->fd, args.out.addr_ptr);
463 if (ptr == MAP_FAILED) {
464 pthread_mutex_unlock(&bo->cpu_access_mutex);
469 bo->cpu_map_count = 1;
470 pthread_mutex_unlock(&bo->cpu_access_mutex);
476 int amdgpu_bo_cpu_unmap(amdgpu_bo_handle bo)
480 pthread_mutex_lock(&bo->cpu_access_mutex);
481 assert(bo->cpu_map_count >= 0);
483 if (bo->cpu_map_count == 0) {
485 pthread_mutex_unlock(&bo->cpu_access_mutex);
490 if (bo->cpu_map_count > 0) {
491 /* mapped multiple times */
492 pthread_mutex_unlock(&bo->cpu_access_mutex);
496 r = drm_munmap(bo->cpu_ptr, bo->alloc_size) == 0 ? 0 : -errno;
498 pthread_mutex_unlock(&bo->cpu_access_mutex);
502 int amdgpu_query_buffer_size_alignment(amdgpu_device_handle dev,
503 struct amdgpu_buffer_size_alignments *info)
505 info->size_local = dev->dev_info.pte_fragment_size;
506 info->size_remote = dev->dev_info.gart_page_size;
510 int amdgpu_bo_wait_for_idle(amdgpu_bo_handle bo,
514 union drm_amdgpu_gem_wait_idle args;
517 memset(&args, 0, sizeof(args));
518 args.in.handle = bo->handle;
519 args.in.timeout = amdgpu_cs_calculate_timeout(timeout_ns);
521 r = drmCommandWriteRead(bo->dev->fd, DRM_AMDGPU_GEM_WAIT_IDLE,
522 &args, sizeof(args));
525 *busy = args.out.status;
528 fprintf(stderr, "amdgpu: GEM_WAIT_IDLE failed with %i\n", r);
533 int amdgpu_create_bo_from_user_mem(amdgpu_device_handle dev,
536 amdgpu_bo_handle *buf_handle)
539 struct amdgpu_bo *bo;
540 struct drm_amdgpu_gem_userptr args;
542 args.addr = (uintptr_t)cpu;
543 args.flags = AMDGPU_GEM_USERPTR_ANONONLY | AMDGPU_GEM_USERPTR_REGISTER |
544 AMDGPU_GEM_USERPTR_VALIDATE;
546 r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_GEM_USERPTR,
547 &args, sizeof(args));
551 bo = calloc(1, sizeof(struct amdgpu_bo));
555 atomic_set(&bo->refcount, 1);
557 bo->alloc_size = size;
558 bo->handle = args.handle;
565 int amdgpu_bo_list_create(amdgpu_device_handle dev,
566 uint32_t number_of_resources,
567 amdgpu_bo_handle *resources,
568 uint8_t *resource_prios,
569 amdgpu_bo_list_handle *result)
571 struct drm_amdgpu_bo_list_entry *list;
572 union drm_amdgpu_bo_list args;
576 if (!number_of_resources)
579 /* overflow check for multiplication */
580 if (number_of_resources > UINT32_MAX / sizeof(struct drm_amdgpu_bo_list_entry))
583 list = malloc(number_of_resources * sizeof(struct drm_amdgpu_bo_list_entry));
587 *result = malloc(sizeof(struct amdgpu_bo_list));
593 memset(&args, 0, sizeof(args));
594 args.in.operation = AMDGPU_BO_LIST_OP_CREATE;
595 args.in.bo_number = number_of_resources;
596 args.in.bo_info_size = sizeof(struct drm_amdgpu_bo_list_entry);
597 args.in.bo_info_ptr = (uint64_t)(uintptr_t)list;
599 for (i = 0; i < number_of_resources; i++) {
600 list[i].bo_handle = resources[i]->handle;
602 list[i].bo_priority = resource_prios[i];
604 list[i].bo_priority = 0;
607 r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_BO_LIST,
608 &args, sizeof(args));
615 (*result)->dev = dev;
616 (*result)->handle = args.out.list_handle;
620 int amdgpu_bo_list_destroy(amdgpu_bo_list_handle list)
622 union drm_amdgpu_bo_list args;
625 memset(&args, 0, sizeof(args));
626 args.in.operation = AMDGPU_BO_LIST_OP_DESTROY;
627 args.in.list_handle = list->handle;
629 r = drmCommandWriteRead(list->dev->fd, DRM_AMDGPU_BO_LIST,
630 &args, sizeof(args));
638 int amdgpu_bo_list_update(amdgpu_bo_list_handle handle,
639 uint32_t number_of_resources,
640 amdgpu_bo_handle *resources,
641 uint8_t *resource_prios)
643 struct drm_amdgpu_bo_list_entry *list;
644 union drm_amdgpu_bo_list args;
648 if (!number_of_resources)
651 /* overflow check for multiplication */
652 if (number_of_resources > UINT32_MAX / sizeof(struct drm_amdgpu_bo_list_entry))
655 list = malloc(number_of_resources * sizeof(struct drm_amdgpu_bo_list_entry));
659 args.in.operation = AMDGPU_BO_LIST_OP_UPDATE;
660 args.in.list_handle = handle->handle;
661 args.in.bo_number = number_of_resources;
662 args.in.bo_info_size = sizeof(struct drm_amdgpu_bo_list_entry);
663 args.in.bo_info_ptr = (uintptr_t)list;
665 for (i = 0; i < number_of_resources; i++) {
666 list[i].bo_handle = resources[i]->handle;
668 list[i].bo_priority = resource_prios[i];
670 list[i].bo_priority = 0;
673 r = drmCommandWriteRead(handle->dev->fd, DRM_AMDGPU_BO_LIST,
674 &args, sizeof(args));
679 int amdgpu_bo_va_op(amdgpu_bo_handle bo,
686 amdgpu_device_handle dev = bo->dev;
688 size = ALIGN(size, getpagesize());
690 return amdgpu_bo_va_op_raw(dev, bo, offset, size, addr,
691 AMDGPU_VM_PAGE_READABLE |
692 AMDGPU_VM_PAGE_WRITEABLE |
693 AMDGPU_VM_PAGE_EXECUTABLE, ops);
696 int amdgpu_bo_va_op_raw(amdgpu_device_handle dev,
704 struct drm_amdgpu_gem_va va;
707 if (ops != AMDGPU_VA_OP_MAP && ops != AMDGPU_VA_OP_UNMAP &&
708 ops != AMDGPU_VA_OP_REPLACE && ops != AMDGPU_VA_OP_CLEAR)
711 memset(&va, 0, sizeof(va));
712 va.handle = bo ? bo->handle : 0;
715 va.va_address = addr;
716 va.offset_in_bo = offset;
719 r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_GEM_VA, &va, sizeof(va));