| /* |
| * Copyright © 2023 Intel Corporation |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a |
| * copy of this software and associated documentation files (the "Software"), |
| * to deal in the Software without restriction, including without limitation |
| * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
| * and/or sell copies of the Software, and to permit persons to whom the |
| * Software is furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be included |
| * in all copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS |
| * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING |
| * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER |
| * DEALINGS IN THE SOFTWARE. |
| */ |
| |
| #include "xe/anv_queue.h" |
| |
| #include "anv_private.h" |
| |
| #include "common/xe/intel_engine.h" |
| #include "common/xe/intel_gem.h" |
| #include "common/xe/intel_queue.h" |
| #include "common/intel_gem.h" |
| |
| #include "xe/anv_device.h" |
| |
| #include "drm-uapi/xe_drm.h" |
| #include "drm-uapi/gpu_scheduler.h" |
| |
| static enum drm_sched_priority |
| anv_vk_priority_to_drm_sched_priority(VkQueueGlobalPriorityKHR vk_priority) |
| { |
| switch (vk_priority) { |
| case VK_QUEUE_GLOBAL_PRIORITY_LOW_KHR: |
| return DRM_SCHED_PRIORITY_MIN; |
| case VK_QUEUE_GLOBAL_PRIORITY_MEDIUM_KHR: |
| return DRM_SCHED_PRIORITY_NORMAL; |
| case VK_QUEUE_GLOBAL_PRIORITY_HIGH_KHR: |
| return DRM_SCHED_PRIORITY_HIGH; |
| default: |
| unreachable("Invalid priority"); |
| return DRM_SCHED_PRIORITY_MIN; |
| } |
| } |
| |
| static void |
| destroy_engine(struct anv_device *device, uint32_t exec_queue_id) |
| { |
| struct drm_xe_exec_queue_destroy destroy = { |
| .exec_queue_id = exec_queue_id, |
| }; |
| |
| intel_ioctl(device->fd, DRM_IOCTL_XE_EXEC_QUEUE_DESTROY, &destroy); |
| } |
| |
| static VkResult |
| create_engine(struct anv_device *device, |
| struct anv_queue *queue, |
| const VkDeviceQueueCreateInfo *pCreateInfo, |
| bool create_companion_rcs_engine) |
| { |
| struct anv_physical_device *physical = device->physical; |
| uint32_t queue_family_index = |
| create_companion_rcs_engine ? |
| anv_get_first_render_queue_index(physical) : |
| pCreateInfo->queueFamilyIndex; |
| struct anv_queue_family *queue_family = |
| &physical->queue.families[queue_family_index]; |
| const struct intel_query_engine_info *engines = physical->engine_info; |
| struct drm_xe_engine_class_instance *instances; |
| const VkDeviceQueueGlobalPriorityCreateInfoKHR *queue_priority = |
| vk_find_struct_const(pCreateInfo->pNext, |
| DEVICE_QUEUE_GLOBAL_PRIORITY_CREATE_INFO_KHR); |
| const VkQueueGlobalPriorityKHR priority = queue_priority ? |
| queue_priority->globalPriority : |
| VK_QUEUE_GLOBAL_PRIORITY_MEDIUM_KHR; |
| |
| /* As per spec, the driver implementation may deny requests to acquire |
| * a priority above the default priority (MEDIUM) if the caller does not |
| * have sufficient privileges. In this scenario VK_ERROR_NOT_PERMITTED_KHR |
| * is returned. |
| */ |
| if (physical->max_context_priority >= VK_QUEUE_GLOBAL_PRIORITY_MEDIUM_KHR) { |
| if (priority > physical->max_context_priority) |
| return vk_error(device, VK_ERROR_NOT_PERMITTED_KHR); |
| } |
| |
| instances = vk_alloc(&device->vk.alloc, |
| sizeof(*instances) * queue_family->queueCount, 8, |
| VK_SYSTEM_ALLOCATION_SCOPE_DEVICE); |
| if (!instances) |
| return VK_ERROR_OUT_OF_HOST_MEMORY; |
| |
| /* Build a list of all compatible HW engines */ |
| uint32_t count = 0; |
| for (uint32_t i = 0; i < engines->num_engines; i++) { |
| const struct intel_engine_class_instance engine = engines->engines[i]; |
| if (engine.engine_class != queue_family->engine_class) |
| continue; |
| |
| instances[count].engine_class = intel_engine_class_to_xe(engine.engine_class); |
| instances[count].engine_instance = engine.engine_instance; |
| instances[count++].gt_id = engine.gt_id; |
| } |
| |
| assert(device->vm_id != 0); |
| const bool pxp_needed = pCreateInfo->flags & VK_DEVICE_QUEUE_CREATE_PROTECTED_BIT; |
| struct drm_xe_ext_set_property priority_ext = { |
| .property = DRM_XE_EXEC_QUEUE_SET_PROPERTY_PRIORITY, |
| .value = anv_vk_priority_to_drm_sched_priority(priority), |
| }; |
| struct drm_xe_ext_set_property pxp_ext = { |
| .property = DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE, |
| .value = DRM_XE_PXP_TYPE_HWDRM, |
| }; |
| struct drm_xe_exec_queue_create create = { |
| /* Allows KMD to pick one of those engines for the submission queue */ |
| .instances = (uintptr_t)instances, |
| .vm_id = device->vm_id, |
| .width = 1, |
| .num_placements = count, |
| }; |
| intel_xe_gem_add_ext((uint64_t *)&create.extensions, |
| DRM_XE_EXEC_QUEUE_EXTENSION_SET_PROPERTY, |
| &priority_ext.base); |
| if (pxp_needed) |
| intel_xe_gem_add_ext((uint64_t *)&create.extensions, |
| DRM_XE_EXEC_QUEUE_EXTENSION_SET_PROPERTY, |
| &pxp_ext.base); |
| int ret; |
| bool retry; |
| do { |
| ret = intel_ioctl(device->fd, DRM_IOCTL_XE_EXEC_QUEUE_CREATE, &create); |
| retry = pxp_needed && ret == -1 && errno == EBUSY; |
| if (retry) |
| usleep(1000); |
| } while (retry); |
| vk_free(&device->vk.alloc, instances); |
| if (ret) |
| return vk_errorf(device, VK_ERROR_UNKNOWN, "Unable to create exec queue"); |
| |
| if (create_companion_rcs_engine) |
| queue->companion_rcs_id = create.exec_queue_id; |
| else |
| queue->exec_queue_id = create.exec_queue_id; |
| |
| if (!create_companion_rcs_engine && |
| queue_family->queueFlags & VK_QUEUE_SPARSE_BINDING_BIT) { |
| struct drm_xe_engine_class_instance bind_instance = { |
| .engine_class = DRM_XE_ENGINE_CLASS_VM_BIND, |
| }; |
| create.num_placements = 1; |
| create.instances = (uintptr_t)&bind_instance; |
| create.extensions = 0; |
| ret = intel_ioctl(device->fd, DRM_IOCTL_XE_EXEC_QUEUE_CREATE, &create); |
| if (ret) { |
| destroy_engine(device, queue->exec_queue_id); |
| return vk_errorf(device, VK_ERROR_UNKNOWN, "Unable to create bind queue"); |
| } |
| queue->bind_queue_id = create.exec_queue_id; |
| } |
| |
| return VK_SUCCESS; |
| } |
| |
| VkResult |
| anv_xe_create_engine(struct anv_device *device, |
| struct anv_queue *queue, |
| const VkDeviceQueueCreateInfo *pCreateInfo) |
| { |
| VkResult result = create_engine(device, queue, pCreateInfo, |
| false /* create_companion_rcs_engine */); |
| |
| if (result != VK_SUCCESS) |
| return result; |
| |
| if (queue->family->engine_class == INTEL_ENGINE_CLASS_COPY || |
| queue->family->engine_class == INTEL_ENGINE_CLASS_COMPUTE) { |
| result = create_engine(device, queue, pCreateInfo, |
| true /* create_companion_rcs_engine */); |
| } |
| |
| return result; |
| } |
| |
| /* |
| * Wait for all previous DRM_IOCTL_XE_EXEC calls over the |
| * drm_xe_exec_queue to complete. |
| **/ |
| int |
| anv_xe_wait_exec_queue_idle(struct anv_device *device, uint32_t exec_queue_id) |
| { |
| struct drm_syncobj_wait syncobj_wait = { |
| .count_handles = 1, |
| .timeout_nsec = INT64_MAX, |
| }; |
| uint32_t syncobj; |
| int ret = xe_queue_get_syncobj_for_idle(device->fd, exec_queue_id, &syncobj); |
| |
| if (ret) { |
| assert(ret == -ECANCELED); |
| return ret; |
| } |
| |
| syncobj_wait.handles = (uintptr_t)&syncobj; |
| ret = intel_ioctl(device->fd, DRM_IOCTL_SYNCOBJ_WAIT, &syncobj_wait); |
| assert(ret == 0); |
| |
| struct drm_syncobj_destroy syncobj_destroy = { |
| .handle = syncobj, |
| }; |
| ret = intel_ioctl(device->fd, DRM_IOCTL_SYNCOBJ_DESTROY, &syncobj_destroy); |
| assert(ret == 0); |
| |
| return ret; |
| } |
| |
| static void |
| bind_engine_idle(struct anv_device *device, uint32_t exec_queue_id) |
| { |
| struct drm_syncobj_create syncobj_create = {}; |
| struct drm_xe_sync xe_sync = { |
| .type = DRM_XE_SYNC_TYPE_SYNCOBJ, |
| .flags = DRM_XE_SYNC_FLAG_SIGNAL, |
| }; |
| struct drm_xe_vm_bind args = { |
| .vm_id = device->vm_id, |
| .num_binds = 0, |
| .exec_queue_id = exec_queue_id, |
| .bind = {}, |
| .num_syncs = 1, |
| .syncs = (uintptr_t)&xe_sync, |
| }; |
| struct drm_syncobj_destroy syncobj_destroy = {}; |
| int ret = intel_ioctl(device->fd, DRM_IOCTL_SYNCOBJ_CREATE, &syncobj_create); |
| |
| assert(ret == 0); |
| |
| xe_sync.handle = syncobj_create.handle; |
| /* Using the special args.num_binds == 0 handling to get syncobj |
| * signaled when the last DRM_IOCTL_XE_VM_BIND is completed. |
| */ |
| ret = intel_ioctl(device->fd, DRM_IOCTL_XE_VM_BIND, &args); |
| if (ret) { |
| /* exec_queue could have been banned, that is why it is being destroyed |
| * so no assert() here |
| */ |
| goto error_bind; |
| } |
| |
| struct drm_syncobj_wait syncobj_wait = { |
| .count_handles = 1, |
| .timeout_nsec = INT64_MAX, |
| .handles = (uintptr_t)&syncobj_create.handle, |
| }; |
| ret = intel_ioctl(device->fd, DRM_IOCTL_SYNCOBJ_WAIT, &syncobj_wait); |
| assert(ret == 0); |
| |
| error_bind: |
| syncobj_destroy.handle = syncobj_create.handle, |
| intel_ioctl(device->fd, DRM_IOCTL_SYNCOBJ_DESTROY, &syncobj_destroy); |
| } |
| |
| void |
| anv_xe_destroy_engine(struct anv_device *device, struct anv_queue *queue) |
| { |
| /* Application could submit a workload and before it is done, destroy the |
| * queue causing job timeouts in Xe KMD as it don't have permanent |
| * exec queues. |
| */ |
| anv_xe_wait_exec_queue_idle(device, queue->exec_queue_id); |
| destroy_engine(device, queue->exec_queue_id); |
| |
| if (queue->companion_rcs_id != 0) { |
| anv_xe_wait_exec_queue_idle(device, queue->companion_rcs_id); |
| destroy_engine(device, queue->companion_rcs_id); |
| } |
| if (queue->bind_queue_id != 0) { |
| bind_engine_idle(device, queue->bind_queue_id); |
| destroy_engine(device, queue->bind_queue_id); |
| } |
| } |