| /* |
| * Copyright © 2019 Google LLC |
| * SPDX-License-Identifier: MIT |
| */ |
| |
| #include "tu_cs.h" |
| |
| #include "tu_suballoc.h" |
| |
| /** |
| * Initialize a command stream. |
| */ |
| void |
| tu_cs_init(struct tu_cs *cs, |
| struct tu_device *device, |
| enum tu_cs_mode mode, |
| uint32_t initial_size, const char *name) |
| { |
| assert(mode != TU_CS_MODE_EXTERNAL); |
| |
| memset(cs, 0, sizeof(*cs)); |
| |
| cs->device = device; |
| cs->mode = mode; |
| cs->next_bo_size = initial_size; |
| cs->name = name; |
| } |
| |
| /** |
| * Initialize a command stream as a wrapper to an external buffer. |
| */ |
| void |
| tu_cs_init_external(struct tu_cs *cs, struct tu_device *device, |
| uint32_t *start, uint32_t *end) |
| { |
| memset(cs, 0, sizeof(*cs)); |
| |
| cs->device = device; |
| cs->mode = TU_CS_MODE_EXTERNAL; |
| cs->start = cs->reserved_end = cs->cur = start; |
| cs->end = end; |
| } |
| |
| /** |
| * Initialize a sub-command stream as a wrapper to an externally sub-allocated |
| * buffer. |
| */ |
| void |
| tu_cs_init_suballoc(struct tu_cs *cs, struct tu_device *device, |
| struct tu_suballoc_bo *suballoc_bo) |
| { |
| uint32_t *start = tu_suballoc_bo_map(suballoc_bo); |
| uint32_t *end = start + (suballoc_bo->size >> 2); |
| |
| memset(cs, 0, sizeof(*cs)); |
| cs->device = device; |
| cs->mode = TU_CS_MODE_SUB_STREAM; |
| cs->start = cs->reserved_end = cs->cur = start; |
| cs->end = end; |
| cs->refcount_bo = tu_bo_get_ref(suballoc_bo->bo); |
| } |
| |
| /** |
| * Finish and release all resources owned by a command stream. |
| */ |
| void |
| tu_cs_finish(struct tu_cs *cs) |
| { |
| for (uint32_t i = 0; i < cs->bo_count; ++i) { |
| tu_bo_finish(cs->device, cs->bos[i]); |
| } |
| |
| if (cs->refcount_bo) |
| tu_bo_finish(cs->device, cs->refcount_bo); |
| |
| free(cs->entries); |
| free(cs->bos); |
| } |
| |
| static struct tu_bo * |
| tu_cs_current_bo(const struct tu_cs *cs) |
| { |
| if (cs->refcount_bo) { |
| return cs->refcount_bo; |
| } else { |
| assert(cs->bo_count); |
| return cs->bos[cs->bo_count - 1]; |
| } |
| } |
| |
| /** |
| * Get the offset of the command packets emitted since the last call to |
| * tu_cs_add_entry. |
| */ |
| static uint32_t |
| tu_cs_get_offset(const struct tu_cs *cs) |
| { |
| return cs->start - (uint32_t *) tu_cs_current_bo(cs)->map; |
| } |
| |
| /* |
| * Allocate and add a BO to a command stream. Following command packets will |
| * be emitted to the new BO. |
| */ |
| static VkResult |
| tu_cs_add_bo(struct tu_cs *cs, uint32_t size) |
| { |
| /* no BO for TU_CS_MODE_EXTERNAL */ |
| assert(cs->mode != TU_CS_MODE_EXTERNAL); |
| /* No adding more BOs if suballocating from a suballoc_bo. */ |
| assert(!cs->refcount_bo); |
| |
| /* no dangling command packet */ |
| assert(tu_cs_is_empty(cs)); |
| |
| /* grow cs->bos if needed */ |
| if (cs->bo_count == cs->bo_capacity) { |
| uint32_t new_capacity = MAX2(4, 2 * cs->bo_capacity); |
| struct tu_bo **new_bos = |
| realloc(cs->bos, new_capacity * sizeof(struct tu_bo *)); |
| if (!new_bos) |
| return VK_ERROR_OUT_OF_HOST_MEMORY; |
| |
| cs->bo_capacity = new_capacity; |
| cs->bos = new_bos; |
| } |
| |
| struct tu_bo *new_bo; |
| |
| VkResult result = |
| tu_bo_init_new(cs->device, &new_bo, size * sizeof(uint32_t), |
| TU_BO_ALLOC_GPU_READ_ONLY | TU_BO_ALLOC_ALLOW_DUMP, cs->name); |
| if (result != VK_SUCCESS) { |
| return result; |
| } |
| |
| result = tu_bo_map(cs->device, new_bo); |
| if (result != VK_SUCCESS) { |
| tu_bo_finish(cs->device, new_bo); |
| return result; |
| } |
| |
| cs->bos[cs->bo_count++] = new_bo; |
| |
| cs->start = cs->cur = cs->reserved_end = (uint32_t *) new_bo->map; |
| cs->end = cs->start + new_bo->size / sizeof(uint32_t); |
| |
| return VK_SUCCESS; |
| } |
| |
| /** |
| * Reserve an IB entry. |
| */ |
| static VkResult |
| tu_cs_reserve_entry(struct tu_cs *cs) |
| { |
| /* entries are only for TU_CS_MODE_GROW */ |
| assert(cs->mode == TU_CS_MODE_GROW); |
| |
| /* grow cs->entries if needed */ |
| if (cs->entry_count == cs->entry_capacity) { |
| uint32_t new_capacity = MAX2(4, cs->entry_capacity * 2); |
| struct tu_cs_entry *new_entries = |
| realloc(cs->entries, new_capacity * sizeof(struct tu_cs_entry)); |
| if (!new_entries) |
| return VK_ERROR_OUT_OF_HOST_MEMORY; |
| |
| cs->entry_capacity = new_capacity; |
| cs->entries = new_entries; |
| } |
| |
| return VK_SUCCESS; |
| } |
| |
| /** |
| * Add an IB entry for the command packets emitted since the last call to this |
| * function. |
| */ |
| static void |
| tu_cs_add_entry(struct tu_cs *cs) |
| { |
| /* entries are only for TU_CS_MODE_GROW */ |
| assert(cs->mode == TU_CS_MODE_GROW); |
| |
| /* disallow empty entry */ |
| assert(!tu_cs_is_empty(cs)); |
| |
| /* |
| * because we disallow empty entry, tu_cs_add_bo and tu_cs_reserve_entry |
| * must both have been called |
| */ |
| assert(cs->bo_count); |
| assert(cs->entry_count < cs->entry_capacity); |
| |
| /* add an entry for [cs->start, cs->cur] */ |
| cs->entries[cs->entry_count++] = (struct tu_cs_entry) { |
| .bo = tu_cs_current_bo(cs), |
| .size = tu_cs_get_size(cs) * sizeof(uint32_t), |
| .offset = tu_cs_get_offset(cs) * sizeof(uint32_t), |
| }; |
| |
| cs->start = cs->cur; |
| } |
| |
| /** |
| * same behavior as tu_cs_emit_call but without the indirect |
| */ |
| VkResult |
| tu_cs_add_entries(struct tu_cs *cs, struct tu_cs *target) |
| { |
| VkResult result; |
| |
| assert(cs->mode == TU_CS_MODE_GROW); |
| assert(target->mode == TU_CS_MODE_GROW); |
| |
| if (!tu_cs_is_empty(cs)) |
| tu_cs_add_entry(cs); |
| |
| for (unsigned i = 0; i < target->entry_count; i++) { |
| result = tu_cs_reserve_entry(cs); |
| if (result != VK_SUCCESS) |
| return result; |
| cs->entries[cs->entry_count++] = target->entries[i]; |
| } |
| |
| return VK_SUCCESS; |
| } |
| |
| /** |
| * Begin (or continue) command packet emission. This does nothing but sanity |
| * checks currently. \a cs must not be in TU_CS_MODE_SUB_STREAM mode. |
| */ |
| void |
| tu_cs_begin(struct tu_cs *cs) |
| { |
| assert(cs->mode != TU_CS_MODE_SUB_STREAM); |
| assert(tu_cs_is_empty(cs)); |
| } |
| |
| /** |
| * End command packet emission. This adds an IB entry when \a cs is in |
| * TU_CS_MODE_GROW mode. |
| */ |
| void |
| tu_cs_end(struct tu_cs *cs) |
| { |
| assert(cs->mode != TU_CS_MODE_SUB_STREAM); |
| |
| if (cs->mode == TU_CS_MODE_GROW && !tu_cs_is_empty(cs)) |
| tu_cs_add_entry(cs); |
| } |
| |
| /** |
| * Begin command packet emission to a sub-stream. \a cs must be in |
| * TU_CS_MODE_SUB_STREAM mode. |
| * |
| * Return \a sub_cs which is in TU_CS_MODE_EXTERNAL mode. tu_cs_begin and |
| * tu_cs_reserve_space are implied and \a sub_cs is ready for command packet |
| * emission. |
| */ |
| VkResult |
| tu_cs_begin_sub_stream(struct tu_cs *cs, uint32_t size, struct tu_cs *sub_cs) |
| { |
| assert(cs->mode == TU_CS_MODE_SUB_STREAM); |
| assert(size); |
| |
| VkResult result = tu_cs_reserve_space(cs, size); |
| if (result != VK_SUCCESS) |
| return result; |
| |
| tu_cs_init_external(sub_cs, cs->device, cs->cur, cs->reserved_end); |
| tu_cs_begin(sub_cs); |
| result = tu_cs_reserve_space(sub_cs, size); |
| assert(result == VK_SUCCESS); |
| |
| return VK_SUCCESS; |
| } |
| |
| /** |
| * Allocate count*size dwords, aligned to size dwords. |
| * \a cs must be in TU_CS_MODE_SUB_STREAM mode. |
| * |
| */ |
| VkResult |
| tu_cs_alloc(struct tu_cs *cs, |
| uint32_t count, |
| uint32_t size, |
| struct tu_cs_memory *memory) |
| { |
| assert(cs->mode == TU_CS_MODE_SUB_STREAM); |
| assert(size && size <= 1024); |
| |
| if (!count) |
| return VK_SUCCESS; |
| |
| /* TODO: smarter way to deal with alignment? */ |
| |
| VkResult result = tu_cs_reserve_space(cs, count * size + (size-1)); |
| if (result != VK_SUCCESS) |
| return result; |
| |
| struct tu_bo *bo = tu_cs_current_bo(cs); |
| size_t offset = align(tu_cs_get_offset(cs), size); |
| |
| memory->map = bo->map + offset * sizeof(uint32_t); |
| memory->iova = bo->iova + offset * sizeof(uint32_t); |
| |
| cs->start = cs->cur = (uint32_t*) bo->map + offset + count * size; |
| |
| return VK_SUCCESS; |
| } |
| |
| /** |
| * End command packet emission to a sub-stream. \a sub_cs becomes invalid |
| * after this call. |
| * |
| * Return an IB entry for the sub-stream. The entry has the same lifetime as |
| * \a cs. |
| */ |
| struct tu_cs_entry |
| tu_cs_end_sub_stream(struct tu_cs *cs, struct tu_cs *sub_cs) |
| { |
| assert(cs->mode == TU_CS_MODE_SUB_STREAM); |
| assert(sub_cs->start == cs->cur && sub_cs->end == cs->reserved_end); |
| tu_cs_sanity_check(sub_cs); |
| |
| tu_cs_end(sub_cs); |
| |
| cs->cur = sub_cs->cur; |
| |
| struct tu_cs_entry entry = { |
| .bo = tu_cs_current_bo(cs), |
| .size = tu_cs_get_size(cs) * sizeof(uint32_t), |
| .offset = tu_cs_get_offset(cs) * sizeof(uint32_t), |
| }; |
| |
| cs->start = cs->cur; |
| |
| return entry; |
| } |
| |
| /** |
| * Reserve space from a command stream for \a reserved_size uint32_t values. |
| * This never fails when \a cs has mode TU_CS_MODE_EXTERNAL. |
| */ |
| VkResult |
| tu_cs_reserve_space(struct tu_cs *cs, uint32_t reserved_size) |
| { |
| if (tu_cs_get_space(cs) < reserved_size) { |
| if (cs->mode == TU_CS_MODE_EXTERNAL) { |
| unreachable("cannot grow external buffer"); |
| return VK_ERROR_OUT_OF_HOST_MEMORY; |
| } |
| |
| /* add an entry for the exiting command packets */ |
| if (!tu_cs_is_empty(cs)) { |
| /* no direct command packet for TU_CS_MODE_SUB_STREAM */ |
| assert(cs->mode != TU_CS_MODE_SUB_STREAM); |
| |
| tu_cs_add_entry(cs); |
| } |
| |
| for (uint32_t i = 0; i < cs->cond_stack_depth; i++) { |
| /* Subtract one here to account for the DWORD field itself. */ |
| *cs->cond_dwords[i] = cs->cur - cs->cond_dwords[i] - 1; |
| |
| /* space for CP_COND_REG_EXEC in next bo */ |
| reserved_size += 3; |
| } |
| |
| /* switch to a new BO */ |
| uint32_t new_size = MAX2(cs->next_bo_size, reserved_size); |
| VkResult result = tu_cs_add_bo(cs, new_size); |
| if (result != VK_SUCCESS) |
| return result; |
| |
| if (cs->cond_stack_depth) { |
| cs->reserved_end = cs->cur + reserved_size; |
| } |
| |
| /* Re-emit CP_COND_REG_EXECs */ |
| for (uint32_t i = 0; i < cs->cond_stack_depth; i++) { |
| tu_cs_emit_pkt7(cs, CP_COND_REG_EXEC, 2); |
| tu_cs_emit(cs, cs->cond_flags[i]); |
| |
| cs->cond_dwords[i] = cs->cur; |
| |
| /* Emit dummy DWORD field here */ |
| tu_cs_emit(cs, CP_COND_REG_EXEC_1_DWORDS(0)); |
| } |
| |
| /* double the size for the next bo, also there is an upper |
| * bound on IB size, which appears to be 0x0fffff |
| */ |
| new_size = MIN2(new_size << 1, 0x0fffff); |
| if (cs->next_bo_size < new_size) |
| cs->next_bo_size = new_size; |
| } |
| |
| assert(tu_cs_get_space(cs) >= reserved_size); |
| cs->reserved_end = cs->cur + reserved_size; |
| |
| if (cs->mode == TU_CS_MODE_GROW) { |
| /* reserve an entry for the next call to this function or tu_cs_end */ |
| return tu_cs_reserve_entry(cs); |
| } |
| |
| return VK_SUCCESS; |
| } |
| |
| /** |
| * Reset a command stream to its initial state. This discards all comand |
| * packets in \a cs, but does not necessarily release all resources. |
| */ |
| void |
| tu_cs_reset(struct tu_cs *cs) |
| { |
| if (cs->mode == TU_CS_MODE_EXTERNAL) { |
| assert(!cs->bo_count && !cs->refcount_bo && !cs->entry_count); |
| cs->reserved_end = cs->cur = cs->start; |
| return; |
| } |
| |
| for (uint32_t i = 0; i + 1 < cs->bo_count; ++i) { |
| tu_bo_finish(cs->device, cs->bos[i]); |
| } |
| |
| if (cs->bo_count) { |
| cs->bos[0] = cs->bos[cs->bo_count - 1]; |
| cs->bo_count = 1; |
| |
| cs->start = cs->cur = cs->reserved_end = (uint32_t *) cs->bos[0]->map; |
| cs->end = cs->start + cs->bos[0]->size / sizeof(uint32_t); |
| } |
| |
| cs->entry_count = 0; |
| } |