We mostly use the same priority for all buffer objects, so I don't think that matter much. This should reduce CPU overhead a little bit. Signed-off-by: Samuel Pitoiset <samuel.pitoiset@gmail.com> Reviewed-by: Bas Nieuwenhuizen <bas@basnieuwenhuizen.nl>tags/18.2-branchpoint
@@ -305,7 +305,7 @@ radv_reset_cmd_buffer(struct radv_cmd_buffer *cmd_buffer) | |||
if (cmd_buffer->upload.upload_bo) | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, | |||
cmd_buffer->upload.upload_bo, 8); | |||
cmd_buffer->upload.upload_bo); | |||
cmd_buffer->upload.offset = 0; | |||
cmd_buffer->record_result = VK_SUCCESS; | |||
@@ -365,7 +365,7 @@ radv_cmd_buffer_resize_upload_buf(struct radv_cmd_buffer *cmd_buffer, | |||
return false; | |||
} | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, bo, 8); | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, bo); | |||
if (cmd_buffer->upload.upload_bo) { | |||
upload = malloc(sizeof(*upload)); | |||
@@ -872,12 +872,12 @@ radv_emit_graphics_pipeline(struct radv_cmd_buffer *cmd_buffer) | |||
continue; | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, | |||
pipeline->shaders[i]->bo, 8); | |||
pipeline->shaders[i]->bo); | |||
} | |||
if (radv_pipeline_has_gs(pipeline)) | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, | |||
pipeline->gs_copy_shader->bo, 8); | |||
pipeline->gs_copy_shader->bo); | |||
if (unlikely(cmd_buffer->device->trace_bo)) | |||
radv_save_pipeline(cmd_buffer, pipeline, RING_GFX); | |||
@@ -1459,7 +1459,7 @@ radv_emit_framebuffer_state(struct radv_cmd_buffer *cmd_buffer) | |||
struct radv_image *image = att->attachment->image; | |||
VkImageLayout layout = subpass->color_attachments[i].layout; | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, att->attachment->bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, att->attachment->bo); | |||
assert(att->attachment->aspect_mask & VK_IMAGE_ASPECT_COLOR_BIT); | |||
radv_emit_fb_color_state(cmd_buffer, i, att, image, layout); | |||
@@ -1472,7 +1472,7 @@ radv_emit_framebuffer_state(struct radv_cmd_buffer *cmd_buffer) | |||
VkImageLayout layout = subpass->depth_stencil_attachment.layout; | |||
struct radv_attachment_info *att = &framebuffer->attachments[idx]; | |||
struct radv_image *image = att->attachment->image; | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, att->attachment->bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, att->attachment->bo); | |||
MAYBE_UNUSED uint32_t queue_mask = radv_image_queue_family_mask(image, | |||
cmd_buffer->queue_family_index, | |||
cmd_buffer->queue_family_index); | |||
@@ -2249,7 +2249,7 @@ static void emit_gfx_buffer_state(struct radv_cmd_buffer *cmd_buffer) | |||
struct radv_device *device = cmd_buffer->device; | |||
if (device->gfx_init) { | |||
uint64_t va = radv_buffer_get_va(device->gfx_init); | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, device->gfx_init, 8); | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, device->gfx_init); | |||
radeon_emit(cmd_buffer->cs, PKT3(PKT3_INDIRECT_BUFFER_CIK, 2, 0)); | |||
radeon_emit(cmd_buffer->cs, va); | |||
radeon_emit(cmd_buffer->cs, va >> 32); | |||
@@ -2317,7 +2317,7 @@ VkResult radv_BeginCommandBuffer( | |||
struct radv_device *device = cmd_buffer->device; | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, | |||
device->trace_bo, 8); | |||
device->trace_bo); | |||
radv_cmd_buffer_trace_emit(cmd_buffer); | |||
} | |||
@@ -2355,7 +2355,7 @@ void radv_CmdBindVertexBuffers( | |||
vb[idx].offset = pOffsets[i]; | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, | |||
vb[idx].buffer->bo, 8); | |||
vb[idx].buffer->bo); | |||
} | |||
if (!changed) { | |||
@@ -2391,7 +2391,7 @@ void radv_CmdBindIndexBuffer( | |||
int index_size_shift = cmd_buffer->state.index_type ? 2 : 1; | |||
cmd_buffer->state.max_index_count = (index_buffer->size - offset) >> index_size_shift; | |||
cmd_buffer->state.dirty |= RADV_CMD_DIRTY_INDEX_BUFFER; | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, index_buffer->bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, index_buffer->bo); | |||
} | |||
@@ -2411,11 +2411,11 @@ radv_bind_descriptor_set(struct radv_cmd_buffer *cmd_buffer, | |||
if (!cmd_buffer->device->use_global_bo_list) { | |||
for (unsigned j = 0; j < set->layout->buffer_count; ++j) | |||
if (set->descriptors[j]) | |||
radv_cs_add_buffer(ws, cmd_buffer->cs, set->descriptors[j], 7); | |||
radv_cs_add_buffer(ws, cmd_buffer->cs, set->descriptors[j]); | |||
} | |||
if(set->bo) | |||
radv_cs_add_buffer(ws, cmd_buffer->cs, set->bo, 8); | |||
radv_cs_add_buffer(ws, cmd_buffer->cs, set->bo); | |||
} | |||
void radv_CmdBindDescriptorSets( | |||
@@ -2639,7 +2639,7 @@ radv_emit_compute_pipeline(struct radv_cmd_buffer *cmd_buffer) | |||
pipeline->max_waves * pipeline->scratch_bytes_per_wave); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, | |||
pipeline->shaders[MESA_SHADER_COMPUTE]->bo, 8); | |||
pipeline->shaders[MESA_SHADER_COMPUTE]->bo); | |||
if (unlikely(cmd_buffer->device->trace_bo)) | |||
radv_save_pipeline(cmd_buffer, pipeline, RING_COMPUTE); | |||
@@ -3252,7 +3252,7 @@ radv_emit_draw_packets(struct radv_cmd_buffer *cmd_buffer, | |||
va += info->indirect->offset + info->indirect_offset; | |||
radv_cs_add_buffer(ws, cs, info->indirect->bo, 8); | |||
radv_cs_add_buffer(ws, cs, info->indirect->bo); | |||
radeon_emit(cs, PKT3(PKT3_SET_BASE, 2, 0)); | |||
radeon_emit(cs, 1); | |||
@@ -3264,7 +3264,7 @@ radv_emit_draw_packets(struct radv_cmd_buffer *cmd_buffer, | |||
count_va += info->count_buffer->offset + | |||
info->count_buffer_offset; | |||
radv_cs_add_buffer(ws, cs, info->count_buffer->bo, 8); | |||
radv_cs_add_buffer(ws, cs, info->count_buffer->bo); | |||
} | |||
if (!state->subpass->view_mask) { | |||
@@ -3724,7 +3724,7 @@ radv_emit_dispatch_packets(struct radv_cmd_buffer *cmd_buffer, | |||
va += info->indirect->offset + info->indirect_offset; | |||
radv_cs_add_buffer(ws, cs, info->indirect->bo, 8); | |||
radv_cs_add_buffer(ws, cs, info->indirect->bo); | |||
if (loc->sgpr_idx != -1) { | |||
for (unsigned i = 0; i < 3; ++i) { | |||
@@ -4236,7 +4236,7 @@ radv_barrier(struct radv_cmd_buffer *cmd_buffer, | |||
RADV_FROM_HANDLE(radv_event, event, info->pEvents[i]); | |||
uint64_t va = radv_buffer_get_va(event->bo); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cs, event->bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cs, event->bo); | |||
MAYBE_UNUSED unsigned cdw_max = radeon_check_space(cmd_buffer->device->ws, cs, 7); | |||
@@ -4322,7 +4322,7 @@ static void write_event(struct radv_cmd_buffer *cmd_buffer, | |||
struct radeon_cmdbuf *cs = cmd_buffer->cs; | |||
uint64_t va = radv_buffer_get_va(event->bo); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cs, event->bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cs, event->bo); | |||
MAYBE_UNUSED unsigned cdw_max = radeon_check_space(cmd_buffer->device->ws, cs, 18); | |||
@@ -721,7 +721,7 @@ static void write_texel_buffer_descriptor(struct radv_device *device, | |||
memcpy(dst, buffer_view->state, 4 * 4); | |||
if (cmd_buffer) | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, buffer_view->bo, 7); | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, buffer_view->bo); | |||
else | |||
*buffer_list = buffer_view->bo; | |||
} | |||
@@ -751,7 +751,7 @@ static void write_buffer_descriptor(struct radv_device *device, | |||
S_008F0C_DATA_FORMAT(V_008F0C_BUF_DATA_FORMAT_32); | |||
if (cmd_buffer) | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, buffer->bo, 7); | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, buffer->bo); | |||
else | |||
*buffer_list = buffer->bo; | |||
} | |||
@@ -795,7 +795,7 @@ write_image_descriptor(struct radv_device *device, | |||
memcpy(dst, descriptor, 16 * 4); | |||
if (cmd_buffer) | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, iview->bo, 7); | |||
radv_cs_add_buffer(device->ws, cmd_buffer->cs, iview->bo); | |||
else | |||
*buffer_list = iview->bo; | |||
} |
@@ -1928,10 +1928,10 @@ radv_emit_gs_ring_sizes(struct radv_queue *queue, struct radeon_cmdbuf *cs, | |||
return; | |||
if (esgs_ring_bo) | |||
radv_cs_add_buffer(queue->device->ws, cs, esgs_ring_bo, 8); | |||
radv_cs_add_buffer(queue->device->ws, cs, esgs_ring_bo); | |||
if (gsvs_ring_bo) | |||
radv_cs_add_buffer(queue->device->ws, cs, gsvs_ring_bo, 8); | |||
radv_cs_add_buffer(queue->device->ws, cs, gsvs_ring_bo); | |||
if (queue->device->physical_device->rad_info.chip_class >= CIK) { | |||
radeon_set_uconfig_reg_seq(cs, R_030900_VGT_ESGS_RING_SIZE, 2); | |||
@@ -1956,7 +1956,7 @@ radv_emit_tess_factor_ring(struct radv_queue *queue, struct radeon_cmdbuf *cs, | |||
tf_va = radv_buffer_get_va(tess_rings_bo); | |||
radv_cs_add_buffer(queue->device->ws, cs, tess_rings_bo, 8); | |||
radv_cs_add_buffer(queue->device->ws, cs, tess_rings_bo); | |||
if (queue->device->physical_device->rad_info.chip_class >= CIK) { | |||
radeon_set_uconfig_reg(cs, R_030938_VGT_TF_RING_SIZE, | |||
@@ -1990,7 +1990,7 @@ radv_emit_compute_scratch(struct radv_queue *queue, struct radeon_cmdbuf *cs, | |||
scratch_va = radv_buffer_get_va(compute_scratch_bo); | |||
radv_cs_add_buffer(queue->device->ws, cs, compute_scratch_bo, 8); | |||
radv_cs_add_buffer(queue->device->ws, cs, compute_scratch_bo); | |||
radeon_set_sh_reg_seq(cs, R_00B900_COMPUTE_USER_DATA_0, 2); | |||
radeon_emit(cs, scratch_va); | |||
@@ -2010,7 +2010,7 @@ radv_emit_global_shader_pointers(struct radv_queue *queue, | |||
va = radv_buffer_get_va(descriptor_bo); | |||
radv_cs_add_buffer(queue->device->ws, cs, descriptor_bo, 8); | |||
radv_cs_add_buffer(queue->device->ws, cs, descriptor_bo); | |||
if (queue->device->physical_device->rad_info.chip_class >= GFX9) { | |||
uint32_t regs[] = {R_00B030_SPI_SHADER_USER_DATA_PS_0, | |||
@@ -2189,7 +2189,7 @@ radv_get_preamble_cs(struct radv_queue *queue, | |||
dest_cs[i] = cs; | |||
if (scratch_bo) | |||
radv_cs_add_buffer(queue->device->ws, cs, scratch_bo, 8); | |||
radv_cs_add_buffer(queue->device->ws, cs, scratch_bo); | |||
if (descriptor_bo != queue->descriptor_bo) { | |||
uint32_t *map = (uint32_t*)queue->device->ws->buffer_map(descriptor_bo); |
@@ -415,7 +415,7 @@ uint32_t radv_fill_buffer(struct radv_cmd_buffer *cmd_buffer, | |||
} else if (size) { | |||
uint64_t va = radv_buffer_get_va(bo); | |||
va += offset; | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, bo); | |||
si_cp_dma_clear_buffer(cmd_buffer, va, size, value); | |||
} | |||
@@ -438,8 +438,8 @@ void radv_copy_buffer(struct radv_cmd_buffer *cmd_buffer, | |||
src_va += src_offset; | |||
dst_va += dst_offset; | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, src_bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, dst_bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, src_bo); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, dst_bo); | |||
si_cp_dma_buffer_copy(cmd_buffer, src_va, dst_va, size); | |||
} | |||
@@ -506,7 +506,7 @@ void radv_CmdUpdateBuffer( | |||
if (dataSize < RADV_BUFFER_OPS_CS_THRESHOLD) { | |||
si_emit_cache_flush(cmd_buffer); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, dst_buffer->bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, dst_buffer->bo); | |||
radeon_check_space(cmd_buffer->device->ws, cmd_buffer->cs, words + 4); | |||
@@ -956,8 +956,8 @@ void radv_CmdCopyQueryPoolResults( | |||
uint64_t dest_va = radv_buffer_get_va(dst_buffer->bo); | |||
dest_va += dst_buffer->offset + dstOffset; | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, pool->bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, dst_buffer->bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, pool->bo); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cmd_buffer->cs, dst_buffer->bo); | |||
switch (pool->type) { | |||
case VK_QUERY_TYPE_OCCLUSION: | |||
@@ -1199,7 +1199,7 @@ void radv_CmdBeginQuery( | |||
struct radeon_cmdbuf *cs = cmd_buffer->cs; | |||
uint64_t va = radv_buffer_get_va(pool->bo); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cs, pool->bo, 8); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cs, pool->bo); | |||
if (cmd_buffer->pending_reset_query) { | |||
if (pool->size >= RADV_BUFFER_OPS_CS_THRESHOLD) { | |||
@@ -1270,7 +1270,7 @@ void radv_CmdWriteTimestamp( | |||
uint64_t avail_va = va + pool->availability_offset + 4 * query; | |||
uint64_t query_va = va + pool->stride * query; | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cs, pool->bo, 5); | |||
radv_cs_add_buffer(cmd_buffer->device->ws, cs, pool->bo); | |||
int num_queries = 1; | |||
if (cmd_buffer->state.subpass && cmd_buffer->state.subpass->view_mask) |
@@ -257,8 +257,7 @@ struct radeon_winsys { | |||
struct radeon_winsys_fence *fence); | |||
void (*cs_add_buffer)(struct radeon_cmdbuf *cs, | |||
struct radeon_winsys_bo *bo, | |||
uint8_t priority); | |||
struct radeon_winsys_bo *bo); | |||
void (*cs_execute_secondary)(struct radeon_cmdbuf *parent, | |||
struct radeon_cmdbuf *child); | |||
@@ -326,13 +325,12 @@ static inline uint64_t radv_buffer_get_va(struct radeon_winsys_bo *bo) | |||
static inline void radv_cs_add_buffer(struct radeon_winsys *ws, | |||
struct radeon_cmdbuf *cs, | |||
struct radeon_winsys_bo *bo, | |||
uint8_t priority) | |||
struct radeon_winsys_bo *bo) | |||
{ | |||
if (bo->is_local) | |||
return; | |||
ws->cs_add_buffer(cs, bo, priority); | |||
ws->cs_add_buffer(cs, bo); | |||
} | |||
#endif /* RADV_RADEON_WINSYS_H */ |
@@ -51,7 +51,6 @@ struct radv_amdgpu_cs { | |||
unsigned max_num_buffers; | |||
unsigned num_buffers; | |||
amdgpu_bo_handle *handles; | |||
uint8_t *priorities; | |||
struct radeon_winsys_bo **old_ib_buffers; | |||
unsigned num_old_ib_buffers; | |||
@@ -66,7 +65,6 @@ struct radv_amdgpu_cs { | |||
unsigned num_virtual_buffers; | |||
unsigned max_num_virtual_buffers; | |||
struct radeon_winsys_bo **virtual_buffers; | |||
uint8_t *virtual_buffer_priorities; | |||
int *virtual_buffer_hash_table; | |||
/* For chips that don't support chaining. */ | |||
@@ -213,10 +211,8 @@ static void radv_amdgpu_cs_destroy(struct radeon_cmdbuf *rcs) | |||
free(cs->old_cs_buffers); | |||
free(cs->old_ib_buffers); | |||
free(cs->virtual_buffers); | |||
free(cs->virtual_buffer_priorities); | |||
free(cs->virtual_buffer_hash_table); | |||
free(cs->handles); | |||
free(cs->priorities); | |||
free(cs); | |||
} | |||
@@ -266,7 +262,7 @@ radv_amdgpu_cs_create(struct radeon_winsys *ws, | |||
cs->ib_size_ptr = &cs->ib.size; | |||
cs->ib.size = 0; | |||
ws->cs_add_buffer(&cs->base, cs->ib_buffer, 8); | |||
ws->cs_add_buffer(&cs->base, cs->ib_buffer); | |||
} else { | |||
cs->base.buf = malloc(16384); | |||
cs->base.max_dw = 4096; | |||
@@ -387,7 +383,7 @@ static void radv_amdgpu_cs_grow(struct radeon_cmdbuf *_cs, size_t min_size) | |||
cs->ib_buffer = cs->old_ib_buffers[--cs->num_old_ib_buffers]; | |||
} | |||
cs->ws->base.cs_add_buffer(&cs->base, cs->ib_buffer, 8); | |||
cs->ws->base.cs_add_buffer(&cs->base, cs->ib_buffer); | |||
radeon_emit(&cs->base, PKT3(PKT3_INDIRECT_BUFFER_CIK, 2, 0)); | |||
radeon_emit(&cs->base, radv_amdgpu_winsys_bo(cs->ib_buffer)->base.va); | |||
@@ -439,7 +435,7 @@ static void radv_amdgpu_cs_reset(struct radeon_cmdbuf *_cs) | |||
cs->num_virtual_buffers = 0; | |||
if (cs->ws->use_ib_bos) { | |||
cs->ws->base.cs_add_buffer(&cs->base, cs->ib_buffer, 8); | |||
cs->ws->base.cs_add_buffer(&cs->base, cs->ib_buffer); | |||
for (unsigned i = 0; i < cs->num_old_ib_buffers; ++i) | |||
cs->ws->base.buffer_destroy(cs->old_ib_buffers[i]); | |||
@@ -483,26 +479,21 @@ static int radv_amdgpu_cs_find_buffer(struct radv_amdgpu_cs *cs, | |||
} | |||
static void radv_amdgpu_cs_add_buffer_internal(struct radv_amdgpu_cs *cs, | |||
amdgpu_bo_handle bo, | |||
uint8_t priority) | |||
amdgpu_bo_handle bo) | |||
{ | |||
unsigned hash; | |||
int index = radv_amdgpu_cs_find_buffer(cs, bo); | |||
if (index != -1) { | |||
cs->priorities[index] = MAX2(cs->priorities[index], priority); | |||
if (index != -1) | |||
return; | |||
} | |||
if (cs->num_buffers == cs->max_num_buffers) { | |||
unsigned new_count = MAX2(1, cs->max_num_buffers * 2); | |||
cs->handles = realloc(cs->handles, new_count * sizeof(amdgpu_bo_handle)); | |||
cs->priorities = realloc(cs->priorities, new_count * sizeof(uint8_t)); | |||
cs->max_num_buffers = new_count; | |||
} | |||
cs->handles[cs->num_buffers] = bo; | |||
cs->priorities[cs->num_buffers] = priority; | |||
hash = ((uintptr_t)bo >> 6) & (ARRAY_SIZE(cs->buffer_hash_table) - 1); | |||
cs->buffer_hash_table[hash] = cs->num_buffers; | |||
@@ -511,8 +502,7 @@ static void radv_amdgpu_cs_add_buffer_internal(struct radv_amdgpu_cs *cs, | |||
} | |||
static void radv_amdgpu_cs_add_virtual_buffer(struct radeon_cmdbuf *_cs, | |||
struct radeon_winsys_bo *bo, | |||
uint8_t priority) | |||
struct radeon_winsys_bo *bo) | |||
{ | |||
struct radv_amdgpu_cs *cs = radv_amdgpu_cs(_cs); | |||
unsigned hash = ((uintptr_t)bo >> 6) & (VIRTUAL_BUFFER_HASH_TABLE_SIZE - 1); | |||
@@ -527,12 +517,10 @@ static void radv_amdgpu_cs_add_virtual_buffer(struct radeon_cmdbuf *_cs, | |||
if (cs->virtual_buffer_hash_table[hash] >= 0) { | |||
int idx = cs->virtual_buffer_hash_table[hash]; | |||
if (cs->virtual_buffers[idx] == bo) { | |||
cs->virtual_buffer_priorities[idx] = MAX2(cs->virtual_buffer_priorities[idx], priority); | |||
return; | |||
} | |||
for (unsigned i = 0; i < cs->num_virtual_buffers; ++i) { | |||
if (cs->virtual_buffers[i] == bo) { | |||
cs->virtual_buffer_priorities[i] = MAX2(cs->virtual_buffer_priorities[i], priority); | |||
cs->virtual_buffer_hash_table[hash] = i; | |||
return; | |||
} | |||
@@ -542,11 +530,9 @@ static void radv_amdgpu_cs_add_virtual_buffer(struct radeon_cmdbuf *_cs, | |||
if(cs->max_num_virtual_buffers <= cs->num_virtual_buffers) { | |||
cs->max_num_virtual_buffers = MAX2(2, cs->max_num_virtual_buffers * 2); | |||
cs->virtual_buffers = realloc(cs->virtual_buffers, sizeof(struct radv_amdgpu_virtual_virtual_buffer*) * cs->max_num_virtual_buffers); | |||
cs->virtual_buffer_priorities = realloc(cs->virtual_buffer_priorities, sizeof(uint8_t) * cs->max_num_virtual_buffers); | |||
} | |||
cs->virtual_buffers[cs->num_virtual_buffers] = bo; | |||
cs->virtual_buffer_priorities[cs->num_virtual_buffers] = priority; | |||
cs->virtual_buffer_hash_table[hash] = cs->num_virtual_buffers; | |||
++cs->num_virtual_buffers; | |||
@@ -554,21 +540,20 @@ static void radv_amdgpu_cs_add_virtual_buffer(struct radeon_cmdbuf *_cs, | |||
} | |||
static void radv_amdgpu_cs_add_buffer(struct radeon_cmdbuf *_cs, | |||
struct radeon_winsys_bo *_bo, | |||
uint8_t priority) | |||
struct radeon_winsys_bo *_bo) | |||
{ | |||
struct radv_amdgpu_cs *cs = radv_amdgpu_cs(_cs); | |||
struct radv_amdgpu_winsys_bo *bo = radv_amdgpu_winsys_bo(_bo); | |||
if (bo->is_virtual) { | |||
radv_amdgpu_cs_add_virtual_buffer(_cs, _bo, priority); | |||
radv_amdgpu_cs_add_virtual_buffer(_cs, _bo); | |||
return; | |||
} | |||
if (bo->base.is_local) | |||
return; | |||
radv_amdgpu_cs_add_buffer_internal(cs, bo->bo, priority); | |||
radv_amdgpu_cs_add_buffer_internal(cs, bo->bo); | |||
} | |||
static void radv_amdgpu_cs_execute_secondary(struct radeon_cmdbuf *_parent, | |||
@@ -578,13 +563,11 @@ static void radv_amdgpu_cs_execute_secondary(struct radeon_cmdbuf *_parent, | |||
struct radv_amdgpu_cs *child = radv_amdgpu_cs(_child); | |||
for (unsigned i = 0; i < child->num_buffers; ++i) { | |||
radv_amdgpu_cs_add_buffer_internal(parent, child->handles[i], | |||
child->priorities[i]); | |||
radv_amdgpu_cs_add_buffer_internal(parent, child->handles[i]); | |||
} | |||
for (unsigned i = 0; i < child->num_virtual_buffers; ++i) { | |||
radv_amdgpu_cs_add_buffer(&parent->base, child->virtual_buffers[i], | |||
child->virtual_buffer_priorities[i]); | |||
radv_amdgpu_cs_add_buffer(&parent->base, child->virtual_buffers[i]); | |||
} | |||
if (parent->ws->use_ib_bos) { | |||
@@ -646,7 +629,7 @@ static int radv_amdgpu_create_bo_list(struct radv_amdgpu_winsys *ws, | |||
return 0; | |||
} | |||
r = amdgpu_bo_list_create(ws->dev, cs->num_buffers, cs->handles, | |||
cs->priorities, bo_list); | |||
NULL, bo_list); | |||
} else { | |||
unsigned total_buffer_count = num_extra_bo; | |||
unsigned unique_bo_count = num_extra_bo; | |||
@@ -670,16 +653,13 @@ static int radv_amdgpu_create_bo_list(struct radv_amdgpu_winsys *ws, | |||
return 0; | |||
} | |||
amdgpu_bo_handle *handles = malloc(sizeof(amdgpu_bo_handle) * total_buffer_count); | |||
uint8_t *priorities = malloc(sizeof(uint8_t) * total_buffer_count); | |||
if (!handles || !priorities) { | |||
if (!handles) { | |||
free(handles); | |||
free(priorities); | |||
return -ENOMEM; | |||
} | |||
for (unsigned i = 0; i < num_extra_bo; i++) { | |||
handles[i] = extra_bo_array[i]->bo; | |||
priorities[i] = 8; | |||
} | |||
for (unsigned i = 0; i < count + !!extra_cs; ++i) { | |||
@@ -695,7 +675,6 @@ static int radv_amdgpu_create_bo_list(struct radv_amdgpu_winsys *ws, | |||
if (unique_bo_count == 0) { | |||
memcpy(handles, cs->handles, cs->num_buffers * sizeof(amdgpu_bo_handle)); | |||
memcpy(priorities, cs->priorities, cs->num_buffers * sizeof(uint8_t)); | |||
unique_bo_count = cs->num_buffers; | |||
continue; | |||
} | |||
@@ -705,14 +684,11 @@ static int radv_amdgpu_create_bo_list(struct radv_amdgpu_winsys *ws, | |||
for (unsigned k = 0; k < unique_bo_so_far; ++k) { | |||
if (handles[k] == cs->handles[j]) { | |||
found = true; | |||
priorities[k] = MAX2(priorities[k], | |||
cs->priorities[j]); | |||
break; | |||
} | |||
} | |||
if (!found) { | |||
handles[unique_bo_count] = cs->handles[j]; | |||
priorities[unique_bo_count] = cs->priorities[j]; | |||
++unique_bo_count; | |||
} | |||
} | |||
@@ -724,14 +700,11 @@ static int radv_amdgpu_create_bo_list(struct radv_amdgpu_winsys *ws, | |||
for (unsigned m = 0; m < unique_bo_count; ++m) { | |||
if (handles[m] == bo->bo) { | |||
found = true; | |||
priorities[m] = MAX2(priorities[m], | |||
cs->virtual_buffer_priorities[j]); | |||
break; | |||
} | |||
} | |||
if (!found) { | |||
handles[unique_bo_count] = bo->bo; | |||
priorities[unique_bo_count] = cs->virtual_buffer_priorities[j]; | |||
++unique_bo_count; | |||
} | |||
} | |||
@@ -740,20 +713,17 @@ static int radv_amdgpu_create_bo_list(struct radv_amdgpu_winsys *ws, | |||
if (radv_bo_list) { | |||
unsigned unique_bo_so_far = unique_bo_count; | |||
const unsigned default_bo_priority = 7; | |||
for (unsigned i = 0; i < radv_bo_list->count; ++i) { | |||
struct radv_amdgpu_winsys_bo *bo = radv_amdgpu_winsys_bo(radv_bo_list->bos[i]); | |||
bool found = false; | |||
for (unsigned j = 0; j < unique_bo_so_far; ++j) { | |||
if (bo->bo == handles[j]) { | |||
found = true; | |||
priorities[j] = MAX2(priorities[j], default_bo_priority); | |||
break; | |||
} | |||
} | |||
if (!found) { | |||
handles[unique_bo_count] = bo->bo; | |||
priorities[unique_bo_count] = default_bo_priority; | |||
++unique_bo_count; | |||
} | |||
} | |||
@@ -761,13 +731,12 @@ static int radv_amdgpu_create_bo_list(struct radv_amdgpu_winsys *ws, | |||
if (unique_bo_count > 0) { | |||
r = amdgpu_bo_list_create(ws->dev, unique_bo_count, handles, | |||
priorities, bo_list); | |||
NULL, bo_list); | |||
} else { | |||
*bo_list = 0; | |||
} | |||
free(handles); | |||
free(priorities); | |||
} | |||
return r; |