blob: e24da87fb699012c8aaf9f7d7c64d1817620b6ee [file] [log] [blame]
/*
* Copyright 2021 Valve Corporation
*
* SPDX-License-Identifier: MIT
*/
#include "ac_spm.h"
#include "util/bitscan.h"
#include "util/u_memory.h"
#include "ac_perfcounter.h"
/* SPM counters definition. */
/* GFX10+ */
static struct ac_spm_counter_descr gfx10_num_l2_hits = {TCP, 0x9};
static struct ac_spm_counter_descr gfx10_num_l2_misses = {TCP, 0x12};
static struct ac_spm_counter_descr gfx10_num_scache_hits = {SQ, 0x14f};
static struct ac_spm_counter_descr gfx10_num_scache_misses = {SQ, 0x150};
static struct ac_spm_counter_descr gfx10_num_scache_misses_dup = {SQ, 0x151};
static struct ac_spm_counter_descr gfx10_num_icache_hits = {SQ, 0x12c};
static struct ac_spm_counter_descr gfx10_num_icache_misses = {SQ, 0x12d};
static struct ac_spm_counter_descr gfx10_num_icache_misses_dup = {SQ, 0x12e};
static struct ac_spm_counter_descr gfx10_num_gl1c_hits = {GL1C, 0xe};
static struct ac_spm_counter_descr gfx10_num_gl1c_misses = {GL1C, 0x12};
static struct ac_spm_counter_descr gfx10_num_gl2c_hits = {GL2C, 0x3};
static struct ac_spm_counter_descr gfx10_num_gl2c_misses = {GL2C, 0x23};
static struct ac_spm_counter_create_info gfx10_spm_counters[] = {
{&gfx10_num_l2_hits},
{&gfx10_num_l2_misses},
{&gfx10_num_scache_hits},
{&gfx10_num_scache_misses},
{&gfx10_num_scache_misses_dup},
{&gfx10_num_icache_hits},
{&gfx10_num_icache_misses},
{&gfx10_num_icache_misses_dup},
{&gfx10_num_gl1c_hits},
{&gfx10_num_gl1c_misses},
{&gfx10_num_gl2c_hits},
{&gfx10_num_gl2c_misses},
};
/* GFX10.3+ */
static struct ac_spm_counter_descr gfx103_num_gl2c_misses = {GL2C, 0x2b};
static struct ac_spm_counter_create_info gfx103_spm_counters[] = {
{&gfx10_num_l2_hits},
{&gfx10_num_l2_misses},
{&gfx10_num_scache_hits},
{&gfx10_num_scache_misses},
{&gfx10_num_scache_misses_dup},
{&gfx10_num_icache_hits},
{&gfx10_num_icache_misses},
{&gfx10_num_icache_misses_dup},
{&gfx10_num_gl1c_hits},
{&gfx10_num_gl1c_misses},
{&gfx10_num_gl2c_hits},
{&gfx103_num_gl2c_misses},
};
/* GFX11+ */
static struct ac_spm_counter_descr gfx11_num_l2_misses = {TCP, 0x11};
static struct ac_spm_counter_descr gfx11_num_scache_hits = {SQ_WGP, 0x126};
static struct ac_spm_counter_descr gfx11_num_scache_misses = {SQ_WGP, 0x127};
static struct ac_spm_counter_descr gfx11_num_scache_misses_dup = {SQ_WGP, 0x128};
static struct ac_spm_counter_descr gfx11_num_icache_hits = {SQ_WGP, 0x10e};
static struct ac_spm_counter_descr gfx11_num_icache_misses = {SQ_WGP, 0x10f};
static struct ac_spm_counter_descr gfx11_num_icache_misses_dup = {SQ_WGP, 0x110};
static struct ac_spm_counter_create_info gfx11_spm_counters[] = {
{&gfx10_num_l2_hits},
{&gfx11_num_l2_misses},
{&gfx11_num_scache_hits},
{&gfx11_num_scache_misses},
{&gfx11_num_scache_misses_dup},
{&gfx11_num_icache_hits},
{&gfx11_num_icache_misses},
{&gfx11_num_icache_misses_dup},
{&gfx10_num_gl1c_hits},
{&gfx10_num_gl1c_misses},
{&gfx10_num_gl2c_hits},
{&gfx103_num_gl2c_misses},
};
static struct ac_spm_block_select *
ac_spm_get_block_select(struct ac_spm *spm, const struct ac_pc_block *block)
{
struct ac_spm_block_select *block_sel, *new_block_sel;
uint32_t num_block_sel;
for (uint32_t i = 0; i < spm->num_block_sel; i++) {
if (spm->block_sel[i].b->b->b->gpu_block == block->b->b->gpu_block)
return &spm->block_sel[i];
}
/* Allocate a new select block if it doesn't already exist. */
num_block_sel = spm->num_block_sel + 1;
block_sel = realloc(spm->block_sel, num_block_sel * sizeof(*block_sel));
if (!block_sel)
return NULL;
spm->num_block_sel = num_block_sel;
spm->block_sel = block_sel;
/* Initialize the new select block. */
new_block_sel = &spm->block_sel[spm->num_block_sel - 1];
memset(new_block_sel, 0, sizeof(*new_block_sel));
new_block_sel->b = block;
new_block_sel->instances =
calloc(block->num_global_instances, sizeof(*new_block_sel->instances));
if (!new_block_sel->instances)
return NULL;
new_block_sel->num_instances = block->num_global_instances;
for (unsigned i = 0; i < new_block_sel->num_instances; i++)
new_block_sel->instances[i].num_counters = block->b->b->num_spm_counters;
return new_block_sel;
}
struct ac_spm_instance_mapping {
uint32_t se_index; /* SE index or 0 if global */
uint32_t sa_index; /* SA index or 0 if global or per-SE */
uint32_t instance_index;
};
static bool
ac_spm_init_instance_mapping(const struct radeon_info *info,
const struct ac_pc_block *block,
const struct ac_spm_counter_info *counter,
struct ac_spm_instance_mapping *mapping)
{
uint32_t instance_index = 0, se_index = 0, sa_index = 0;
if (block->b->b->flags & AC_PC_BLOCK_SE) {
if (block->b->b->gpu_block == SQ) {
/* Per-SE blocks. */
se_index = counter->instance / block->num_instances;
instance_index = counter->instance % block->num_instances;
} else {
/* Per-SA blocks. */
assert(block->b->b->gpu_block == GL1C ||
block->b->b->gpu_block == TCP ||
block->b->b->gpu_block == SQ_WGP);
se_index = (counter->instance / block->num_instances) / info->max_sa_per_se;
sa_index = (counter->instance / block->num_instances) % info->max_sa_per_se;
instance_index = counter->instance % block->num_instances;
}
} else {
/* Global blocks. */
assert(block->b->b->gpu_block == GL2C);
instance_index = counter->instance;
}
if (se_index >= info->num_se ||
sa_index >= info->max_sa_per_se ||
instance_index >= block->num_instances)
return false;
mapping->se_index = se_index;
mapping->sa_index = sa_index;
mapping->instance_index = instance_index;
return true;
}
static void
ac_spm_init_muxsel(const struct radeon_info *info,
const struct ac_pc_block *block,
const struct ac_spm_instance_mapping *mapping,
struct ac_spm_counter_info *counter,
uint32_t spm_wire)
{
const uint16_t counter_idx = 2 * spm_wire + (counter->is_even ? 0 : 1);
union ac_spm_muxsel *muxsel = &counter->muxsel;
if (info->gfx_level >= GFX11) {
muxsel->gfx11.counter = counter_idx;
muxsel->gfx11.block = block->b->b->spm_block_select;
muxsel->gfx11.shader_array = mapping->sa_index;
muxsel->gfx11.instance = mapping->instance_index;
} else {
muxsel->gfx10.counter = counter_idx;
muxsel->gfx10.block = block->b->b->spm_block_select;
muxsel->gfx10.shader_array = mapping->sa_index;
muxsel->gfx10.instance = mapping->instance_index;
}
}
static uint32_t
ac_spm_init_grbm_gfx_index(const struct ac_pc_block *block,
const struct ac_spm_instance_mapping *mapping)
{
uint32_t instance = mapping->instance_index;
uint32_t grbm_gfx_index = 0;
grbm_gfx_index |= S_030800_SE_INDEX(mapping->se_index) |
S_030800_SH_INDEX(mapping->sa_index);
switch (block->b->b->gpu_block) {
case GL2C:
/* Global blocks. */
grbm_gfx_index |= S_030800_SE_BROADCAST_WRITES(1);
break;
case SQ:
/* Per-SE blocks. */
grbm_gfx_index |= S_030800_SH_BROADCAST_WRITES(1);
break;
default:
/* Other blocks shouldn't broadcast. */
break;
}
if (block->b->b->gpu_block == SQ_WGP) {
union {
struct {
uint32_t block_index : 2; /* Block index withing WGP */
uint32_t wgp_index : 3;
uint32_t is_below_spi : 1; /* 0: lower WGP numbers, 1: higher WGP numbers */
uint32_t reserved : 26;
};
uint32_t value;
} instance_index = {0};
const uint32_t num_wgp_above_spi = 4;
const bool is_below_spi = mapping->instance_index >= num_wgp_above_spi;
instance_index.wgp_index =
is_below_spi ? (mapping->instance_index - num_wgp_above_spi) : mapping->instance_index;
instance_index.is_below_spi = is_below_spi;
instance = instance_index.value;
}
grbm_gfx_index |= S_030800_INSTANCE_INDEX(instance);
return grbm_gfx_index;
}
static bool
ac_spm_map_counter(struct ac_spm *spm, struct ac_spm_block_select *block_sel,
struct ac_spm_counter_info *counter,
const struct ac_spm_instance_mapping *mapping,
uint32_t *spm_wire)
{
uint32_t instance = counter->instance;
if (block_sel->b->b->b->gpu_block == SQ_WGP) {
if (!spm->sq_wgp[instance].grbm_gfx_index) {
spm->sq_wgp[instance].grbm_gfx_index =
ac_spm_init_grbm_gfx_index(block_sel->b, mapping);
}
for (unsigned i = 0; i < ARRAY_SIZE(spm->sq_wgp[instance].counters); i++) {
struct ac_spm_counter_select *cntr_sel = &spm->sq_wgp[instance].counters[i];
if (i < spm->sq_wgp[instance].num_counters)
continue;
cntr_sel->sel0 |= S_036700_PERF_SEL(counter->event_id) |
S_036700_SPM_MODE(1) | /* 16-bit clamp */
S_036700_PERF_MODE(0);
/* Each SQ_WQP modules (GFX11+) share one 32-bit accumulator/wire
* per pair of selects.
*/
cntr_sel->active |= 1 << (i % 2);
*spm_wire = i / 2;
if (cntr_sel->active & 0x1)
counter->is_even = true;
spm->sq_wgp[instance].num_counters++;
return true;
}
} else if (block_sel->b->b->b->gpu_block == SQ) {
for (unsigned i = 0; i < ARRAY_SIZE(spm->sqg[instance].counters); i++) {
struct ac_spm_counter_select *cntr_sel = &spm->sqg[instance].counters[i];
if (i < spm->sqg[instance].num_counters)
continue;
/* SQ doesn't support 16-bit counters. */
cntr_sel->sel0 |= S_036700_PERF_SEL(counter->event_id) |
S_036700_SPM_MODE(3) | /* 32-bit clamp */
S_036700_PERF_MODE(0);
cntr_sel->active |= 0x3;
/* 32-bits counter are always even. */
counter->is_even = true;
/* One wire per SQ module. */
*spm_wire = i;
spm->sqg[instance].num_counters++;
return true;
}
} else {
/* Generic blocks. */
struct ac_spm_block_instance *block_instance =
&block_sel->instances[instance];
if (!block_instance->grbm_gfx_index) {
block_instance->grbm_gfx_index =
ac_spm_init_grbm_gfx_index(block_sel->b, mapping);
}
for (unsigned i = 0; i < block_instance->num_counters; i++) {
struct ac_spm_counter_select *cntr_sel = &block_instance->counters[i];
int index = ffs(~cntr_sel->active) - 1;
switch (index) {
case 0: /* use S_037004_PERF_SEL */
cntr_sel->sel0 |= S_037004_PERF_SEL(counter->event_id) |
S_037004_CNTR_MODE(1) | /* 16-bit clamp */
S_037004_PERF_MODE(0); /* accum */
break;
case 1: /* use S_037004_PERF_SEL1 */
cntr_sel->sel0 |= S_037004_PERF_SEL1(counter->event_id) |
S_037004_PERF_MODE1(0);
break;
case 2: /* use S_037004_PERF_SEL2 */
cntr_sel->sel1 |= S_037008_PERF_SEL2(counter->event_id) |
S_037008_PERF_MODE2(0);
break;
case 3: /* use S_037004_PERF_SEL3 */
cntr_sel->sel1 |= S_037008_PERF_SEL3(counter->event_id) |
S_037008_PERF_MODE3(0);
break;
default:
return false;
}
/* Mark this 16-bit counter as used. */
cntr_sel->active |= 1 << index;
/* Determine if the counter is even or odd. */
counter->is_even = !(index % 2);
/* Determine the SPM wire (one wire holds two 16-bit counters). */
*spm_wire = !!(index >= 2);
return true;
}
}
return false;
}
static bool
ac_spm_add_counter(const struct radeon_info *info,
const struct ac_perfcounters *pc,
struct ac_spm *spm,
const struct ac_spm_counter_create_info *counter_info)
{
struct ac_spm_instance_mapping instance_mapping = {0};
struct ac_spm_counter_info *counter;
struct ac_spm_block_select *block_sel;
struct ac_pc_block *block;
uint32_t spm_wire;
/* Check if the GPU block is valid. */
block = ac_pc_get_block(pc, counter_info->b->gpu_block);
if (!block) {
fprintf(stderr, "ac/spm: Invalid GPU block.\n");
return false;
}
/* Check if the number of instances is valid. */
if (counter_info->instance > block->num_global_instances - 1) {
fprintf(stderr, "ac/spm: Invalid instance ID.\n");
return false;
}
/* Check if the event ID is valid. */
if (counter_info->b->event_id > block->b->selectors) {
fprintf(stderr, "ac/spm: Invalid event ID.\n");
return false;
}
counter = &spm->counters[spm->num_counters];
spm->num_counters++;
counter->gpu_block = counter_info->b->gpu_block;
counter->event_id = counter_info->b->event_id;
counter->instance = counter_info->instance;
/* Get the select block used to configure the counter. */
block_sel = ac_spm_get_block_select(spm, block);
if (!block_sel)
return false;
/* Initialize instance mapping for the counter. */
if (!ac_spm_init_instance_mapping(info, block, counter, &instance_mapping)) {
fprintf(stderr, "ac/spm: Failed to initialize instance mapping.\n");
return false;
}
/* Map the counter to the select block. */
if (!ac_spm_map_counter(spm, block_sel, counter, &instance_mapping, &spm_wire)) {
fprintf(stderr, "ac/spm: No free slots available!\n");
return false;
}
/* Determine the counter segment type. */
if (block->b->b->flags & AC_PC_BLOCK_SE) {
counter->segment_type = instance_mapping.se_index;
} else {
counter->segment_type = AC_SPM_SEGMENT_TYPE_GLOBAL;
}
/* Configure the muxsel for SPM. */
ac_spm_init_muxsel(info, block, &instance_mapping, counter, spm_wire);
return true;
}
static void
ac_spm_fill_muxsel_ram(const struct radeon_info *info,
struct ac_spm *spm,
enum ac_spm_segment_type segment_type,
uint32_t offset)
{
struct ac_spm_muxsel_line *mappings = spm->muxsel_lines[segment_type];
uint32_t even_counter_idx = 0, even_line_idx = 0;
uint32_t odd_counter_idx = 0, odd_line_idx = 1;
/* Add the global timestamps first. */
if (segment_type == AC_SPM_SEGMENT_TYPE_GLOBAL) {
if (info->gfx_level >= GFX11) {
mappings[even_line_idx].muxsel[even_counter_idx++].value = 0xf840;
mappings[even_line_idx].muxsel[even_counter_idx++].value = 0xf841;
mappings[even_line_idx].muxsel[even_counter_idx++].value = 0xf842;
mappings[even_line_idx].muxsel[even_counter_idx++].value = 0xf843;
} else {
for (unsigned i = 0; i < 4; i++) {
mappings[even_line_idx].muxsel[even_counter_idx++].value = 0xf0f0;
}
}
}
for (unsigned i = 0; i < spm->num_counters; i++) {
struct ac_spm_counter_info *counter = &spm->counters[i];
if (counter->segment_type != segment_type)
continue;
if (counter->is_even) {
counter->offset =
(offset + even_line_idx) * AC_SPM_NUM_COUNTER_PER_MUXSEL + even_counter_idx;
mappings[even_line_idx].muxsel[even_counter_idx] = spm->counters[i].muxsel;
if (++even_counter_idx == AC_SPM_NUM_COUNTER_PER_MUXSEL) {
even_counter_idx = 0;
even_line_idx += 2;
}
} else {
counter->offset =
(offset + odd_line_idx) * AC_SPM_NUM_COUNTER_PER_MUXSEL + odd_counter_idx;
mappings[odd_line_idx].muxsel[odd_counter_idx] = spm->counters[i].muxsel;
if (++odd_counter_idx == AC_SPM_NUM_COUNTER_PER_MUXSEL) {
odd_counter_idx = 0;
odd_line_idx += 2;
}
}
}
}
bool ac_init_spm(const struct radeon_info *info,
const struct ac_perfcounters *pc,
struct ac_spm *spm)
{
const struct ac_spm_counter_create_info *create_info;
unsigned create_info_count;
unsigned num_counters = 0;
switch (info->gfx_level) {
case GFX10:
create_info_count = ARRAY_SIZE(gfx10_spm_counters);
create_info = gfx10_spm_counters;
break;
case GFX10_3:
create_info_count = ARRAY_SIZE(gfx103_spm_counters);
create_info = gfx103_spm_counters;
break;
case GFX11:
case GFX11_5:
create_info_count = ARRAY_SIZE(gfx11_spm_counters);
create_info = gfx11_spm_counters;
break;
default:
fprintf(stderr, "radv: Failed to initialize SPM because SPM counters aren't implemented.\n");
return false; /* not implemented */
}
/* Count the total number of counters. */
for (unsigned i = 0; i < create_info_count; i++) {
const struct ac_pc_block *block = ac_pc_get_block(pc, create_info[i].b->gpu_block);
if (!block)
return false;
num_counters += block->num_global_instances;
}
spm->counters = CALLOC(num_counters, sizeof(*spm->counters));
if (!spm->counters)
return false;
for (unsigned i = 0; i < create_info_count; i++) {
const struct ac_pc_block *block = ac_pc_get_block(pc, create_info[i].b->gpu_block);
struct ac_spm_counter_create_info counter = create_info[i];
for (unsigned j = 0; j < block->num_global_instances; j++) {
counter.instance = j;
if (!ac_spm_add_counter(info, pc, spm, &counter)) {
fprintf(stderr, "ac/spm: Failed to add SPM counter (%d).\n", i);
return false;
}
}
}
/* Determine the segment size and create a muxsel ram for every segment. */
for (unsigned s = 0; s < AC_SPM_SEGMENT_TYPE_COUNT; s++) {
unsigned num_even_counters = 0, num_odd_counters = 0;
if (s == AC_SPM_SEGMENT_TYPE_GLOBAL) {
/* The global segment always start with a 64-bit timestamp. */
num_even_counters += AC_SPM_GLOBAL_TIMESTAMP_COUNTERS;
}
/* Count the number of even/odd counters for this segment. */
for (unsigned c = 0; c < spm->num_counters; c++) {
struct ac_spm_counter_info *counter = &spm->counters[c];
if (counter->segment_type != s)
continue;
if (counter->is_even) {
num_even_counters++;
} else {
num_odd_counters++;
}
}
/* Compute the number of lines. */
unsigned even_lines =
DIV_ROUND_UP(num_even_counters, AC_SPM_NUM_COUNTER_PER_MUXSEL);
unsigned odd_lines =
DIV_ROUND_UP(num_odd_counters, AC_SPM_NUM_COUNTER_PER_MUXSEL);
unsigned num_lines = (even_lines > odd_lines) ? (2 * even_lines - 1) : (2 * odd_lines);
spm->muxsel_lines[s] = CALLOC(num_lines, sizeof(*spm->muxsel_lines[s]));
if (!spm->muxsel_lines[s])
return false;
spm->num_muxsel_lines[s] = num_lines;
}
/* Compute the maximum number of muxsel lines among all SEs. On GFX11,
* there is only one SE segment size value and the highest value is used.
*/
for (unsigned s = 0; s < AC_SPM_SEGMENT_TYPE_GLOBAL; s++) {
spm->max_se_muxsel_lines =
MAX2(spm->num_muxsel_lines[s], spm->max_se_muxsel_lines);
}
/* RLC uses the following order: Global, SE0, SE1, SE2, SE3, SE4, SE5. */
ac_spm_fill_muxsel_ram(info, spm, AC_SPM_SEGMENT_TYPE_GLOBAL, 0);
const uint32_t num_global_lines = spm->num_muxsel_lines[AC_SPM_SEGMENT_TYPE_GLOBAL];
if (info->gfx_level >= GFX11) {
/* On GFX11, RLC uses one segment size for every single SE. */
for (unsigned i = 0; i < info->num_se; i++) {
assert(i < AC_SPM_SEGMENT_TYPE_GLOBAL);
uint32_t offset = num_global_lines + i * spm->max_se_muxsel_lines;
ac_spm_fill_muxsel_ram(info, spm, i, offset);
}
} else {
uint32_t offset = num_global_lines;
for (unsigned i = 0; i < info->num_se; i++) {
assert(i < AC_SPM_SEGMENT_TYPE_GLOBAL);
ac_spm_fill_muxsel_ram(info, spm, i, offset);
offset += spm->num_muxsel_lines[i];
}
}
/* On GFX11, the data size written by the hw is in units of segment. */
spm->ptr_granularity = info->gfx_level >= GFX11 ? 32 : 1;
return true;
}
void ac_destroy_spm(struct ac_spm *spm)
{
for (unsigned s = 0; s < AC_SPM_SEGMENT_TYPE_COUNT; s++) {
FREE(spm->muxsel_lines[s]);
}
for (unsigned i = 0; i < spm->num_block_sel; i++) {
FREE(spm->block_sel[i].instances);
}
FREE(spm->block_sel);
FREE(spm->counters);
}
static uint32_t ac_spm_get_sample_size(const struct ac_spm *spm)
{
uint32_t sample_size = 0; /* in bytes */
for (unsigned s = 0; s < AC_SPM_SEGMENT_TYPE_COUNT; s++) {
sample_size += spm->num_muxsel_lines[s] * AC_SPM_MUXSEL_LINE_SIZE * 4;
}
return sample_size;
}
static bool ac_spm_get_num_samples(const struct ac_spm *spm, uint32_t *num_samples)
{
uint32_t sample_size = ac_spm_get_sample_size(spm);
uint32_t *ptr = (uint32_t *)spm->ptr;
uint32_t data_size, num_lines_written;
/* Get the data size (in bytes) written by the hw to the ring buffer. */
data_size = ptr[0] * spm->ptr_granularity;
/* Compute the number of 256 bits (16 * 16-bits counters) lines written. */
num_lines_written = data_size / (2 * AC_SPM_NUM_COUNTER_PER_MUXSEL);
/* Check for overflow. */
if (num_lines_written % (sample_size / 32)) {
/* Buffer is too small and it needs to be resized. */
return false;
}
*num_samples = num_lines_written / (sample_size / 32);
return true;
}
bool ac_spm_get_trace(const struct ac_spm *spm, struct ac_spm_trace *trace)
{
memset(trace, 0, sizeof(*trace));
trace->ptr = spm->ptr;
trace->sample_interval = spm->sample_interval;
trace->num_counters = spm->num_counters;
trace->counters = spm->counters;
trace->sample_size_in_bytes = ac_spm_get_sample_size(spm);
return ac_spm_get_num_samples(spm, &trace->num_samples);
}