mesa/src/freedreno/vulkan/tu_perfetto.cc

426 lines
13 KiB
C++

/*
* Copyright © 2021 Google, Inc.
* SPDX-License-Identifier: MIT
*/
#include <perfetto.h>
#include "tu_perfetto.h"
#include "util/hash_table.h"
#include "util/perf/u_perfetto.h"
#include "util/perf/u_perfetto_renderpass.h"
#include "tu_tracepoints.h"
#include "tu_tracepoints_perfetto.h"
/* we can't include tu_knl.h and tu_device.h */
int
tu_device_get_gpu_timestamp(struct tu_device *dev,
uint64_t *ts);
int
tu_device_get_suspend_count(struct tu_device *dev,
uint64_t *suspend_count);
uint64_t
tu_device_ticks_to_ns(struct tu_device *dev, uint64_t ts);
struct u_trace_context *
tu_device_get_u_trace(struct tu_device *device);
/**
* Queue-id's
*/
enum {
DEFAULT_HW_QUEUE_ID,
};
/**
* Render-stage id's
*/
enum tu_stage_id {
CMD_BUFFER_STAGE_ID,
RENDER_PASS_STAGE_ID,
BINNING_STAGE_ID,
GMEM_STAGE_ID,
BYPASS_STAGE_ID,
BLIT_STAGE_ID,
COMPUTE_STAGE_ID,
CLEAR_SYSMEM_STAGE_ID,
CLEAR_GMEM_STAGE_ID,
GMEM_LOAD_STAGE_ID,
GMEM_STORE_STAGE_ID,
SYSMEM_RESOLVE_STAGE_ID,
// TODO add the rest from fd_stage_id
};
static const struct {
const char *name;
const char *desc;
} queues[] = {
[DEFAULT_HW_QUEUE_ID] = {"GPU Queue 0", "Default Adreno Hardware Queue"},
};
static const struct {
const char *name;
const char *desc;
} stages[] = {
[CMD_BUFFER_STAGE_ID] = { "Command Buffer" },
[RENDER_PASS_STAGE_ID] = { "Render Pass" },
[BINNING_STAGE_ID] = { "Binning", "Perform Visibility pass and determine target bins" },
[GMEM_STAGE_ID] = { "GMEM", "Rendering to GMEM" },
[BYPASS_STAGE_ID] = { "Bypass", "Rendering to system memory" },
[BLIT_STAGE_ID] = { "Blit", "Performing a Blit operation" },
[COMPUTE_STAGE_ID] = { "Compute", "Compute job" },
[CLEAR_SYSMEM_STAGE_ID] = { "Clear Sysmem", "" },
[CLEAR_GMEM_STAGE_ID] = { "Clear GMEM", "Per-tile (GMEM) clear" },
[GMEM_LOAD_STAGE_ID] = { "GMEM Load", "Per tile system memory to GMEM load" },
[GMEM_STORE_STAGE_ID] = { "GMEM Store", "Per tile GMEM to system memory store" },
[SYSMEM_RESOLVE_STAGE_ID] = { "SysMem Resolve", "System memory MSAA resolve" },
// TODO add the rest
};
static uint32_t gpu_clock_id;
static uint64_t next_clock_sync_ns; /* cpu time of next clk sync */
/**
* The timestamp at the point where we first emitted the clock_sync..
* this will be a *later* timestamp that the first GPU traces (since
* we capture the first clock_sync from the CPU *after* the first GPU
* tracepoints happen). To avoid confusing perfetto we need to drop
* the GPU traces with timestamps before this.
*/
static uint64_t sync_gpu_ts;
static uint64_t last_suspend_count;
static uint64_t gpu_max_timestamp;
static uint64_t gpu_timestamp_offset;
struct TuRenderpassIncrementalState {
bool was_cleared = true;
};
struct TuRenderpassTraits : public perfetto::DefaultDataSourceTraits {
using IncrementalStateType = TuRenderpassIncrementalState;
};
class TuRenderpassDataSource : public MesaRenderpassDataSource<TuRenderpassDataSource,
TuRenderpassTraits> {
void OnStart(const StartArgs &args) override
{
MesaRenderpassDataSource<TuRenderpassDataSource, TuRenderpassTraits>::OnStart(args);
/* Note: clock_id's below 128 are reserved.. for custom clock sources,
* using the hash of a namespaced string is the recommended approach.
* See: https://perfetto.dev/docs/concepts/clock-sync
*/
gpu_clock_id =
_mesa_hash_string("org.freedesktop.mesa.freedreno") | 0x80000000;
gpu_timestamp_offset = 0;
gpu_max_timestamp = 0;
last_suspend_count = 0;
}
};
PERFETTO_DECLARE_DATA_SOURCE_STATIC_MEMBERS(TuRenderpassDataSource);
PERFETTO_DEFINE_DATA_SOURCE_STATIC_MEMBERS(TuRenderpassDataSource);
static void
send_descriptors(TuRenderpassDataSource::TraceContext &ctx, uint64_t ts_ns)
{
PERFETTO_LOG("Sending renderstage descriptors");
auto packet = ctx.NewTracePacket();
packet->set_timestamp(0);
auto event = packet->set_gpu_render_stage_event();
event->set_gpu_id(0);
auto spec = event->set_specifications();
for (unsigned i = 0; i < ARRAY_SIZE(queues); i++) {
auto desc = spec->add_hw_queue();
desc->set_name(queues[i].name);
desc->set_description(queues[i].desc);
}
for (unsigned i = 0; i < ARRAY_SIZE(stages); i++) {
auto desc = spec->add_stage();
desc->set_name(stages[i].name);
if (stages[i].desc)
desc->set_description(stages[i].desc);
}
}
static struct tu_perfetto_stage *
stage_push(struct tu_device *dev)
{
struct tu_perfetto_state *p = tu_device_get_perfetto_state(dev);
if (p->stage_depth >= ARRAY_SIZE(p->stages)) {
p->skipped_depth++;
return NULL;
}
return &p->stages[p->stage_depth++];
}
typedef void (*trace_payload_as_extra_func)(perfetto::protos::pbzero::GpuRenderStageEvent *, const void*);
static struct tu_perfetto_stage *
stage_pop(struct tu_device *dev)
{
struct tu_perfetto_state *p = tu_device_get_perfetto_state(dev);
if (!p->stage_depth)
return NULL;
if (p->skipped_depth) {
p->skipped_depth--;
return NULL;
}
return &p->stages[--p->stage_depth];
}
static void
stage_start(struct tu_device *dev,
uint64_t ts_ns,
enum tu_stage_id stage_id,
const void *payload = nullptr,
size_t payload_size = 0,
trace_payload_as_extra_func payload_as_extra = nullptr)
{
struct tu_perfetto_stage *stage = stage_push(dev);
if (!stage) {
PERFETTO_ELOG("stage %d is nested too deep", stage_id);
return;
}
if (payload) {
void* new_payload = malloc(payload_size);
if (new_payload)
memcpy(new_payload, payload, payload_size);
else
PERFETTO_ELOG("Failed to allocate payload for stage %d", stage_id);
payload = new_payload;
}
*stage = (struct tu_perfetto_stage) {
.stage_id = stage_id,
.start_ts = ts_ns,
.payload = payload,
.start_payload_function = (void *) payload_as_extra,
};
}
static void
stage_end(struct tu_device *dev, uint64_t ts_ns, enum tu_stage_id stage_id,
uint32_t submission_id, const void* payload = nullptr,
trace_payload_as_extra_func payload_as_extra = nullptr)
{
struct tu_perfetto_stage *stage = stage_pop(dev);
if (!stage)
return;
if (stage->stage_id != stage_id) {
PERFETTO_ELOG("stage %d ended while stage %d is expected",
stage_id, stage->stage_id);
return;
}
/* If we haven't managed to calibrate the alignment between GPU and CPU
* timestamps yet, then skip this trace, otherwise perfetto won't know
* what to do with it.
*/
if (!sync_gpu_ts)
return;
TuRenderpassDataSource::Trace([=](TuRenderpassDataSource::TraceContext tctx) {
if (auto state = tctx.GetIncrementalState(); state->was_cleared) {
send_descriptors(tctx, stage->start_ts);
state->was_cleared = false;
}
auto packet = tctx.NewTracePacket();
gpu_max_timestamp = MAX2(gpu_max_timestamp, ts_ns + gpu_timestamp_offset);
packet->set_timestamp(stage->start_ts + gpu_timestamp_offset);
packet->set_timestamp_clock_id(gpu_clock_id);
auto event = packet->set_gpu_render_stage_event();
event->set_event_id(0); // ???
event->set_hw_queue_id(DEFAULT_HW_QUEUE_ID);
event->set_duration(ts_ns - stage->start_ts);
event->set_stage_id(stage->stage_id);
event->set_context((uintptr_t)dev);
event->set_submission_id(submission_id);
if (stage->payload) {
if (stage->start_payload_function)
((trace_payload_as_extra_func) stage->start_payload_function)(
event, stage->payload);
free((void *)stage->payload);
}
if (payload && payload_as_extra)
payload_as_extra(event, payload);
});
}
#ifdef __cplusplus
extern "C" {
#endif
void
tu_perfetto_init(void)
{
util_perfetto_init();
perfetto::DataSourceDescriptor dsd;
dsd.set_name("gpu.renderstages.msm");
TuRenderpassDataSource::Register(dsd);
}
static void
sync_timestamp(struct tu_device *dev)
{
uint64_t cpu_ts = perfetto::base::GetBootTimeNs().count();
uint64_t gpu_ts = 0;
if (cpu_ts < next_clock_sync_ns)
return;
if (tu_device_get_gpu_timestamp(dev, &gpu_ts)) {
PERFETTO_ELOG("Could not sync CPU and GPU clocks");
return;
}
/* get cpu timestamp again because tu_device_get_gpu_timestamp can take
* >100us
*/
cpu_ts = perfetto::base::GetBootTimeNs().count();
uint64_t current_suspend_count = 0;
/* If we fail to get it we will use a fallback */
tu_device_get_suspend_count(dev, &current_suspend_count);
/* convert GPU ts into ns: */
gpu_ts = tu_device_ticks_to_ns(dev, gpu_ts);
/* GPU timestamp is being reset after suspend-resume cycle.
* Perfetto requires clock snapshots to be monotonic,
* so we have to fix-up the time.
*/
if (current_suspend_count != last_suspend_count) {
gpu_timestamp_offset = gpu_max_timestamp;
last_suspend_count = current_suspend_count;
}
gpu_ts += gpu_timestamp_offset;
/* Fallback check, detect non-monotonic cases which would happen
* if we cannot retrieve suspend count.
*/
if (sync_gpu_ts > gpu_ts) {
gpu_ts += (gpu_max_timestamp - gpu_timestamp_offset);
gpu_timestamp_offset = gpu_max_timestamp;
}
if (sync_gpu_ts > gpu_ts) {
PERFETTO_ELOG("Non-monotonic gpu timestamp detected, bailing out");
return;
}
TuRenderpassDataSource::Trace([=](auto tctx) {
MesaRenderpassDataSource<TuRenderpassDataSource,
TuRenderpassTraits>::EmitClockSync(tctx, cpu_ts,
gpu_ts, gpu_clock_id);
});
gpu_max_timestamp = gpu_ts;
sync_gpu_ts = gpu_ts;
next_clock_sync_ns = cpu_ts + 30000000;
}
static void
emit_submit_id(uint32_t submission_id)
{
TuRenderpassDataSource::Trace([=](TuRenderpassDataSource::TraceContext tctx) {
auto packet = tctx.NewTracePacket();
packet->set_timestamp(perfetto::base::GetBootTimeNs().count());
auto event = packet->set_vulkan_api_event();
auto submit = event->set_vk_queue_submit();
submit->set_submission_id(submission_id);
});
}
void
tu_perfetto_submit(struct tu_device *dev, uint32_t submission_id)
{
/* sync_timestamp isn't free */
if (!u_trace_perfetto_active(tu_device_get_u_trace(dev)))
return;
sync_timestamp(dev);
emit_submit_id(submission_id);
}
/*
* Trace callbacks, called from u_trace once the timestamps from GPU have been
* collected.
*/
#define CREATE_EVENT_CALLBACK(event_name, stage_id) \
void tu_perfetto_start_##event_name( \
struct tu_device *dev, uint64_t ts_ns, const void *flush_data, \
const struct trace_start_##event_name *payload) \
{ \
stage_start( \
dev, ts_ns, stage_id, payload, \
sizeof(struct trace_start_##event_name), \
(trace_payload_as_extra_func) &trace_payload_as_extra_start_##event_name); \
} \
\
void tu_perfetto_end_##event_name( \
struct tu_device *dev, uint64_t ts_ns, const void *flush_data, \
const struct trace_end_##event_name *payload) \
{ \
auto trace_flush_data = \
(const struct tu_u_trace_submission_data *) flush_data; \
uint32_t submission_id = \
tu_u_trace_submission_data_get_submit_id(trace_flush_data); \
stage_end( \
dev, ts_ns, stage_id, submission_id, payload, \
(trace_payload_as_extra_func) &trace_payload_as_extra_end_##event_name); \
}
CREATE_EVENT_CALLBACK(cmd_buffer, CMD_BUFFER_STAGE_ID)
CREATE_EVENT_CALLBACK(render_pass, RENDER_PASS_STAGE_ID)
CREATE_EVENT_CALLBACK(binning_ib, BINNING_STAGE_ID)
CREATE_EVENT_CALLBACK(draw_ib_gmem, GMEM_STAGE_ID)
CREATE_EVENT_CALLBACK(draw_ib_sysmem, BYPASS_STAGE_ID)
CREATE_EVENT_CALLBACK(blit, BLIT_STAGE_ID)
CREATE_EVENT_CALLBACK(compute, COMPUTE_STAGE_ID)
CREATE_EVENT_CALLBACK(gmem_clear, CLEAR_GMEM_STAGE_ID)
CREATE_EVENT_CALLBACK(sysmem_clear, CLEAR_SYSMEM_STAGE_ID)
CREATE_EVENT_CALLBACK(sysmem_clear_all, CLEAR_SYSMEM_STAGE_ID)
CREATE_EVENT_CALLBACK(gmem_load, GMEM_LOAD_STAGE_ID)
CREATE_EVENT_CALLBACK(gmem_store, GMEM_STORE_STAGE_ID)
CREATE_EVENT_CALLBACK(sysmem_resolve, SYSMEM_RESOLVE_STAGE_ID)
#ifdef __cplusplus
}
#endif