blob: db6d099b828312dcc42e45ca9df480d735cad8b4 [file] [log] [blame]
/*
* Copyright (C) 2018 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "src/trace_processor/trace_processor_impl.h"
#include <algorithm>
#include <memory>
#include "perfetto/base/logging.h"
#include "perfetto/base/status.h"
#include "perfetto/base/time.h"
#include "perfetto/ext/base/string_splitter.h"
#include "perfetto/ext/base/string_utils.h"
#include "perfetto/ext/trace_processor/demangle.h"
#include "src/trace_processor/dynamic/ancestor_generator.h"
#include "src/trace_processor/dynamic/connected_flow_generator.h"
#include "src/trace_processor/dynamic/descendant_generator.h"
#include "src/trace_processor/dynamic/describe_slice_generator.h"
#include "src/trace_processor/dynamic/experimental_annotated_stack_generator.h"
#include "src/trace_processor/dynamic/experimental_counter_dur_generator.h"
#include "src/trace_processor/dynamic/experimental_flamegraph_generator.h"
#include "src/trace_processor/dynamic/experimental_flat_slice_generator.h"
#include "src/trace_processor/dynamic/experimental_sched_upid_generator.h"
#include "src/trace_processor/dynamic/experimental_slice_layout_generator.h"
#include "src/trace_processor/dynamic/thread_state_generator.h"
#include "src/trace_processor/export_json.h"
#include "src/trace_processor/importers/additional_modules.h"
#include "src/trace_processor/importers/ftrace/sched_event_tracker.h"
#include "src/trace_processor/importers/fuchsia/fuchsia_trace_parser.h"
#include "src/trace_processor/importers/fuchsia/fuchsia_trace_tokenizer.h"
#include "src/trace_processor/importers/gzip/gzip_trace_parser.h"
#include "src/trace_processor/importers/json/json_trace_parser.h"
#include "src/trace_processor/importers/json/json_trace_tokenizer.h"
#include "src/trace_processor/importers/proto/metadata_tracker.h"
#include "src/trace_processor/importers/systrace/systrace_trace_parser.h"
#include "src/trace_processor/iterator_impl.h"
#include "src/trace_processor/sqlite/create_function.h"
#include "src/trace_processor/sqlite/create_view_function.h"
#include "src/trace_processor/sqlite/register_function.h"
#include "src/trace_processor/sqlite/scoped_db.h"
#include "src/trace_processor/sqlite/span_join_operator_table.h"
#include "src/trace_processor/sqlite/sql_stats_table.h"
#include "src/trace_processor/sqlite/sqlite3_str_split.h"
#include "src/trace_processor/sqlite/sqlite_raw_table.h"
#include "src/trace_processor/sqlite/sqlite_table.h"
#include "src/trace_processor/sqlite/sqlite_utils.h"
#include "src/trace_processor/sqlite/stats_table.h"
#include "src/trace_processor/sqlite/window_operator_table.h"
#include "src/trace_processor/tp_metatrace.h"
#include "src/trace_processor/types/variadic.h"
#include "src/trace_processor/util/protozero_to_text.h"
#include "src/trace_processor/util/status_macros.h"
#include "protos/perfetto/trace/perfetto/perfetto_metatrace.pbzero.h"
#include "protos/perfetto/trace/trace.pbzero.h"
#include "protos/perfetto/trace/trace_packet.pbzero.h"
#include "src/trace_processor/metrics/all_chrome_metrics.descriptor.h"
#include "src/trace_processor/metrics/metrics.descriptor.h"
#include "src/trace_processor/metrics/metrics.h"
#include "src/trace_processor/metrics/sql/amalgamated_sql_metrics.h"
// In Android and Chromium tree builds, we don't have the percentile module.
// Just don't include it.
#if PERFETTO_BUILDFLAG(PERFETTO_TP_PERCENTILE)
// defined in sqlite_src/ext/misc/percentile.c
extern "C" int sqlite3_percentile_init(sqlite3* db,
char** error,
const sqlite3_api_routines* api);
#endif // PERFETTO_BUILDFLAG(PERFETTO_TP_PERCENTILE)
namespace perfetto {
namespace trace_processor {
namespace {
const char kAllTablesQuery[] =
"SELECT tbl_name, type FROM (SELECT * FROM sqlite_master UNION ALL SELECT "
"* FROM sqlite_temp_master)";
template <typename SqlFunction, typename Ptr = typename SqlFunction::Context*>
void RegisterFunction(sqlite3* db,
const char* name,
int argc,
Ptr context = nullptr,
bool deterministic = true) {
auto status = RegisterSqlFunction<SqlFunction>(
db, name, argc, std::move(context), deterministic);
if (!status.ok())
PERFETTO_ELOG("%s", status.c_message());
}
void InitializeSqlite(sqlite3* db) {
char* error = nullptr;
sqlite3_exec(db, "PRAGMA temp_store=2", 0, 0, &error);
if (error) {
PERFETTO_FATAL("Error setting pragma temp_store: %s", error);
}
sqlite3_str_split_init(db);
// In Android tree builds, we don't have the percentile module.
// Just don't include it.
#if PERFETTO_BUILDFLAG(PERFETTO_TP_PERCENTILE)
sqlite3_percentile_init(db, &error, nullptr);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
#endif
}
void BuildBoundsTable(sqlite3* db, std::pair<int64_t, int64_t> bounds) {
char* error = nullptr;
sqlite3_exec(db, "DELETE FROM trace_bounds", nullptr, nullptr, &error);
if (error) {
PERFETTO_ELOG("Error deleting from bounds table: %s", error);
sqlite3_free(error);
return;
}
char* insert_sql = sqlite3_mprintf("INSERT INTO trace_bounds VALUES(%" PRId64
", %" PRId64 ")",
bounds.first, bounds.second);
sqlite3_exec(db, insert_sql, 0, 0, &error);
sqlite3_free(insert_sql);
if (error) {
PERFETTO_ELOG("Error inserting bounds table: %s", error);
sqlite3_free(error);
}
}
void CreateBuiltinTables(sqlite3* db) {
char* error = nullptr;
sqlite3_exec(db, "CREATE TABLE perfetto_tables(name STRING)", 0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
sqlite3_exec(db,
"CREATE TABLE trace_bounds(start_ts BIG INT, end_ts BIG INT)", 0,
0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
// Ensure that the entries in power_profile are unique to prevent duplicates
// when the power_profile is augmented with additional profiles.
sqlite3_exec(db,
"CREATE TABLE power_profile("
"device STRING, cpu INT, cluster INT, freq INT, power DOUBLE,"
"UNIQUE(device, cpu, cluster, freq));",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
sqlite3_exec(db, "CREATE TABLE trace_metrics(name STRING)", 0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
// This is a table intended to be used for metric debugging/developing. Data
// in the table is shown specially in the UI, and users can insert rows into
// this table to draw more things.
sqlite3_exec(db,
"CREATE TABLE debug_slices (id BIG INT, name STRING, ts BIG INT,"
"dur BIG INT, depth BIG INT)",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
// Initialize the bounds table with some data so even before parsing any data,
// we still have a valid table.
BuildBoundsTable(db, std::make_pair(0, 0));
}
void CreateBuiltinViews(sqlite3* db) {
char* error = nullptr;
sqlite3_exec(db,
"CREATE VIEW counter_definitions AS "
"SELECT "
" *, "
" id AS counter_id "
"FROM counter_track",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
sqlite3_exec(db,
"CREATE VIEW counter_values AS "
"SELECT "
" *, "
" track_id as counter_id "
"FROM counter",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
sqlite3_exec(db,
"CREATE VIEW counters AS "
"SELECT * "
"FROM counter_values v "
"INNER JOIN counter_track t "
"ON v.track_id = t.id "
"ORDER BY ts;",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
sqlite3_exec(db,
"CREATE VIEW slice AS "
"SELECT "
" *, "
" category AS cat, "
" id AS slice_id "
"FROM internal_slice;",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
sqlite3_exec(db,
"CREATE VIEW instants AS "
"SELECT "
"*, "
"0.0 as value "
"FROM instant;",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
sqlite3_exec(db,
"CREATE VIEW sched AS "
"SELECT "
"*, "
"ts + dur as ts_end "
"FROM sched_slice;",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
// Legacy view for "slice" table with a deprecated table name.
// TODO(eseckler): Remove this view when all users have switched to "slice".
sqlite3_exec(db,
"CREATE VIEW slices AS "
"SELECT * FROM slice;",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
sqlite3_exec(db,
"CREATE VIEW thread AS "
"SELECT "
"id as utid, "
"* "
"FROM internal_thread;",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
sqlite3_exec(db,
"CREATE VIEW process AS "
"SELECT "
"id as upid, "
"* "
"FROM internal_process;",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
// This should be kept in sync with GlobalArgsTracker::AddArgSet.
sqlite3_exec(db,
"CREATE VIEW args AS "
"SELECT "
"*, "
"CASE value_type "
" WHEN 'int' THEN CAST(int_value AS text) "
" WHEN 'uint' THEN CAST(int_value AS text) "
" WHEN 'string' THEN string_value "
" WHEN 'real' THEN CAST(real_value AS text) "
" WHEN 'pointer' THEN printf('0x%x', int_value) "
" WHEN 'bool' THEN ( "
" CASE WHEN int_value <> 0 THEN 'true' "
" ELSE 'false' END) "
" WHEN 'json' THEN string_value "
"ELSE NULL END AS display_value "
"FROM internal_args;",
0, 0, &error);
if (error) {
PERFETTO_ELOG("Error initializing: %s", error);
sqlite3_free(error);
}
}
struct ExportJson : public SqlFunction {
using Context = TraceStorage;
static base::Status Run(TraceStorage* storage,
size_t /*argc*/,
sqlite3_value** argv,
SqlValue& /*out*/,
Destructors&);
};
base::Status ExportJson::Run(TraceStorage* storage,
size_t /*argc*/,
sqlite3_value** argv,
SqlValue& /*out*/,
Destructors&) {
FILE* output;
if (sqlite3_value_type(argv[0]) == SQLITE_INTEGER) {
// Assume input is an FD.
output = fdopen(sqlite3_value_int(argv[0]), "w");
if (!output) {
return base::ErrStatus(
"EXPORT_JSON: Couldn't open output file from given FD");
}
} else {
const char* filename =
reinterpret_cast<const char*>(sqlite3_value_text(argv[0]));
output = fopen(filename, "w");
if (!output) {
return base::ErrStatus("EXPORT_JSON: Couldn't open output file");
}
}
return json::ExportJson(storage, output);
}
struct Hash : public SqlFunction {
static base::Status Run(void*,
size_t argc,
sqlite3_value** argv,
SqlValue& out,
Destructors&);
};
base::Status Hash::Run(void*,
size_t argc,
sqlite3_value** argv,
SqlValue& out,
Destructors&) {
base::Hash hash;
for (size_t i = 0; i < argc; ++i) {
sqlite3_value* value = argv[i];
int type = sqlite3_value_type(value);
switch (type) {
case SQLITE_INTEGER:
hash.Update(sqlite3_value_int64(value));
break;
case SQLITE_TEXT: {
const char* ptr =
reinterpret_cast<const char*>(sqlite3_value_text(value));
hash.Update(ptr, strlen(ptr));
break;
}
default:
return base::ErrStatus("HASH: arg %zu has unknown type %d", i, type);
}
}
out = SqlValue::Long(static_cast<int64_t>(hash.digest()));
return base::OkStatus();
}
struct Demangle : public SqlFunction {
static base::Status Run(void*,
size_t argc,
sqlite3_value** argv,
SqlValue& out,
Destructors& destructors);
};
base::Status Demangle::Run(void*,
size_t argc,
sqlite3_value** argv,
SqlValue& out,
Destructors& destructors) {
if (argc != 1)
return base::ErrStatus("Unsupported number of arg passed to DEMANGLE");
sqlite3_value* value = argv[0];
if (sqlite3_value_type(value) == SQLITE_NULL)
return base::OkStatus();
if (sqlite3_value_type(value) != SQLITE_TEXT)
return base::ErrStatus("Unsupported type of arg passed to DEMANGLE");
const char* mangled =
reinterpret_cast<const char*>(sqlite3_value_text(value));
std::unique_ptr<char, base::FreeDeleter> demangled =
demangle::Demangle(mangled);
if (!demangled)
return base::OkStatus();
destructors.string_destructor = free;
out = SqlValue::String(demangled.release());
return base::OkStatus();
}
void LastNonNullStep(sqlite3_context* ctx, int argc, sqlite3_value** argv) {
if (argc != 1) {
sqlite3_result_error(
ctx, "Unsupported number of args passed to LAST_NON_NULL", -1);
return;
}
sqlite3_value* value = argv[0];
if (sqlite3_value_type(value) == SQLITE_NULL) {
return;
}
sqlite3_value** ptr = reinterpret_cast<sqlite3_value**>(
sqlite3_aggregate_context(ctx, sizeof(sqlite3_value*)));
if (ptr) {
if (*ptr != nullptr) {
sqlite3_value_free(*ptr);
}
*ptr = sqlite3_value_dup(value);
}
}
void LastNonNullInverse(sqlite3_context* ctx, int argc, sqlite3_value** argv) {
// Do nothing.
base::ignore_result(ctx);
base::ignore_result(argc);
base::ignore_result(argv);
}
void LastNonNullValue(sqlite3_context* ctx) {
sqlite3_value** ptr =
reinterpret_cast<sqlite3_value**>(sqlite3_aggregate_context(ctx, 0));
if (!ptr || !*ptr) {
sqlite3_result_null(ctx);
} else {
sqlite3_result_value(ctx, *ptr);
}
}
void LastNonNullFinal(sqlite3_context* ctx) {
sqlite3_value** ptr =
reinterpret_cast<sqlite3_value**>(sqlite3_aggregate_context(ctx, 0));
if (!ptr || !*ptr) {
sqlite3_result_null(ctx);
} else {
sqlite3_result_value(ctx, *ptr);
sqlite3_value_free(*ptr);
}
}
void RegisterLastNonNullFunction(sqlite3* db) {
auto ret = sqlite3_create_window_function(
db, "LAST_NON_NULL", 1, SQLITE_UTF8 | SQLITE_DETERMINISTIC, nullptr,
&LastNonNullStep, &LastNonNullFinal, &LastNonNullValue,
&LastNonNullInverse, nullptr);
if (ret) {
PERFETTO_ELOG("Error initializing LAST_NON_NULL");
}
}
struct ValueAtMaxTsContext {
bool initialized;
int value_type;
int64_t max_ts;
int64_t int_value_at_max_ts;
double double_value_at_max_ts;
};
void ValueAtMaxTsStep(sqlite3_context* ctx, int, sqlite3_value** argv) {
sqlite3_value* ts = argv[0];
sqlite3_value* value = argv[1];
// Note that sqlite3_aggregate_context zeros the memory for us so all the
// variables of the struct should be zero.
ValueAtMaxTsContext* fn_ctx = reinterpret_cast<ValueAtMaxTsContext*>(
sqlite3_aggregate_context(ctx, sizeof(ValueAtMaxTsContext)));
// For performance reasons, we only do the check for the type of ts and value
// on the first call of the function.
if (PERFETTO_UNLIKELY(!fn_ctx->initialized)) {
if (sqlite3_value_type(ts) != SQLITE_INTEGER) {
sqlite3_result_error(ctx, "VALUE_AT_MAX_TS: ts passed was not an integer",
-1);
return;
}
fn_ctx->value_type = sqlite3_value_type(value);
if (fn_ctx->value_type != SQLITE_INTEGER &&
fn_ctx->value_type != SQLITE_FLOAT) {
sqlite3_result_error(
ctx, "VALUE_AT_MAX_TS: value passed was not an integer or float", -1);
return;
}
fn_ctx->max_ts = std::numeric_limits<int64_t>::min();
fn_ctx->initialized = true;
}
// On dcheck builds however, we check every passed ts and value.
#if PERFETTO_DCHECK_IS_ON()
if (sqlite3_value_type(ts) != SQLITE_INTEGER) {
sqlite3_result_error(ctx, "VALUE_AT_MAX_TS: ts passed was not an integer",
-1);
return;
}
if (sqlite3_value_type(value) != fn_ctx->value_type) {
sqlite3_result_error(ctx, "VALUE_AT_MAX_TS: value type is inconsistent",
-1);
return;
}
#endif
int64_t ts_int = sqlite3_value_int64(ts);
if (PERFETTO_LIKELY(fn_ctx->max_ts <= ts_int)) {
fn_ctx->max_ts = ts_int;
if (fn_ctx->value_type == SQLITE_INTEGER) {
fn_ctx->int_value_at_max_ts = sqlite3_value_int64(value);
} else {
fn_ctx->double_value_at_max_ts = sqlite3_value_double(value);
}
}
}
void ValueAtMaxTsFinal(sqlite3_context* ctx) {
ValueAtMaxTsContext* fn_ctx =
reinterpret_cast<ValueAtMaxTsContext*>(sqlite3_aggregate_context(ctx, 0));
if (!fn_ctx) {
sqlite3_result_null(ctx);
return;
}
if (fn_ctx->value_type == SQLITE_INTEGER) {
sqlite3_result_int64(ctx, fn_ctx->int_value_at_max_ts);
} else {
sqlite3_result_double(ctx, fn_ctx->double_value_at_max_ts);
}
}
void RegisterValueAtMaxTsFunction(sqlite3* db) {
auto ret = sqlite3_create_function_v2(
db, "VALUE_AT_MAX_TS", 2, SQLITE_UTF8 | SQLITE_DETERMINISTIC, nullptr,
nullptr, &ValueAtMaxTsStep, &ValueAtMaxTsFinal, nullptr);
if (ret) {
PERFETTO_ELOG("Error initializing VALUE_AT_MAX_TS");
}
}
struct ExtractArg : public SqlFunction {
using Context = TraceStorage;
static base::Status Run(TraceStorage* storage,
size_t argc,
sqlite3_value** argv,
SqlValue& out,
Destructors& destructors);
};
base::Status ExtractArg::Run(TraceStorage* storage,
size_t argc,
sqlite3_value** argv,
SqlValue& out,
Destructors& destructors) {
if (argc != 2)
return base::ErrStatus("EXTRACT_ARG: 2 args required");
// If the arg set id is null, just return null as the result.
if (sqlite3_value_type(argv[0]) == SQLITE_NULL)
return base::OkStatus();
if (sqlite3_value_type(argv[0]) != SQLITE_INTEGER)
return base::ErrStatus("EXTRACT_ARG: 1st argument should be arg set id");
if (sqlite3_value_type(argv[1]) != SQLITE_TEXT)
return base::ErrStatus("EXTRACT_ARG: 2nd argument should be key");
uint32_t arg_set_id = static_cast<uint32_t>(sqlite3_value_int(argv[0]));
const char* key = reinterpret_cast<const char*>(sqlite3_value_text(argv[1]));
base::Optional<Variadic> opt_value;
RETURN_IF_ERROR(storage->ExtractArg(arg_set_id, key, &opt_value));
if (!opt_value)
return base::OkStatus();
// This function always returns static strings (i.e. scoped to lifetime
// of the TraceStorage thread pool) so prevent SQLite from making copies.
destructors.string_destructor = sqlite_utils::kSqliteStatic;
switch (opt_value->type) {
case Variadic::kNull:
return base::OkStatus();
case Variadic::kInt:
out = SqlValue::Long(opt_value->int_value);
return base::OkStatus();
case Variadic::kUint:
out = SqlValue::Long(static_cast<int64_t>(opt_value->uint_value));
return base::OkStatus();
case Variadic::kString:
out =
SqlValue::String(storage->GetString(opt_value->string_value).data());
return base::OkStatus();
case Variadic::kReal:
out = SqlValue::Double(opt_value->real_value);
return base::OkStatus();
case Variadic::kBool:
out = SqlValue::Long(opt_value->bool_value);
return base::OkStatus();
case Variadic::kPointer:
out = SqlValue::Long(static_cast<int64_t>(opt_value->pointer_value));
return base::OkStatus();
case Variadic::kJson:
out = SqlValue::String(storage->GetString(opt_value->json_value).data());
return base::OkStatus();
}
PERFETTO_FATAL("For GCC");
}
std::vector<std::string> SanitizeMetricMountPaths(
const std::vector<std::string>& mount_paths) {
std::vector<std::string> sanitized;
for (const auto& path : mount_paths) {
if (path.length() == 0)
continue;
sanitized.push_back(path);
if (path.back() != '/')
sanitized.back().append("/");
}
return sanitized;
}
struct SourceGeq : public SqlFunction {
static base::Status Run(void*,
size_t,
sqlite3_value**,
SqlValue&,
Destructors&) {
return base::ErrStatus(
"SOURCE_GEQ should not be called from the global scope");
}
};
struct Glob : public SqlFunction {
static base::Status Run(void*,
size_t,
sqlite3_value** argv,
SqlValue& out,
Destructors&) {
const char* pattern =
reinterpret_cast<const char*>(sqlite3_value_text(argv[0]));
const char* text =
reinterpret_cast<const char*>(sqlite3_value_text(argv[1]));
if (pattern && text) {
out = SqlValue::Long(sqlite3_strglob(pattern, text) == 0);
}
return base::OkStatus();
}
};
void SetupMetrics(TraceProcessor* tp,
sqlite3* db,
std::vector<metrics::SqlMetricFile>* sql_metrics,
const std::vector<std::string>& extension_paths) {
const std::vector<std::string> sanitized_extension_paths =
SanitizeMetricMountPaths(extension_paths);
std::vector<std::string> skip_prefixes;
skip_prefixes.reserve(sanitized_extension_paths.size());
for (const auto& path : sanitized_extension_paths) {
skip_prefixes.push_back(kMetricProtoRoot + path);
}
tp->ExtendMetricsProto(kMetricsDescriptor.data(), kMetricsDescriptor.size(),
skip_prefixes);
tp->ExtendMetricsProto(kAllChromeMetricsDescriptor.data(),
kAllChromeMetricsDescriptor.size(), skip_prefixes);
// TODO(lalitm): remove this special casing and change
// SanitizeMetricMountPaths if/when we move all protos for builtin metrics to
// match extension protos.
bool skip_all_sql = std::find(extension_paths.begin(), extension_paths.end(),
"") != extension_paths.end();
if (!skip_all_sql) {
for (const auto& file_to_sql : metrics::sql_metrics::kFileToSql) {
if (base::StartsWithAny(file_to_sql.path, sanitized_extension_paths))
continue;
tp->RegisterMetric(file_to_sql.path, file_to_sql.sql);
}
}
RegisterFunction<metrics::NullIfEmpty>(db, "NULL_IF_EMPTY", 1);
RegisterFunction<metrics::UnwrapMetricProto>(db, "UNWRAP_METRIC_PROTO", 2);
RegisterFunction<metrics::RunMetric>(
db, "RUN_METRIC", -1,
std::unique_ptr<metrics::RunMetric::Context>(
new metrics::RunMetric::Context{tp, sql_metrics}));
// TODO(lalitm): migrate this over to using RegisterFunction once aggregate
// functions are supported.
{
auto ret = sqlite3_create_function_v2(
db, "RepeatedField", 1, SQLITE_UTF8, nullptr, nullptr,
metrics::RepeatedFieldStep, metrics::RepeatedFieldFinal, nullptr);
if (ret)
PERFETTO_FATAL("Error initializing RepeatedField");
}
}
void EnsureSqliteInitialized() {
// sqlite3_initialize isn't actually thread-safe despite being documented
// as such; we need to make sure multiple TraceProcessorImpl instances don't
// call it concurrently and only gets called once per process, instead.
static bool init_once = [] { return sqlite3_initialize() == SQLITE_OK; }();
PERFETTO_CHECK(init_once);
}
void InsertIntoTraceMetricsTable(sqlite3* db, const std::string& metric_name) {
char* insert_sql = sqlite3_mprintf(
"INSERT INTO trace_metrics(name) VALUES('%q')", metric_name.c_str());
char* insert_error = nullptr;
sqlite3_exec(db, insert_sql, nullptr, nullptr, &insert_error);
sqlite3_free(insert_sql);
if (insert_error) {
PERFETTO_ELOG("Error registering table: %s", insert_error);
sqlite3_free(insert_error);
}
}
void IncrementCountForStmt(sqlite3_stmt* stmt,
IteratorImpl::StmtMetadata* metadata) {
metadata->statement_count++;
// If the stmt is already done, it clearly didn't have any output.
if (sqlite_utils::IsStmtDone(stmt))
return;
// If the statement only has a single column and that column is named
// "suppress_query_output", treat it as a statement without output for
// accounting purposes. This is done so that embedders (e.g. shell) can
// strictly check that only the last query produces output while also
// providing an escape hatch for SELECT RUN_METRIC() invocations (which
// sadly produce output).
if (sqlite3_column_count(stmt) == 1 &&
strcmp(sqlite3_column_name(stmt, 0), "suppress_query_output") == 0) {
return;
}
// Otherwise, the statement has output and so increment the count.
metadata->statement_count_with_output++;
}
base::Status PrepareAndStepUntilLastValidStmt(
sqlite3* db,
const std::string& sql,
ScopedStmt* output_stmt,
IteratorImpl::StmtMetadata* metadata) {
ScopedStmt prev_stmt;
// A sql string can contain several statements. Some of them might be comment
// only, e.g. "SELECT 1; /* comment */; SELECT 2;". Here we process one
// statement on each iteration. SQLite's sqlite_prepare_v2 (wrapped by
// PrepareStmt) returns on each iteration a pointer to the unprocessed string.
//
// Unfortunately we cannot call PrepareStmt and tokenize all statements
// upfront because sqlite_prepare_v2 also semantically checks the statement
// against the schema. In some cases statements might depend on the execution
// of previous ones (e.e. CREATE VIEW x; SELECT FROM x; DELETE VIEW x;).
//
// Also, unfortunately, we need to PrepareStmt to find out if a statement is a
// comment or a real statement.
//
// The logic here is the following:
// - We invoke PrepareStmt on each statement.
// - If the statement is a comment we simply skip it.
// - If the statement is valid, we step once to make sure side effects take
// effect.
// - If we encounter a valid statement afterwards, we step internally through
// all rows of the previous one. This ensures that any further side effects
// take hold *before* we step into the next statement.
// - Once no further non-comment statements are encountered, we return an
// iterator to the last valid statement.
for (const char* rem_sql = sql.c_str(); rem_sql && rem_sql[0];) {
ScopedStmt cur_stmt;
{
PERFETTO_TP_TRACE("QUERY_PREPARE");
const char* tail = nullptr;
RETURN_IF_ERROR(sqlite_utils::PrepareStmt(db, rem_sql, &cur_stmt, &tail));
rem_sql = tail;
}
// The only situation where we'd have an ok status but also no prepared
// statement is if the statement of SQL we parsed was a pure comment. In
// this case, just continue to the next statement.
if (!cur_stmt)
continue;
// Before stepping into |cur_stmt|, we need to finish iterating through
// the previous statement so we don't have two clashing statements (e.g.
// SELECT * FROM v and DROP VIEW v) partially stepped into.
if (prev_stmt)
RETURN_IF_ERROR(sqlite_utils::StepStmtUntilDone(prev_stmt.get()));
PERFETTO_DLOG("Executing statement: %s", sqlite3_sql(*cur_stmt));
// Now step once into |cur_stmt| so that when we prepare the next statment
// we will have executed any dependent bytecode in this one.
int err = sqlite3_step(*cur_stmt);
if (err != SQLITE_ROW && err != SQLITE_DONE)
return base::ErrStatus("%s (errcode: %d)", sqlite3_errmsg(db), err);
// Increment the neecessary counts for the statement.
IncrementCountForStmt(cur_stmt.get(), metadata);
// Propogate the current statement to the next iteration.
prev_stmt = std::move(cur_stmt);
}
// If we didn't manage to prepare a single statment, that means everything
// in the SQL was treated as a comment.
if (!prev_stmt)
return base::ErrStatus("No valid SQL to run");
// Update the output statment and column count.
*output_stmt = std::move(prev_stmt);
metadata->column_count =
static_cast<uint32_t>(sqlite3_column_count(output_stmt->get()));
return base::OkStatus();
}
} // namespace
TraceProcessorImpl::TraceProcessorImpl(const Config& cfg)
: TraceProcessorStorageImpl(cfg) {
context_.fuchsia_trace_tokenizer.reset(new FuchsiaTraceTokenizer(&context_));
context_.fuchsia_trace_parser.reset(new FuchsiaTraceParser(&context_));
context_.systrace_trace_parser.reset(new SystraceTraceParser(&context_));
if (util::IsGzipSupported())
context_.gzip_trace_parser.reset(new GzipTraceParser(&context_));
if (json::IsJsonSupported()) {
context_.json_trace_tokenizer.reset(new JsonTraceTokenizer(&context_));
context_.json_trace_parser.reset(new JsonTraceParser(&context_));
}
RegisterAdditionalModules(&context_);
sqlite3* db = nullptr;
EnsureSqliteInitialized();
PERFETTO_CHECK(sqlite3_open(":memory:", &db) == SQLITE_OK);
InitializeSqlite(db);
CreateBuiltinTables(db);
CreateBuiltinViews(db);
db_.reset(std::move(db));
// New style function registration.
RegisterFunction<Glob>(db, "glob", 2);
RegisterFunction<Hash>(db, "HASH", -1);
RegisterFunction<Demangle>(db, "DEMANGLE", 1);
RegisterFunction<SourceGeq>(db, "SOURCE_GEQ", -1);
RegisterFunction<ExportJson>(db, "EXPORT_JSON", 1, context_.storage.get(),
false);
RegisterFunction<ExtractArg>(db, "EXTRACT_ARG", 2, context_.storage.get());
RegisterFunction<CreateFunction>(
db, "CREATE_FUNCTION", 3,
std::unique_ptr<CreateFunction::Context>(
new CreateFunction::Context{db_.get(), &create_function_state_}));
RegisterFunction<CreateViewFunction>(
db, "CREATE_VIEW_FUNCTION", 3,
std::unique_ptr<CreateViewFunction::Context>(
new CreateViewFunction::Context{db_.get()}));
// Old style function registration.
// TODO(lalitm): migrate this over to using RegisterFunction once aggregate
// functions are supported.
RegisterLastNonNullFunction(db);
RegisterValueAtMaxTsFunction(db);
SetupMetrics(this, *db_, &sql_metrics_, cfg.skip_builtin_metric_paths);
// Setup the query cache.
query_cache_.reset(new QueryCache());
const TraceStorage* storage = context_.storage.get();
SqlStatsTable::RegisterTable(*db_, storage);
StatsTable::RegisterTable(*db_, storage);
// Operator tables.
SpanJoinOperatorTable::RegisterTable(*db_, storage);
WindowOperatorTable::RegisterTable(*db_, storage);
CreateViewFunction::RegisterTable(*db_, &create_view_function_state_);
// New style tables but with some custom logic.
SqliteRawTable::RegisterTable(*db_, query_cache_.get(), &context_);
// Tables dynamically generated at query time.
RegisterDynamicTable(std::unique_ptr<ExperimentalFlamegraphGenerator>(
new ExperimentalFlamegraphGenerator(&context_)));
RegisterDynamicTable(std::unique_ptr<ExperimentalCounterDurGenerator>(
new ExperimentalCounterDurGenerator(storage->counter_table())));
RegisterDynamicTable(std::unique_ptr<DescribeSliceGenerator>(
new DescribeSliceGenerator(&context_)));
RegisterDynamicTable(std::unique_ptr<ExperimentalSliceLayoutGenerator>(
new ExperimentalSliceLayoutGenerator(
context_.storage.get()->mutable_string_pool(),
&storage->slice_table())));
RegisterDynamicTable(std::unique_ptr<AncestorGenerator>(
new AncestorGenerator(AncestorGenerator::Ancestor::kSlice, &context_)));
RegisterDynamicTable(std::unique_ptr<AncestorGenerator>(new AncestorGenerator(
AncestorGenerator::Ancestor::kStackProfileCallsite, &context_)));
RegisterDynamicTable(std::unique_ptr<AncestorGenerator>(new AncestorGenerator(
AncestorGenerator::Ancestor::kSliceByStack, &context_)));
RegisterDynamicTable(
std::unique_ptr<DescendantGenerator>(new DescendantGenerator(
DescendantGenerator::Descendant::kSlice, &context_)));
RegisterDynamicTable(
std::unique_ptr<DescendantGenerator>(new DescendantGenerator(
DescendantGenerator::Descendant::kSliceByStack, &context_)));
RegisterDynamicTable(
std::unique_ptr<ConnectedFlowGenerator>(new ConnectedFlowGenerator(
ConnectedFlowGenerator::Mode::kDirectlyConnectedFlow, &context_)));
RegisterDynamicTable(
std::unique_ptr<ConnectedFlowGenerator>(new ConnectedFlowGenerator(
ConnectedFlowGenerator::Mode::kPrecedingFlow, &context_)));
RegisterDynamicTable(
std::unique_ptr<ConnectedFlowGenerator>(new ConnectedFlowGenerator(
ConnectedFlowGenerator::Mode::kFollowingFlow, &context_)));
RegisterDynamicTable(std::unique_ptr<ExperimentalSchedUpidGenerator>(
new ExperimentalSchedUpidGenerator(storage->sched_slice_table(),
storage->thread_table())));
RegisterDynamicTable(std::unique_ptr<ThreadStateGenerator>(
new ThreadStateGenerator(&context_)));
RegisterDynamicTable(std::unique_ptr<ExperimentalAnnotatedStackGenerator>(
new ExperimentalAnnotatedStackGenerator(&context_)));
RegisterDynamicTable(std::unique_ptr<ExperimentalFlatSliceGenerator>(
new ExperimentalFlatSliceGenerator(&context_)));
// New style db-backed tables.
RegisterDbTable(storage->arg_table());
RegisterDbTable(storage->thread_table());
RegisterDbTable(storage->process_table());
RegisterDbTable(storage->slice_table());
RegisterDbTable(storage->flow_table());
RegisterDbTable(storage->thread_slice_table());
RegisterDbTable(storage->sched_slice_table());
RegisterDbTable(storage->instant_table());
RegisterDbTable(storage->gpu_slice_table());
RegisterDbTable(storage->track_table());
RegisterDbTable(storage->thread_track_table());
RegisterDbTable(storage->process_track_table());
RegisterDbTable(storage->gpu_track_table());
RegisterDbTable(storage->counter_table());
RegisterDbTable(storage->counter_track_table());
RegisterDbTable(storage->process_counter_track_table());
RegisterDbTable(storage->thread_counter_track_table());
RegisterDbTable(storage->cpu_counter_track_table());
RegisterDbTable(storage->irq_counter_track_table());
RegisterDbTable(storage->softirq_counter_track_table());
RegisterDbTable(storage->gpu_counter_track_table());
RegisterDbTable(storage->gpu_counter_group_table());
RegisterDbTable(storage->perf_counter_track_table());
RegisterDbTable(storage->heap_graph_object_table());
RegisterDbTable(storage->heap_graph_reference_table());
RegisterDbTable(storage->heap_graph_class_table());
RegisterDbTable(storage->symbol_table());
RegisterDbTable(storage->heap_profile_allocation_table());
RegisterDbTable(storage->cpu_profile_stack_sample_table());
RegisterDbTable(storage->perf_sample_table());
RegisterDbTable(storage->stack_profile_callsite_table());
RegisterDbTable(storage->stack_profile_mapping_table());
RegisterDbTable(storage->stack_profile_frame_table());
RegisterDbTable(storage->package_list_table());
RegisterDbTable(storage->profiler_smaps_table());
RegisterDbTable(storage->android_log_table());
RegisterDbTable(storage->vulkan_memory_allocations_table());
RegisterDbTable(storage->graphics_frame_slice_table());
RegisterDbTable(storage->expected_frame_timeline_slice_table());
RegisterDbTable(storage->actual_frame_timeline_slice_table());
RegisterDbTable(storage->metadata_table());
RegisterDbTable(storage->cpu_table());
RegisterDbTable(storage->cpu_freq_table());
RegisterDbTable(storage->clock_snapshot_table());
RegisterDbTable(storage->memory_snapshot_table());
RegisterDbTable(storage->process_memory_snapshot_table());
RegisterDbTable(storage->memory_snapshot_node_table());
RegisterDbTable(storage->memory_snapshot_edge_table());
}
TraceProcessorImpl::~TraceProcessorImpl() = default;
base::Status TraceProcessorImpl::Parse(TraceBlobView blob) {
bytes_parsed_ += blob.size();
return TraceProcessorStorageImpl::Parse(std::move(blob));
}
std::string TraceProcessorImpl::GetCurrentTraceName() {
if (current_trace_name_.empty())
return "";
auto size = " (" + std::to_string(bytes_parsed_ / 1024 / 1024) + " MB)";
return current_trace_name_ + size;
}
void TraceProcessorImpl::SetCurrentTraceName(const std::string& name) {
current_trace_name_ = name;
}
void TraceProcessorImpl::NotifyEndOfFile() {
if (current_trace_name_.empty())
current_trace_name_ = "Unnamed trace";
TraceProcessorStorageImpl::NotifyEndOfFile();
SchedEventTracker::GetOrCreate(&context_)->FlushPendingEvents();
context_.metadata_tracker->SetMetadata(
metadata::trace_size_bytes,
Variadic::Integer(static_cast<int64_t>(bytes_parsed_)));
BuildBoundsTable(*db_, context_.storage->GetTraceTimestampBoundsNs());
// Create a snapshot of all tables and views created so far. This is so later
// we can drop all extra tables created by the UI and reset to the original
// state (see RestoreInitialTables).
initial_tables_.clear();
auto it = ExecuteQuery(kAllTablesQuery);
while (it.Next()) {
auto value = it.Get(0);
PERFETTO_CHECK(value.type == SqlValue::Type::kString);
initial_tables_.push_back(value.string_value);
}
}
size_t TraceProcessorImpl::RestoreInitialTables() {
// Step 1: figure out what tables/views/indices we need to delete.
std::vector<std::pair<std::string, std::string>> deletion_list;
std::string msg = "Resetting DB to initial state, deleting table/views:";
for (auto it = ExecuteQuery(kAllTablesQuery); it.Next();) {
std::string name(it.Get(0).string_value);
std::string type(it.Get(1).string_value);
if (std::find(initial_tables_.begin(), initial_tables_.end(), name) ==
initial_tables_.end()) {
msg += " " + name;
deletion_list.push_back(std::make_pair(type, name));
}
}
PERFETTO_LOG("%s", msg.c_str());
// Step 2: actually delete those tables/views/indices.
for (const auto& tn : deletion_list) {
std::string query = "DROP " + tn.first + " " + tn.second;
auto it = ExecuteQuery(query);
while (it.Next()) {
}
// Index deletion can legitimately fail. If one creates an index "i" on a
// table "t" but issues the deletion in the order (t, i), the DROP index i
// will fail with "no such index" because deleting the table "t"
// automatically deletes all associated indexes.
if (!it.Status().ok() && tn.first != "index")
PERFETTO_FATAL("%s -> %s", query.c_str(), it.Status().c_message());
}
return deletion_list.size();
}
Iterator TraceProcessorImpl::ExecuteQuery(const std::string& sql) {
PERFETTO_TP_TRACE("QUERY_EXECUTE");
uint32_t sql_stats_row =
context_.storage->mutable_sql_stats()->RecordQueryBegin(
sql, base::GetWallTimeNs().count());
ScopedStmt stmt;
IteratorImpl::StmtMetadata metadata;
base::Status status =
PrepareAndStepUntilLastValidStmt(*db_, sql, &stmt, &metadata);
PERFETTO_DCHECK((status.ok() && stmt) || (!status.ok() && !stmt));
std::unique_ptr<IteratorImpl> impl(new IteratorImpl(
this, *db_, status, std::move(stmt), std::move(metadata), sql_stats_row));
return Iterator(std::move(impl));
}
void TraceProcessorImpl::InterruptQuery() {
if (!db_)
return;
query_interrupted_.store(true);
sqlite3_interrupt(db_.get());
}
bool TraceProcessorImpl::IsRootMetricField(const std::string& metric_name) {
base::Optional<uint32_t> desc_idx =
pool_.FindDescriptorIdx(".perfetto.protos.TraceMetrics");
if (!desc_idx.has_value())
return false;
auto field_idx = pool_.descriptors()[*desc_idx].FindFieldByName(metric_name);
return field_idx != nullptr;
}
base::Status TraceProcessorImpl::RegisterMetric(const std::string& path,
const std::string& sql) {
std::string stripped_sql;
for (base::StringSplitter sp(sql, '\n'); sp.Next();) {
if (strncmp(sp.cur_token(), "--", 2) != 0) {
stripped_sql.append(sp.cur_token());
stripped_sql.push_back('\n');
}
}
// Check if the metric with the given path already exists and if it does, just
// update the SQL associated with it.
auto it = std::find_if(
sql_metrics_.begin(), sql_metrics_.end(),
[&path](const metrics::SqlMetricFile& m) { return m.path == path; });
if (it != sql_metrics_.end()) {
it->sql = stripped_sql;
return base::OkStatus();
}
auto sep_idx = path.rfind('/');
std::string basename =
sep_idx == std::string::npos ? path : path.substr(sep_idx + 1);
auto sql_idx = basename.rfind(".sql");
if (sql_idx == std::string::npos) {
return base::ErrStatus("Unable to find .sql extension for metric");
}
auto no_ext_name = basename.substr(0, sql_idx);
metrics::SqlMetricFile metric;
metric.path = path;
metric.sql = stripped_sql;
if (IsRootMetricField(no_ext_name)) {
metric.proto_field_name = no_ext_name;
metric.output_table_name = no_ext_name + "_output";
auto field_it_and_inserted =
proto_field_to_sql_metric_path_.emplace(*metric.proto_field_name, path);
if (!field_it_and_inserted.second) {
// We already had a metric with this field name in the map. However, if
// this was the case, we should have found the metric in
// |path_to_sql_metric_file_| above if we are simply overriding the
// metric. Return an error since this means we have two different SQL
// files which are trying to output the same metric.
const auto& prev_path = field_it_and_inserted.first->second;
PERFETTO_DCHECK(prev_path != path);
return base::ErrStatus(
"RegisterMetric Error: Metric paths %s (which is already registered) "
"and %s are both trying to output the proto field %s",
prev_path.c_str(), path.c_str(), metric.proto_field_name->c_str());
}
InsertIntoTraceMetricsTable(*db_, no_ext_name);
}
sql_metrics_.emplace_back(metric);
return base::OkStatus();
}
base::Status TraceProcessorImpl::ExtendMetricsProto(const uint8_t* data,
size_t size) {
return ExtendMetricsProto(data, size, /*skip_prefixes*/ {});
}
base::Status TraceProcessorImpl::ExtendMetricsProto(
const uint8_t* data,
size_t size,
const std::vector<std::string>& skip_prefixes) {
base::Status status =
pool_.AddFromFileDescriptorSet(data, size, skip_prefixes);
if (!status.ok())
return status;
for (uint32_t i = 0; i < pool_.descriptors().size(); ++i) {
// Convert the full name (e.g. .perfetto.protos.TraceMetrics.SubMetric)
// into a function name of the form (TraceMetrics_SubMetric).
const auto& desc = pool_.descriptors()[i];
auto fn_name = desc.full_name().substr(desc.package_name().size() + 1);
std::replace(fn_name.begin(), fn_name.end(), '.', '_');
RegisterFunction<metrics::BuildProto>(
db_.get(), fn_name.c_str(), -1,
std::unique_ptr<metrics::BuildProto::Context>(
new metrics::BuildProto::Context{this, &pool_, i}));
}
return base::OkStatus();
}
base::Status TraceProcessorImpl::ComputeMetric(
const std::vector<std::string>& metric_names,
std::vector<uint8_t>* metrics_proto) {
auto opt_idx = pool_.FindDescriptorIdx(".perfetto.protos.TraceMetrics");
if (!opt_idx.has_value())
return base::Status("Root metrics proto descriptor not found");
const auto& root_descriptor = pool_.descriptors()[opt_idx.value()];
return metrics::ComputeMetrics(this, metric_names, sql_metrics_, pool_,
root_descriptor, metrics_proto);
}
base::Status TraceProcessorImpl::ComputeMetricText(
const std::vector<std::string>& metric_names,
TraceProcessor::MetricResultFormat format,
std::string* metrics_string) {
std::vector<uint8_t> metrics_proto;
base::Status status = ComputeMetric(metric_names, &metrics_proto);
if (!status.ok())
return status;
switch (format) {
case TraceProcessor::MetricResultFormat::kProtoText:
*metrics_string = protozero_to_text::ProtozeroToText(
pool_, ".perfetto.protos.TraceMetrics",
protozero::ConstBytes{metrics_proto.data(), metrics_proto.size()},
protozero_to_text::kIncludeNewLines);
break;
case TraceProcessor::MetricResultFormat::kJson:
// TODO(dproy): Implement this.
PERFETTO_FATAL("Json formatted metrics not supported yet.");
break;
}
return status;
}
std::vector<uint8_t> TraceProcessorImpl::GetMetricDescriptors() {
return pool_.SerializeAsDescriptorSet();
}
void TraceProcessorImpl::EnableMetatrace() {
metatrace::Enable();
}
base::Status TraceProcessorImpl::DisableAndReadMetatrace(
std::vector<uint8_t>* trace_proto) {
protozero::HeapBuffered<protos::pbzero::Trace> trace;
metatrace::DisableAndReadBuffer([&trace](metatrace::Record* record) {
auto packet = trace->add_packet();
packet->set_timestamp(record->timestamp_ns);
auto* evt = packet->set_perfetto_metatrace();
evt->set_event_name(record->event_name);
evt->set_event_duration_ns(record->duration_ns);
evt->set_thread_id(1); // Not really important, just required for the ui.
if (record->args_buffer_size == 0)
return;
base::StringSplitter s(record->args_buffer, record->args_buffer_size, '\0');
for (; s.Next();) {
auto* arg_proto = evt->add_args();
arg_proto->set_key(s.cur_token());
bool has_next = s.Next();
PERFETTO_CHECK(has_next);
arg_proto->set_value(s.cur_token());
}
});
*trace_proto = trace.SerializeAsArray();
return base::OkStatus();
}
} // namespace trace_processor
} // namespace perfetto