| /* |
| * Copyright (C) 2017 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef SRC_TRACE_PROCESSOR_STORAGE_TRACE_STORAGE_H_ |
| #define SRC_TRACE_PROCESSOR_STORAGE_TRACE_STORAGE_H_ |
| |
| #include <array> |
| #include <deque> |
| #include <map> |
| #include <optional> |
| #include <string> |
| #include <unordered_map> |
| #include <utility> |
| #include <vector> |
| |
| #include "perfetto/base/logging.h" |
| #include "perfetto/base/time.h" |
| #include "perfetto/ext/base/hash.h" |
| #include "perfetto/ext/base/string_view.h" |
| #include "perfetto/ext/base/utils.h" |
| #include "perfetto/trace_processor/basic_types.h" |
| #include "perfetto/trace_processor/status.h" |
| #include "src/trace_processor/containers/string_pool.h" |
| #include "src/trace_processor/storage/metadata.h" |
| #include "src/trace_processor/storage/stats.h" |
| #include "src/trace_processor/tables/android_tables_py.h" |
| #include "src/trace_processor/tables/counter_tables_py.h" |
| #include "src/trace_processor/tables/flow_tables_py.h" |
| #include "src/trace_processor/tables/memory_tables_py.h" |
| #include "src/trace_processor/tables/metadata_tables_py.h" |
| #include "src/trace_processor/tables/profiler_tables_py.h" |
| #include "src/trace_processor/tables/slice_tables_py.h" |
| #include "src/trace_processor/tables/trace_proto_tables_py.h" |
| #include "src/trace_processor/tables/track_tables_py.h" |
| #include "src/trace_processor/types/variadic.h" |
| #include "src/trace_processor/views/slice_views.h" |
| |
| namespace perfetto { |
| namespace trace_processor { |
| |
| // UniquePid is an offset into |unique_processes_|. This is necessary because |
| // Unix pids are reused and thus not guaranteed to be unique over a long |
| // period of time. |
| using UniquePid = uint32_t; |
| |
| // UniqueTid is an offset into |unique_threads_|. Necessary because tids can |
| // be reused. |
| using UniqueTid = uint32_t; |
| |
| // StringId is an offset into |string_pool_|. |
| using StringId = StringPool::Id; |
| static const StringId kNullStringId = StringId::Null(); |
| |
| using ArgSetId = uint32_t; |
| static const ArgSetId kInvalidArgSetId = 0; |
| |
| using TrackId = tables::TrackTable::Id; |
| |
| using CounterId = tables::CounterTable::Id; |
| |
| using SliceId = tables::SliceTable::Id; |
| |
| using SchedId = tables::SchedSliceTable::Id; |
| |
| using MappingId = tables::StackProfileMappingTable::Id; |
| |
| using FrameId = tables::StackProfileFrameTable::Id; |
| |
| using SymbolId = tables::SymbolTable::Id; |
| |
| using CallsiteId = tables::StackProfileCallsiteTable::Id; |
| |
| using MetadataId = tables::MetadataTable::Id; |
| |
| using RawId = tables::RawTable::Id; |
| |
| using FlamegraphId = tables::ExperimentalFlamegraphNodesTable::Id; |
| |
| using VulkanAllocId = tables::VulkanMemoryAllocationsTable::Id; |
| |
| using ProcessMemorySnapshotId = tables::ProcessMemorySnapshotTable::Id; |
| |
| using SnapshotNodeId = tables::MemorySnapshotNodeTable::Id; |
| |
| static const TrackId kInvalidTrackId = |
| TrackId(std::numeric_limits<uint32_t>::max()); |
| |
| enum class RefType { |
| kRefNoRef = 0, |
| kRefUtid = 1, |
| kRefCpuId = 2, |
| kRefIrq = 3, |
| kRefSoftIrq = 4, |
| kRefUpid = 5, |
| kRefGpuId = 6, |
| kRefTrack = 7, |
| kRefMax |
| }; |
| |
| const std::vector<NullTermStringView>& GetRefTypeStringMap(); |
| |
| // Stores a data inside a trace file in a columnar form. This makes it efficient |
| // to read or search across a single field of the trace (e.g. all the thread |
| // names for a given CPU). |
| class TraceStorage { |
| public: |
| TraceStorage(const Config& = Config()); |
| |
| virtual ~TraceStorage(); |
| |
| class VirtualTrackSlices { |
| public: |
| inline uint32_t AddVirtualTrackSlice(SliceId slice_id, |
| int64_t thread_timestamp_ns, |
| int64_t thread_duration_ns, |
| int64_t thread_instruction_count, |
| int64_t thread_instruction_delta) { |
| slice_ids_.emplace_back(slice_id); |
| thread_timestamp_ns_.emplace_back(thread_timestamp_ns); |
| thread_duration_ns_.emplace_back(thread_duration_ns); |
| thread_instruction_counts_.emplace_back(thread_instruction_count); |
| thread_instruction_deltas_.emplace_back(thread_instruction_delta); |
| return slice_count() - 1; |
| } |
| |
| uint32_t slice_count() const { |
| return static_cast<uint32_t>(slice_ids_.size()); |
| } |
| |
| const std::deque<SliceId>& slice_ids() const { return slice_ids_; } |
| const std::deque<int64_t>& thread_timestamp_ns() const { |
| return thread_timestamp_ns_; |
| } |
| const std::deque<int64_t>& thread_duration_ns() const { |
| return thread_duration_ns_; |
| } |
| const std::deque<int64_t>& thread_instruction_counts() const { |
| return thread_instruction_counts_; |
| } |
| const std::deque<int64_t>& thread_instruction_deltas() const { |
| return thread_instruction_deltas_; |
| } |
| |
| std::optional<uint32_t> FindRowForSliceId(SliceId slice_id) const { |
| auto it = |
| std::lower_bound(slice_ids().begin(), slice_ids().end(), slice_id); |
| if (it != slice_ids().end() && *it == slice_id) { |
| return static_cast<uint32_t>(std::distance(slice_ids().begin(), it)); |
| } |
| return std::nullopt; |
| } |
| |
| void UpdateThreadDeltasForSliceId(SliceId slice_id, |
| int64_t end_thread_timestamp_ns, |
| int64_t end_thread_instruction_count) { |
| auto opt_row = FindRowForSliceId(slice_id); |
| if (!opt_row) |
| return; |
| uint32_t row = *opt_row; |
| int64_t begin_ns = thread_timestamp_ns_[row]; |
| thread_duration_ns_[row] = end_thread_timestamp_ns - begin_ns; |
| int64_t begin_ticount = thread_instruction_counts_[row]; |
| thread_instruction_deltas_[row] = |
| end_thread_instruction_count - begin_ticount; |
| } |
| |
| private: |
| std::deque<SliceId> slice_ids_; |
| std::deque<int64_t> thread_timestamp_ns_; |
| std::deque<int64_t> thread_duration_ns_; |
| std::deque<int64_t> thread_instruction_counts_; |
| std::deque<int64_t> thread_instruction_deltas_; |
| }; |
| |
| class SqlStats { |
| public: |
| static constexpr size_t kMaxLogEntries = 100; |
| uint32_t RecordQueryBegin(const std::string& query, int64_t time_started); |
| void RecordQueryFirstNext(uint32_t row, int64_t time_first_next); |
| void RecordQueryEnd(uint32_t row, int64_t time_end); |
| size_t size() const { return queries_.size(); } |
| const std::deque<std::string>& queries() const { return queries_; } |
| const std::deque<int64_t>& times_started() const { return times_started_; } |
| const std::deque<int64_t>& times_first_next() const { |
| return times_first_next_; |
| } |
| const std::deque<int64_t>& times_ended() const { return times_ended_; } |
| |
| private: |
| uint32_t popped_queries_ = 0; |
| |
| std::deque<std::string> queries_; |
| std::deque<int64_t> times_started_; |
| std::deque<int64_t> times_first_next_; |
| std::deque<int64_t> times_ended_; |
| }; |
| |
| struct Stats { |
| using IndexMap = std::map<int, int64_t>; |
| int64_t value = 0; |
| IndexMap indexed_values; |
| }; |
| using StatsMap = std::array<Stats, stats::kNumKeys>; |
| |
| // Return an unqiue identifier for the contents of each string. |
| // The string is copied internally and can be destroyed after this called. |
| // Virtual for testing. |
| virtual StringId InternString(base::StringView str) { |
| return string_pool_.InternString(str); |
| } |
| |
| // Example usage: SetStats(stats::android_log_num_failed, 42); |
| void SetStats(size_t key, int64_t value) { |
| PERFETTO_DCHECK(key < stats::kNumKeys); |
| PERFETTO_DCHECK(stats::kTypes[key] == stats::kSingle); |
| stats_[key].value = value; |
| } |
| |
| // Example usage: IncrementStats(stats::android_log_num_failed, -1); |
| void IncrementStats(size_t key, int64_t increment = 1) { |
| PERFETTO_DCHECK(key < stats::kNumKeys); |
| PERFETTO_DCHECK(stats::kTypes[key] == stats::kSingle); |
| stats_[key].value += increment; |
| } |
| |
| // Example usage: IncrementIndexedStats(stats::cpu_failure, 1); |
| void IncrementIndexedStats(size_t key, int index, int64_t increment = 1) { |
| PERFETTO_DCHECK(key < stats::kNumKeys); |
| PERFETTO_DCHECK(stats::kTypes[key] == stats::kIndexed); |
| stats_[key].indexed_values[index] += increment; |
| } |
| |
| // Example usage: SetIndexedStats(stats::cpu_failure, 1, 42); |
| void SetIndexedStats(size_t key, int index, int64_t value) { |
| PERFETTO_DCHECK(key < stats::kNumKeys); |
| PERFETTO_DCHECK(stats::kTypes[key] == stats::kIndexed); |
| stats_[key].indexed_values[index] = value; |
| } |
| |
| // Example usage: opt_cpu_failure = GetIndexedStats(stats::cpu_failure, 1); |
| std::optional<int64_t> GetIndexedStats(size_t key, int index) { |
| PERFETTO_DCHECK(key < stats::kNumKeys); |
| PERFETTO_DCHECK(stats::kTypes[key] == stats::kIndexed); |
| auto kv = stats_[key].indexed_values.find(index); |
| if (kv != stats_[key].indexed_values.end()) { |
| return kv->second; |
| } |
| return std::nullopt; |
| } |
| |
| class ScopedStatsTracer { |
| public: |
| ScopedStatsTracer(TraceStorage* storage, size_t key) |
| : storage_(storage), key_(key), start_ns_(base::GetWallTimeNs()) {} |
| |
| ~ScopedStatsTracer() { |
| if (!storage_) |
| return; |
| auto delta_ns = base::GetWallTimeNs() - start_ns_; |
| storage_->IncrementStats(key_, delta_ns.count()); |
| } |
| |
| ScopedStatsTracer(ScopedStatsTracer&& other) noexcept { MoveImpl(&other); } |
| |
| ScopedStatsTracer& operator=(ScopedStatsTracer&& other) { |
| MoveImpl(&other); |
| return *this; |
| } |
| |
| private: |
| ScopedStatsTracer(const ScopedStatsTracer&) = delete; |
| ScopedStatsTracer& operator=(const ScopedStatsTracer&) = delete; |
| |
| void MoveImpl(ScopedStatsTracer* other) { |
| storage_ = other->storage_; |
| key_ = other->key_; |
| start_ns_ = other->start_ns_; |
| other->storage_ = nullptr; |
| } |
| |
| TraceStorage* storage_; |
| size_t key_; |
| base::TimeNanos start_ns_; |
| }; |
| |
| ScopedStatsTracer TraceExecutionTimeIntoStats(size_t key) { |
| return ScopedStatsTracer(this, key); |
| } |
| |
| // Reading methods. |
| // Virtual for testing. |
| virtual NullTermStringView GetString(StringId id) const { |
| return string_pool_.Get(id); |
| } |
| |
| // Requests the removal of unused capacity. |
| // Matches the semantics of std::vector::shrink_to_fit. |
| void ShrinkToFitTables() { |
| // At the moment, we only bother calling ShrinkToFit on a set group |
| // of tables. If we wanted to extend this to every table, we'd need to deal |
| // with tracking all the tables in the storage: this is not worth doing |
| // given most memory is used by these tables. |
| thread_table_.ShrinkToFit(); |
| process_table_.ShrinkToFit(); |
| track_table_.ShrinkToFit(); |
| counter_table_.ShrinkToFit(); |
| slice_table_.ShrinkToFit(); |
| raw_table_.ShrinkToFit(); |
| sched_slice_table_.ShrinkToFit(); |
| thread_state_table_.ShrinkToFit(); |
| arg_table_.ShrinkToFit(); |
| } |
| |
| const tables::ThreadTable& thread_table() const { return thread_table_; } |
| tables::ThreadTable* mutable_thread_table() { return &thread_table_; } |
| |
| const tables::ProcessTable& process_table() const { return process_table_; } |
| tables::ProcessTable* mutable_process_table() { return &process_table_; } |
| |
| const tables::FiledescriptorTable& filedescriptor_table() const { |
| return filedescriptor_table_; |
| } |
| tables::FiledescriptorTable* mutable_filedescriptor_table() { |
| return &filedescriptor_table_; |
| } |
| |
| const tables::TrackTable& track_table() const { return track_table_; } |
| tables::TrackTable* mutable_track_table() { return &track_table_; } |
| |
| const tables::CounterTrackTable& counter_track_table() const { |
| return counter_track_table_; |
| } |
| tables::CounterTrackTable* mutable_counter_track_table() { |
| return &counter_track_table_; |
| } |
| |
| const tables::CpuCounterTrackTable& cpu_counter_track_table() const { |
| return cpu_counter_track_table_; |
| } |
| tables::CpuCounterTrackTable* mutable_cpu_counter_track_table() { |
| return &cpu_counter_track_table_; |
| } |
| |
| const tables::GpuCounterGroupTable& gpu_counter_group_table() const { |
| return gpu_counter_group_table_; |
| } |
| tables::GpuCounterGroupTable* mutable_gpu_counter_group_table() { |
| return &gpu_counter_group_table_; |
| } |
| |
| const tables::GpuCounterTrackTable& gpu_counter_track_table() const { |
| return gpu_counter_track_table_; |
| } |
| tables::GpuCounterTrackTable* mutable_gpu_counter_track_table() { |
| return &gpu_counter_track_table_; |
| } |
| |
| const tables::EnergyCounterTrackTable& energy_counter_track_table() const { |
| return energy_counter_track_table_; |
| } |
| tables::EnergyCounterTrackTable* mutable_energy_counter_track_table() { |
| return &energy_counter_track_table_; |
| } |
| |
| const tables::UidCounterTrackTable& uid_counter_track_table() const { |
| return uid_counter_track_table_; |
| } |
| tables::UidCounterTrackTable* mutable_uid_counter_track_table() { |
| return &uid_counter_track_table_; |
| } |
| |
| const tables::EnergyPerUidCounterTrackTable& |
| energy_per_uid_counter_track_table() const { |
| return energy_per_uid_counter_track_table_; |
| } |
| tables::EnergyPerUidCounterTrackTable* |
| mutable_energy_per_uid_counter_track_table() { |
| return &energy_per_uid_counter_track_table_; |
| } |
| |
| const tables::IrqCounterTrackTable& irq_counter_track_table() const { |
| return irq_counter_track_table_; |
| } |
| tables::IrqCounterTrackTable* mutable_irq_counter_track_table() { |
| return &irq_counter_track_table_; |
| } |
| |
| const tables::PerfCounterTrackTable& perf_counter_track_table() const { |
| return perf_counter_track_table_; |
| } |
| tables::PerfCounterTrackTable* mutable_perf_counter_track_table() { |
| return &perf_counter_track_table_; |
| } |
| |
| const tables::ProcessCounterTrackTable& process_counter_track_table() const { |
| return process_counter_track_table_; |
| } |
| tables::ProcessCounterTrackTable* mutable_process_counter_track_table() { |
| return &process_counter_track_table_; |
| } |
| |
| const tables::ProcessTrackTable& process_track_table() const { |
| return process_track_table_; |
| } |
| tables::ProcessTrackTable* mutable_process_track_table() { |
| return &process_track_table_; |
| } |
| |
| const tables::ThreadTrackTable& thread_track_table() const { |
| return thread_track_table_; |
| } |
| tables::ThreadTrackTable* mutable_thread_track_table() { |
| return &thread_track_table_; |
| } |
| |
| const tables::ThreadStateTable& thread_state_table() const { |
| return thread_state_table_; |
| } |
| tables::ThreadStateTable* mutable_thread_state_table() { |
| return &thread_state_table_; |
| } |
| |
| const tables::ThreadCounterTrackTable& thread_counter_track_table() const { |
| return thread_counter_track_table_; |
| } |
| tables::ThreadCounterTrackTable* mutable_thread_counter_track_table() { |
| return &thread_counter_track_table_; |
| } |
| |
| const tables::SoftirqCounterTrackTable& softirq_counter_track_table() const { |
| return softirq_counter_track_table_; |
| } |
| tables::SoftirqCounterTrackTable* mutable_softirq_counter_track_table() { |
| return &softirq_counter_track_table_; |
| } |
| |
| const tables::SchedSliceTable& sched_slice_table() const { |
| return sched_slice_table_; |
| } |
| tables::SchedSliceTable* mutable_sched_slice_table() { |
| return &sched_slice_table_; |
| } |
| |
| const tables::SliceTable& slice_table() const { return slice_table_; } |
| tables::SliceTable* mutable_slice_table() { return &slice_table_; } |
| |
| const tables::FlowTable& flow_table() const { return flow_table_; } |
| tables::FlowTable* mutable_flow_table() { return &flow_table_; } |
| |
| const VirtualTrackSlices& virtual_track_slices() const { |
| return virtual_track_slices_; |
| } |
| VirtualTrackSlices* mutable_virtual_track_slices() { |
| return &virtual_track_slices_; |
| } |
| |
| const tables::GpuSliceTable& gpu_slice_table() const { |
| return gpu_slice_table_; |
| } |
| tables::GpuSliceTable* mutable_gpu_slice_table() { return &gpu_slice_table_; } |
| |
| const tables::CounterTable& counter_table() const { return counter_table_; } |
| tables::CounterTable* mutable_counter_table() { return &counter_table_; } |
| |
| const SqlStats& sql_stats() const { return sql_stats_; } |
| SqlStats* mutable_sql_stats() { return &sql_stats_; } |
| |
| const tables::AndroidLogTable& android_log_table() const { |
| return android_log_table_; |
| } |
| tables::AndroidLogTable* mutable_android_log_table() { |
| return &android_log_table_; |
| } |
| |
| const tables::AndroidDumpstateTable& android_dumpstate_table() const { |
| return android_dumpstate_table_; |
| } |
| |
| tables::AndroidDumpstateTable* mutable_android_dumpstate_table() { |
| return &android_dumpstate_table_; |
| } |
| |
| const StatsMap& stats() const { return stats_; } |
| |
| const tables::MetadataTable& metadata_table() const { |
| return metadata_table_; |
| } |
| tables::MetadataTable* mutable_metadata_table() { return &metadata_table_; } |
| |
| const tables::ClockSnapshotTable& clock_snapshot_table() const { |
| return clock_snapshot_table_; |
| } |
| tables::ClockSnapshotTable* mutable_clock_snapshot_table() { |
| return &clock_snapshot_table_; |
| } |
| |
| const tables::ArgTable& arg_table() const { return arg_table_; } |
| tables::ArgTable* mutable_arg_table() { return &arg_table_; } |
| |
| const tables::RawTable& raw_table() const { return raw_table_; } |
| tables::RawTable* mutable_raw_table() { return &raw_table_; } |
| |
| const tables::FtraceEventTable& ftrace_event_table() const { |
| return ftrace_event_table_; |
| } |
| tables::FtraceEventTable* mutable_ftrace_event_table() { |
| return &ftrace_event_table_; |
| } |
| |
| const tables::CpuTable& cpu_table() const { return cpu_table_; } |
| tables::CpuTable* mutable_cpu_table() { return &cpu_table_; } |
| |
| const tables::CpuFreqTable& cpu_freq_table() const { return cpu_freq_table_; } |
| tables::CpuFreqTable* mutable_cpu_freq_table() { return &cpu_freq_table_; } |
| |
| const tables::StackProfileMappingTable& stack_profile_mapping_table() const { |
| return stack_profile_mapping_table_; |
| } |
| tables::StackProfileMappingTable* mutable_stack_profile_mapping_table() { |
| return &stack_profile_mapping_table_; |
| } |
| |
| const tables::StackProfileFrameTable& stack_profile_frame_table() const { |
| return stack_profile_frame_table_; |
| } |
| tables::StackProfileFrameTable* mutable_stack_profile_frame_table() { |
| return &stack_profile_frame_table_; |
| } |
| |
| const tables::StackProfileCallsiteTable& stack_profile_callsite_table() |
| const { |
| return stack_profile_callsite_table_; |
| } |
| tables::StackProfileCallsiteTable* mutable_stack_profile_callsite_table() { |
| return &stack_profile_callsite_table_; |
| } |
| |
| const tables::HeapProfileAllocationTable& heap_profile_allocation_table() |
| const { |
| return heap_profile_allocation_table_; |
| } |
| tables::HeapProfileAllocationTable* mutable_heap_profile_allocation_table() { |
| return &heap_profile_allocation_table_; |
| } |
| |
| const tables::PackageListTable& package_list_table() const { |
| return package_list_table_; |
| } |
| tables::PackageListTable* mutable_package_list_table() { |
| return &package_list_table_; |
| } |
| |
| const tables::AndroidGameInterventionListTable& |
| android_game_intervention_list_table() const { |
| return android_game_intervention_list_table_; |
| } |
| tables::AndroidGameInterventionListTable* |
| mutable_android_game_intervenion_list_table() { |
| return &android_game_intervention_list_table_; |
| } |
| |
| const tables::ProfilerSmapsTable& profiler_smaps_table() const { |
| return profiler_smaps_table_; |
| } |
| tables::ProfilerSmapsTable* mutable_profiler_smaps_table() { |
| return &profiler_smaps_table_; |
| } |
| |
| const tables::StackSampleTable& stack_sample_table() const { |
| return stack_sample_table_; |
| } |
| tables::StackSampleTable* mutable_stack_sample_table() { |
| return &stack_sample_table_; |
| } |
| |
| const tables::CpuProfileStackSampleTable& cpu_profile_stack_sample_table() |
| const { |
| return cpu_profile_stack_sample_table_; |
| } |
| tables::CpuProfileStackSampleTable* mutable_cpu_profile_stack_sample_table() { |
| return &cpu_profile_stack_sample_table_; |
| } |
| |
| const tables::PerfSampleTable& perf_sample_table() const { |
| return perf_sample_table_; |
| } |
| tables::PerfSampleTable* mutable_perf_sample_table() { |
| return &perf_sample_table_; |
| } |
| |
| const tables::SymbolTable& symbol_table() const { return symbol_table_; } |
| |
| tables::SymbolTable* mutable_symbol_table() { return &symbol_table_; } |
| |
| const tables::HeapGraphObjectTable& heap_graph_object_table() const { |
| return heap_graph_object_table_; |
| } |
| |
| tables::HeapGraphObjectTable* mutable_heap_graph_object_table() { |
| return &heap_graph_object_table_; |
| } |
| const tables::HeapGraphClassTable& heap_graph_class_table() const { |
| return heap_graph_class_table_; |
| } |
| |
| tables::HeapGraphClassTable* mutable_heap_graph_class_table() { |
| return &heap_graph_class_table_; |
| } |
| |
| const tables::HeapGraphReferenceTable& heap_graph_reference_table() const { |
| return heap_graph_reference_table_; |
| } |
| |
| tables::HeapGraphReferenceTable* mutable_heap_graph_reference_table() { |
| return &heap_graph_reference_table_; |
| } |
| |
| const tables::CpuTrackTable& cpu_track_table() const { |
| return cpu_track_table_; |
| } |
| tables::CpuTrackTable* mutable_cpu_track_table() { return &cpu_track_table_; } |
| |
| const tables::GpuTrackTable& gpu_track_table() const { |
| return gpu_track_table_; |
| } |
| tables::GpuTrackTable* mutable_gpu_track_table() { return &gpu_track_table_; } |
| |
| const tables::VulkanMemoryAllocationsTable& vulkan_memory_allocations_table() |
| const { |
| return vulkan_memory_allocations_table_; |
| } |
| |
| tables::VulkanMemoryAllocationsTable* |
| mutable_vulkan_memory_allocations_table() { |
| return &vulkan_memory_allocations_table_; |
| } |
| |
| const tables::GraphicsFrameSliceTable& graphics_frame_slice_table() const { |
| return graphics_frame_slice_table_; |
| } |
| |
| tables::GraphicsFrameSliceTable* mutable_graphics_frame_slice_table() { |
| return &graphics_frame_slice_table_; |
| } |
| |
| const tables::MemorySnapshotTable& memory_snapshot_table() const { |
| return memory_snapshot_table_; |
| } |
| tables::MemorySnapshotTable* mutable_memory_snapshot_table() { |
| return &memory_snapshot_table_; |
| } |
| |
| const tables::ProcessMemorySnapshotTable& process_memory_snapshot_table() |
| const { |
| return process_memory_snapshot_table_; |
| } |
| tables::ProcessMemorySnapshotTable* mutable_process_memory_snapshot_table() { |
| return &process_memory_snapshot_table_; |
| } |
| |
| const tables::MemorySnapshotNodeTable& memory_snapshot_node_table() const { |
| return memory_snapshot_node_table_; |
| } |
| tables::MemorySnapshotNodeTable* mutable_memory_snapshot_node_table() { |
| return &memory_snapshot_node_table_; |
| } |
| |
| const tables::MemorySnapshotEdgeTable& memory_snapshot_edge_table() const { |
| return memory_snapshot_edge_table_; |
| } |
| tables::MemorySnapshotEdgeTable* mutable_memory_snapshot_edge_table() { |
| return &memory_snapshot_edge_table_; |
| } |
| |
| const tables::ExpectedFrameTimelineSliceTable& |
| expected_frame_timeline_slice_table() const { |
| return expected_frame_timeline_slice_table_; |
| } |
| |
| tables::ExpectedFrameTimelineSliceTable* |
| mutable_expected_frame_timeline_slice_table() { |
| return &expected_frame_timeline_slice_table_; |
| } |
| |
| const tables::ActualFrameTimelineSliceTable& |
| actual_frame_timeline_slice_table() const { |
| return actual_frame_timeline_slice_table_; |
| } |
| tables::ActualFrameTimelineSliceTable* |
| mutable_actual_frame_timeline_slice_table() { |
| return &actual_frame_timeline_slice_table_; |
| } |
| |
| const tables::ExperimentalProtoPathTable& experimental_proto_path_table() |
| const { |
| return experimental_proto_path_table_; |
| } |
| tables::ExperimentalProtoPathTable* mutable_experimental_proto_path_table() { |
| return &experimental_proto_path_table_; |
| } |
| |
| const tables::ExperimentalProtoContentTable& |
| experimental_proto_content_table() const { |
| return experimental_proto_content_table_; |
| } |
| tables::ExperimentalProtoContentTable* |
| mutable_experimental_proto_content_table() { |
| return &experimental_proto_content_table_; |
| } |
| |
| const tables::ExpMissingChromeProcTable& |
| experimental_missing_chrome_processes_table() const { |
| return experimental_missing_chrome_processes_table_; |
| } |
| tables::ExpMissingChromeProcTable* |
| mutable_experimental_missing_chrome_processes_table() { |
| return &experimental_missing_chrome_processes_table_; |
| } |
| |
| const views::ThreadSliceView& thread_slice_view() const { |
| return thread_slice_view_; |
| } |
| |
| const StringPool& string_pool() const { return string_pool_; } |
| StringPool* mutable_string_pool() { return &string_pool_; } |
| |
| // Number of interned strings in the pool. Includes the empty string w/ ID=0. |
| size_t string_count() const { return string_pool_.size(); } |
| |
| // Start / end ts (in nanoseconds) across the parsed trace events. |
| // Returns (0, 0) if the trace is empty. |
| std::pair<int64_t, int64_t> GetTraceTimestampBoundsNs() const; |
| |
| util::Status ExtractArg(uint32_t arg_set_id, |
| const char* key, |
| std::optional<Variadic>* result) { |
| const auto& args = arg_table(); |
| RowMap filtered = args.FilterToRowMap( |
| {args.arg_set_id().eq(arg_set_id), args.key().eq(key)}); |
| if (filtered.empty()) { |
| *result = std::nullopt; |
| return util::OkStatus(); |
| } |
| if (filtered.size() > 1) { |
| return util::ErrStatus( |
| "EXTRACT_ARG: received multiple args matching arg set id and key"); |
| } |
| uint32_t idx = filtered.Get(0); |
| *result = GetArgValue(idx); |
| return util::OkStatus(); |
| } |
| |
| Variadic GetArgValue(uint32_t row) const { |
| Variadic v; |
| v.type = *GetVariadicTypeForId(arg_table_.value_type()[row]); |
| |
| // Force initialization of union to stop GCC complaining. |
| v.int_value = 0; |
| |
| switch (v.type) { |
| case Variadic::Type::kBool: |
| v.bool_value = static_cast<bool>(*arg_table_.int_value()[row]); |
| break; |
| case Variadic::Type::kInt: |
| v.int_value = *arg_table_.int_value()[row]; |
| break; |
| case Variadic::Type::kUint: |
| v.uint_value = static_cast<uint64_t>(*arg_table_.int_value()[row]); |
| break; |
| case Variadic::Type::kString: { |
| auto opt_value = arg_table_.string_value()[row]; |
| v.string_value = opt_value ? *opt_value : kNullStringId; |
| break; |
| } |
| case Variadic::Type::kPointer: |
| v.pointer_value = static_cast<uint64_t>(*arg_table_.int_value()[row]); |
| break; |
| case Variadic::Type::kReal: |
| v.real_value = *arg_table_.real_value()[row]; |
| break; |
| case Variadic::Type::kJson: { |
| auto opt_value = arg_table_.string_value()[row]; |
| v.json_value = opt_value ? *opt_value : kNullStringId; |
| break; |
| } |
| case Variadic::Type::kNull: |
| break; |
| } |
| return v; |
| } |
| |
| StringId GetIdForVariadicType(Variadic::Type type) const { |
| return variadic_type_ids_[type]; |
| } |
| |
| std::optional<Variadic::Type> GetVariadicTypeForId(StringId id) const { |
| auto it = |
| std::find(variadic_type_ids_.begin(), variadic_type_ids_.end(), id); |
| if (it == variadic_type_ids_.end()) |
| return std::nullopt; |
| |
| int64_t idx = std::distance(variadic_type_ids_.begin(), it); |
| return static_cast<Variadic::Type>(idx); |
| } |
| |
| private: |
| using StringHash = uint64_t; |
| |
| TraceStorage(const TraceStorage&) = delete; |
| TraceStorage& operator=(const TraceStorage&) = delete; |
| |
| TraceStorage(TraceStorage&&) = delete; |
| TraceStorage& operator=(TraceStorage&&) = delete; |
| |
| // One entry for each unique string in the trace. |
| StringPool string_pool_; |
| |
| // Stats about parsing the trace. |
| StatsMap stats_{}; |
| |
| // Extra data extracted from the trace. Includes: |
| // * metadata from chrome and benchmarking infrastructure |
| // * descriptions of android packages |
| tables::MetadataTable metadata_table_{&string_pool_}; |
| |
| // Contains data from all the clock snapshots in the trace. |
| tables::ClockSnapshotTable clock_snapshot_table_{&string_pool_}; |
| |
| // Metadata for tracks. |
| tables::TrackTable track_table_{&string_pool_}; |
| tables::ThreadStateTable thread_state_table_{&string_pool_}; |
| tables::CpuTrackTable cpu_track_table_{&string_pool_, &track_table_}; |
| tables::GpuTrackTable gpu_track_table_{&string_pool_, &track_table_}; |
| tables::ProcessTrackTable process_track_table_{&string_pool_, &track_table_}; |
| tables::ThreadTrackTable thread_track_table_{&string_pool_, &track_table_}; |
| |
| // Track tables for counter events. |
| tables::CounterTrackTable counter_track_table_{&string_pool_, &track_table_}; |
| tables::ThreadCounterTrackTable thread_counter_track_table_{ |
| &string_pool_, &counter_track_table_}; |
| tables::ProcessCounterTrackTable process_counter_track_table_{ |
| &string_pool_, &counter_track_table_}; |
| tables::CpuCounterTrackTable cpu_counter_track_table_{&string_pool_, |
| &counter_track_table_}; |
| tables::IrqCounterTrackTable irq_counter_track_table_{&string_pool_, |
| &counter_track_table_}; |
| tables::SoftirqCounterTrackTable softirq_counter_track_table_{ |
| &string_pool_, &counter_track_table_}; |
| tables::GpuCounterTrackTable gpu_counter_track_table_{&string_pool_, |
| &counter_track_table_}; |
| tables::EnergyCounterTrackTable energy_counter_track_table_{ |
| &string_pool_, &counter_track_table_}; |
| tables::UidCounterTrackTable uid_counter_track_table_{&string_pool_, |
| &counter_track_table_}; |
| tables::EnergyPerUidCounterTrackTable energy_per_uid_counter_track_table_{ |
| &string_pool_, &uid_counter_track_table_}; |
| tables::GpuCounterGroupTable gpu_counter_group_table_{&string_pool_}; |
| tables::PerfCounterTrackTable perf_counter_track_table_{ |
| &string_pool_, &counter_track_table_}; |
| |
| // Args for all other tables. |
| tables::ArgTable arg_table_{&string_pool_}; |
| |
| // Information about all the threads and processes in the trace. |
| tables::ThreadTable thread_table_{&string_pool_}; |
| tables::ProcessTable process_table_{&string_pool_}; |
| tables::FiledescriptorTable filedescriptor_table_{&string_pool_}; |
| |
| // Slices coming from userspace events (e.g. Chromium TRACE_EVENT macros). |
| tables::SliceTable slice_table_{&string_pool_}; |
| |
| // Flow events from userspace events (e.g. Chromium TRACE_EVENT macros). |
| tables::FlowTable flow_table_{&string_pool_}; |
| |
| // Slices from CPU scheduling data. |
| tables::SchedSliceTable sched_slice_table_{&string_pool_}; |
| |
| // Additional attributes for virtual track slices (sub-type of |
| // NestableSlices). |
| VirtualTrackSlices virtual_track_slices_; |
| |
| // Additional attributes for gpu track slices (sub-type of |
| // NestableSlices). |
| tables::GpuSliceTable gpu_slice_table_{&string_pool_, &slice_table_}; |
| |
| // The values from the Counter events from the trace. This includes CPU |
| // frequency events as well systrace trace_marker counter events. |
| tables::CounterTable counter_table_{&string_pool_}; |
| |
| SqlStats sql_stats_; |
| |
| tables::RawTable raw_table_{&string_pool_}; |
| tables::FtraceEventTable ftrace_event_table_{&string_pool_, &raw_table_}; |
| |
| tables::CpuTable cpu_table_{&string_pool_}; |
| |
| tables::CpuFreqTable cpu_freq_table_{&string_pool_}; |
| |
| tables::AndroidLogTable android_log_table_{&string_pool_}; |
| |
| tables::AndroidDumpstateTable android_dumpstate_table_{&string_pool_}; |
| |
| tables::StackProfileMappingTable stack_profile_mapping_table_{&string_pool_}; |
| tables::StackProfileFrameTable stack_profile_frame_table_{&string_pool_}; |
| tables::StackProfileCallsiteTable stack_profile_callsite_table_{ |
| &string_pool_}; |
| tables::StackSampleTable stack_sample_table_{&string_pool_}; |
| tables::HeapProfileAllocationTable heap_profile_allocation_table_{ |
| &string_pool_}; |
| tables::CpuProfileStackSampleTable cpu_profile_stack_sample_table_{ |
| &string_pool_, &stack_sample_table_}; |
| tables::PerfSampleTable perf_sample_table_{&string_pool_}; |
| tables::PackageListTable package_list_table_{&string_pool_}; |
| tables::AndroidGameInterventionListTable |
| android_game_intervention_list_table_{&string_pool_}; |
| tables::ProfilerSmapsTable profiler_smaps_table_{&string_pool_}; |
| |
| // Symbol tables (mappings from frames to symbol names) |
| tables::SymbolTable symbol_table_{&string_pool_}; |
| tables::HeapGraphObjectTable heap_graph_object_table_{&string_pool_}; |
| tables::HeapGraphClassTable heap_graph_class_table_{&string_pool_}; |
| tables::HeapGraphReferenceTable heap_graph_reference_table_{&string_pool_}; |
| |
| tables::VulkanMemoryAllocationsTable vulkan_memory_allocations_table_{ |
| &string_pool_}; |
| |
| tables::GraphicsFrameSliceTable graphics_frame_slice_table_{&string_pool_, |
| &slice_table_}; |
| |
| // Metadata for memory snapshot. |
| tables::MemorySnapshotTable memory_snapshot_table_{&string_pool_}; |
| tables::ProcessMemorySnapshotTable process_memory_snapshot_table_{ |
| &string_pool_}; |
| tables::MemorySnapshotNodeTable memory_snapshot_node_table_{&string_pool_}; |
| tables::MemorySnapshotEdgeTable memory_snapshot_edge_table_{&string_pool_}; |
| |
| // FrameTimeline tables |
| tables::ExpectedFrameTimelineSliceTable expected_frame_timeline_slice_table_{ |
| &string_pool_, &slice_table_}; |
| tables::ActualFrameTimelineSliceTable actual_frame_timeline_slice_table_{ |
| &string_pool_, &slice_table_}; |
| |
| tables::ExperimentalProtoPathTable experimental_proto_path_table_{ |
| &string_pool_}; |
| tables::ExperimentalProtoContentTable experimental_proto_content_table_{ |
| &string_pool_}; |
| |
| tables::ExpMissingChromeProcTable |
| experimental_missing_chrome_processes_table_{&string_pool_}; |
| |
| views::ThreadSliceView thread_slice_view_{&slice_table_, &thread_track_table_, |
| &thread_table_}; |
| |
| // The below array allow us to map between enums and their string |
| // representations. |
| std::array<StringId, Variadic::kMaxType + 1> variadic_type_ids_; |
| }; |
| |
| } // namespace trace_processor |
| } // namespace perfetto |
| |
| template <> |
| struct std::hash<::perfetto::trace_processor::BaseId> { |
| using argument_type = ::perfetto::trace_processor::BaseId; |
| using result_type = size_t; |
| |
| result_type operator()(const argument_type& r) const { |
| return std::hash<uint32_t>{}(r.value); |
| } |
| }; |
| |
| template <> |
| struct std::hash<::perfetto::trace_processor::TrackId> |
| : std::hash<::perfetto::trace_processor::BaseId> {}; |
| template <> |
| struct std::hash<::perfetto::trace_processor::MappingId> |
| : std::hash<::perfetto::trace_processor::BaseId> {}; |
| template <> |
| struct std::hash<::perfetto::trace_processor::CallsiteId> |
| : std::hash<::perfetto::trace_processor::BaseId> {}; |
| template <> |
| struct std::hash<::perfetto::trace_processor::FrameId> |
| : std::hash<::perfetto::trace_processor::BaseId> {}; |
| template <> |
| struct std::hash<::perfetto::trace_processor::tables::HeapGraphObjectTable::Id> |
| : std::hash<::perfetto::trace_processor::BaseId> {}; |
| |
| template <> |
| struct std::hash< |
| ::perfetto::trace_processor::tables::StackProfileFrameTable::Row> { |
| using argument_type = |
| ::perfetto::trace_processor::tables::StackProfileFrameTable::Row; |
| using result_type = size_t; |
| |
| result_type operator()(const argument_type& r) const { |
| return std::hash<::perfetto::trace_processor::StringId>{}(r.name) ^ |
| std::hash<std::optional<::perfetto::trace_processor::MappingId>>{}( |
| r.mapping) ^ |
| std::hash<int64_t>{}(r.rel_pc); |
| } |
| }; |
| |
| template <> |
| struct std::hash< |
| ::perfetto::trace_processor::tables::StackProfileCallsiteTable::Row> { |
| using argument_type = |
| ::perfetto::trace_processor::tables::StackProfileCallsiteTable::Row; |
| using result_type = size_t; |
| |
| result_type operator()(const argument_type& r) const { |
| return std::hash<int64_t>{}(r.depth) ^ |
| std::hash<std::optional<::perfetto::trace_processor::CallsiteId>>{}( |
| r.parent_id) ^ |
| std::hash<::perfetto::trace_processor::FrameId>{}(r.frame_id); |
| } |
| }; |
| |
| template <> |
| struct std::hash< |
| ::perfetto::trace_processor::tables::StackProfileMappingTable::Row> { |
| using argument_type = |
| ::perfetto::trace_processor::tables::StackProfileMappingTable::Row; |
| using result_type = size_t; |
| |
| result_type operator()(const argument_type& r) const { |
| return std::hash<::perfetto::trace_processor::StringId>{}(r.build_id) ^ |
| std::hash<int64_t>{}(r.exact_offset) ^ |
| std::hash<int64_t>{}(r.start_offset) ^ |
| std::hash<int64_t>{}(r.start) ^ std::hash<int64_t>{}(r.end) ^ |
| std::hash<int64_t>{}(r.load_bias) ^ |
| std::hash<::perfetto::trace_processor::StringId>{}(r.name); |
| } |
| }; |
| |
| #endif // SRC_TRACE_PROCESSOR_STORAGE_TRACE_STORAGE_H_ |