| /* |
| * Copyright (C) 2020 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #include "src/profiling/perf/unwinding.h" |
| |
| #include <cinttypes> |
| #include <mutex> |
| |
| #include <unwindstack/Unwinder.h> |
| |
| #include "perfetto/ext/base/metatrace.h" |
| #include "perfetto/ext/base/no_destructor.h" |
| #include "perfetto/ext/base/thread_utils.h" |
| #include "perfetto/ext/base/utils.h" |
| |
| namespace { |
| constexpr size_t kUnwindingMaxFrames = 1000; |
| constexpr uint32_t kDataSourceShutdownRetryDelayMs = 400; |
| } // namespace |
| |
| namespace perfetto { |
| namespace profiling { |
| |
| Unwinder::Delegate::~Delegate() = default; |
| |
| Unwinder::Unwinder(Delegate* delegate, base::UnixTaskRunner* task_runner) |
| : task_runner_(task_runner), delegate_(delegate) { |
| ResetAndEnableUnwindstackCache(); |
| base::MaybeSetThreadName("stack-unwinding"); |
| } |
| |
| void Unwinder::PostStartDataSource(DataSourceInstanceID ds_id, |
| bool kernel_frames) { |
| // No need for a weak pointer as the associated task runner quits (stops |
| // running tasks) strictly before the Unwinder's destruction. |
| task_runner_->PostTask( |
| [this, ds_id, kernel_frames] { StartDataSource(ds_id, kernel_frames); }); |
| } |
| |
| void Unwinder::StartDataSource(DataSourceInstanceID ds_id, bool kernel_frames) { |
| PERFETTO_DCHECK_THREAD(thread_checker_); |
| PERFETTO_DLOG("Unwinder::StartDataSource(%zu)", static_cast<size_t>(ds_id)); |
| |
| auto it_and_inserted = data_sources_.emplace(ds_id, DataSourceState{}); |
| PERFETTO_DCHECK(it_and_inserted.second); |
| |
| if (kernel_frames) { |
| kernel_symbolizer_.GetOrCreateKernelSymbolMap(); |
| } |
| } |
| |
| // c++11: use shared_ptr to transfer resource handles, so that the resources get |
| // released even if the task runner is destroyed with pending tasks. |
| // "Cleverness" warning: |
| // the task will be executed on a different thread, and will mutate the |
| // pointed-to memory. It may be the case that this posting thread will not |
| // decrement its shared_ptr refcount until *after* the task has executed. In |
| // that scenario, the destruction of the pointed-to memory will be happening on |
| // the posting thread. This implies a data race between the mutation on the task |
| // thread, and the destruction on the posting thread. *However*, we assume that |
| // there is no race in practice due to refcount decrements having |
| // release-acquire semantics. The refcount decrements pair with each other, and |
| // therefore also serve as a memory barrier between the destructor, and any |
| // previous modifications of the pointed-to memory. |
| // TODO(rsavitski): present a more convincing argument, or reimplement |
| // without relying on shared_ptr implementation details. |
| void Unwinder::PostAdoptProcDescriptors(DataSourceInstanceID ds_id, |
| pid_t pid, |
| base::ScopedFile maps_fd, |
| base::ScopedFile mem_fd) { |
| auto shared_maps = std::make_shared<base::ScopedFile>(std::move(maps_fd)); |
| auto shared_mem = std::make_shared<base::ScopedFile>(std::move(mem_fd)); |
| task_runner_->PostTask([this, ds_id, pid, shared_maps, shared_mem] { |
| base::ScopedFile maps = std::move(*shared_maps.get()); |
| base::ScopedFile mem = std::move(*shared_mem.get()); |
| AdoptProcDescriptors(ds_id, pid, std::move(maps), std::move(mem)); |
| }); |
| } |
| |
| void Unwinder::AdoptProcDescriptors(DataSourceInstanceID ds_id, |
| pid_t pid, |
| base::ScopedFile maps_fd, |
| base::ScopedFile mem_fd) { |
| PERFETTO_DCHECK_THREAD(thread_checker_); |
| PERFETTO_DLOG("Unwinder::AdoptProcDescriptors(%zu, %d, %d, %d)", |
| static_cast<size_t>(ds_id), static_cast<int>(pid), |
| maps_fd.get(), mem_fd.get()); |
| |
| auto it = data_sources_.find(ds_id); |
| if (it == data_sources_.end()) |
| return; |
| DataSourceState& ds = it->second; |
| |
| ProcessState& proc_state = ds.process_states[pid]; // insert if new |
| PERFETTO_DCHECK(proc_state.status == ProcessState::Status::kInitial || |
| proc_state.status == ProcessState::Status::kFdsTimedOut); |
| PERFETTO_DCHECK(!proc_state.unwind_state.has_value()); |
| |
| PERFETTO_METATRACE_SCOPED(TAG_PRODUCER, PROFILER_MAPS_PARSE); |
| |
| proc_state.status = ProcessState::Status::kFdsResolved; |
| proc_state.unwind_state = |
| UnwindingMetadata{std::move(maps_fd), std::move(mem_fd)}; |
| } |
| |
| void Unwinder::PostRecordTimedOutProcDescriptors(DataSourceInstanceID ds_id, |
| pid_t pid) { |
| task_runner_->PostTask([this, ds_id, pid] { |
| UpdateProcessStateStatus(ds_id, pid, ProcessState::Status::kFdsTimedOut); |
| }); |
| } |
| |
| void Unwinder::PostRecordNoUserspaceProcess(DataSourceInstanceID ds_id, |
| pid_t pid) { |
| task_runner_->PostTask([this, ds_id, pid] { |
| UpdateProcessStateStatus(ds_id, pid, ProcessState::Status::kNoUserspace); |
| }); |
| } |
| |
| void Unwinder::UpdateProcessStateStatus(DataSourceInstanceID ds_id, |
| pid_t pid, |
| ProcessState::Status new_status) { |
| PERFETTO_DCHECK_THREAD(thread_checker_); |
| PERFETTO_DLOG("Unwinder::UpdateProcessStateStatus(%zu, %d, %d)", |
| static_cast<size_t>(ds_id), static_cast<int>(pid), |
| static_cast<int>(new_status)); |
| |
| auto it = data_sources_.find(ds_id); |
| if (it == data_sources_.end()) |
| return; |
| DataSourceState& ds = it->second; |
| |
| ProcessState& proc_state = ds.process_states[pid]; // insert if new |
| proc_state.status = new_status; |
| } |
| |
| void Unwinder::PostProcessQueue() { |
| task_runner_->PostTask([this] { ProcessQueue(); }); |
| } |
| |
| // Note: we always walk the queue in order. So if there are multiple data |
| // sources, one of which is shutting down, its shutdown can be delayed by |
| // unwinding of other sources' samples. Instead, we could scan the queue |
| // multiple times, prioritizing the samples for shutting-down sources. At the |
| // time of writing, the earlier is considered to be fair enough. |
| void Unwinder::ProcessQueue() { |
| PERFETTO_DCHECK_THREAD(thread_checker_); |
| PERFETTO_METATRACE_SCOPED(TAG_PRODUCER, PROFILER_UNWIND_TICK); |
| PERFETTO_DLOG("Unwinder::ProcessQueue"); |
| |
| base::FlatSet<DataSourceInstanceID> pending_sample_sources = |
| ConsumeAndUnwindReadySamples(); |
| |
| // Deal with the possiblity of data sources that are shutting down. |
| bool post_delayed_reprocess = false; |
| base::FlatSet<DataSourceInstanceID> sources_to_stop; |
| for (auto& id_and_ds : data_sources_) { |
| DataSourceInstanceID ds_id = id_and_ds.first; |
| const DataSourceState& ds = id_and_ds.second; |
| |
| if (ds.status == DataSourceState::Status::kActive) |
| continue; |
| |
| // Data source that is shutting down. If we're still waiting on proc-fds (or |
| // the lookup to time out) for samples in the queue - repost a later |
| // attempt (as there is no guarantee that there are any readers waking up |
| // the unwinder anymore). |
| if (pending_sample_sources.count(ds_id)) { |
| PERFETTO_DLOG( |
| "Unwinder delaying DS(%zu) stop: waiting on a pending sample", |
| static_cast<size_t>(ds_id)); |
| post_delayed_reprocess = true; |
| } else { |
| // Otherwise, proceed with tearing down data source state (after |
| // completing the loop, to avoid invalidating the iterator). |
| sources_to_stop.insert(ds_id); |
| } |
| } |
| |
| for (auto ds_id : sources_to_stop) |
| FinishDataSourceStop(ds_id); |
| |
| if (post_delayed_reprocess) |
| task_runner_->PostDelayedTask([this] { ProcessQueue(); }, |
| kDataSourceShutdownRetryDelayMs); |
| } |
| |
| base::FlatSet<DataSourceInstanceID> Unwinder::ConsumeAndUnwindReadySamples() { |
| PERFETTO_DCHECK_THREAD(thread_checker_); |
| base::FlatSet<DataSourceInstanceID> pending_sample_sources; |
| |
| // Use a single snapshot of the ring buffer pointers. |
| ReadView read_view = unwind_queue_.BeginRead(); |
| |
| PERFETTO_METATRACE_COUNTER( |
| TAG_PRODUCER, PROFILER_UNWIND_QUEUE_SZ, |
| static_cast<int32_t>(read_view.write_pos - read_view.read_pos)); |
| |
| if (read_view.read_pos == read_view.write_pos) |
| return pending_sample_sources; |
| |
| // Walk the queue. |
| for (auto read_pos = read_view.read_pos; read_pos < read_view.write_pos; |
| read_pos++) { |
| UnwindEntry& entry = unwind_queue_.at(read_pos); |
| |
| if (!entry.valid) |
| continue; // already processed |
| |
| uint64_t sampled_stack_bytes = entry.sample.stack.size(); |
| |
| // Data source might be gone due to an abrupt stop. |
| auto it = data_sources_.find(entry.data_source_id); |
| if (it == data_sources_.end()) { |
| entry = UnwindEntry::Invalid(); |
| DecrementEnqueuedFootprint(sampled_stack_bytes); |
| continue; |
| } |
| DataSourceState& ds = it->second; |
| |
| pid_t pid = entry.sample.common.pid; |
| ProcessState& proc_state = ds.process_states[pid]; // insert if new |
| |
| // Giving up on the sample (proc-fd lookup timed out). |
| if (proc_state.status == ProcessState::Status::kFdsTimedOut) { |
| PERFETTO_DLOG("Unwinder skipping sample for pid [%d]: kFdsTimedOut", |
| static_cast<int>(pid)); |
| |
| // free up the sampled stack as the main thread has no use for it |
| entry.sample.stack.clear(); |
| entry.sample.stack.shrink_to_fit(); |
| |
| delegate_->PostEmitUnwinderSkippedSample(entry.data_source_id, |
| std::move(entry.sample)); |
| entry = UnwindEntry::Invalid(); |
| DecrementEnqueuedFootprint(sampled_stack_bytes); |
| continue; |
| } |
| |
| // Still waiting to be notified how to handle this process. |
| if (proc_state.status == ProcessState::Status::kInitial) { |
| PERFETTO_DLOG("Unwinder deferring sample for pid [%d]", |
| static_cast<int>(pid)); |
| |
| pending_sample_sources.insert(entry.data_source_id); |
| continue; |
| } |
| |
| // b/324757089: we are not precisely tracking process lifetimes, so the |
| // sample might be for a different process that reused the pid since the |
| // start of the session. Normally this is both infrequent and not a problem |
| // since the unwinding will fail due to invalidated procfs descriptors. |
| // However we need this explicit skip for the specific case of a kernel |
| // thread reusing a userspace pid, as the unwinding doesn't expect absent |
| // userspace state for a thought-to-be-userspace process. |
| // TODO(rsavitski): start tracking process exits more accurately, either |
| // via PERF_RECORD_EXIT records or by checking the validity of the procfs |
| // descriptors. |
| if (PERFETTO_UNLIKELY(!entry.sample.regs && |
| proc_state.status == |
| ProcessState::Status::kFdsResolved)) { |
| PERFETTO_DLOG( |
| "Unwinder discarding sample for pid [%d]: uspace->kthread pid reuse", |
| static_cast<int>(pid)); |
| |
| PERFETTO_CHECK(sampled_stack_bytes == 0); |
| entry = UnwindEntry::Invalid(); |
| continue; |
| } |
| |
| // Sample ready - process it. |
| if (proc_state.status == ProcessState::Status::kFdsResolved || |
| proc_state.status == ProcessState::Status::kNoUserspace) { |
| // Metatrace: emit both a scoped slice, as well as a "counter" |
| // representing the pid being unwound. |
| PERFETTO_METATRACE_SCOPED(TAG_PRODUCER, PROFILER_UNWIND_SAMPLE); |
| PERFETTO_METATRACE_COUNTER(TAG_PRODUCER, PROFILER_UNWIND_CURRENT_PID, |
| static_cast<int32_t>(pid)); |
| |
| PERFETTO_CHECK(proc_state.status == ProcessState::Status::kNoUserspace || |
| proc_state.unwind_state.has_value()); |
| |
| UnwindingMetadata* opt_user_state = |
| (proc_state.unwind_state.has_value() |
| ? &proc_state.unwind_state.value() |
| : nullptr); |
| CompletedSample unwound_sample = UnwindSample( |
| entry.sample, opt_user_state, proc_state.attempted_unwinding); |
| proc_state.attempted_unwinding = true; |
| |
| PERFETTO_METATRACE_COUNTER(TAG_PRODUCER, PROFILER_UNWIND_CURRENT_PID, 0); |
| |
| delegate_->PostEmitSample(entry.data_source_id, |
| std::move(unwound_sample)); |
| entry = UnwindEntry::Invalid(); |
| DecrementEnqueuedFootprint(sampled_stack_bytes); |
| continue; |
| } |
| } |
| |
| // Consume all leading processed entries in the queue. |
| auto new_read_pos = read_view.read_pos; |
| for (; new_read_pos < read_view.write_pos; new_read_pos++) { |
| UnwindEntry& entry = unwind_queue_.at(new_read_pos); |
| if (entry.valid) |
| break; |
| } |
| if (new_read_pos != read_view.read_pos) |
| unwind_queue_.CommitNewReadPosition(new_read_pos); |
| |
| PERFETTO_METATRACE_COUNTER( |
| TAG_PRODUCER, PROFILER_UNWIND_QUEUE_SZ, |
| static_cast<int32_t>(read_view.write_pos - new_read_pos)); |
| |
| PERFETTO_DLOG("Unwind queue drain: [%" PRIu64 "]->[%" PRIu64 "]", |
| read_view.write_pos - read_view.read_pos, |
| read_view.write_pos - new_read_pos); |
| |
| return pending_sample_sources; |
| } |
| |
| CompletedSample Unwinder::UnwindSample(const ParsedSample& sample, |
| UnwindingMetadata* opt_user_state, |
| bool pid_unwound_before) { |
| PERFETTO_DCHECK_THREAD(thread_checker_); |
| |
| CompletedSample ret; |
| ret.common = sample.common; |
| |
| // Symbolize kernel-unwound kernel frames, if appropriate. |
| std::vector<unwindstack::FrameData> kernel_frames = |
| SymbolizeKernelCallchain(sample); |
| |
| size_t kernel_frames_size = kernel_frames.size(); |
| ret.frames = std::move(kernel_frames); |
| ret.build_ids.resize(kernel_frames_size, ""); |
| |
| // Perform userspace unwinding using libunwindstack, if appropriate. |
| if (!opt_user_state) |
| return ret; |
| |
| // Overlay the stack bytes over /proc/<pid>/mem. |
| UnwindingMetadata* unwind_state = opt_user_state; |
| std::shared_ptr<unwindstack::Memory> overlay_memory = |
| std::make_shared<StackOverlayMemory>( |
| unwind_state->fd_mem, sample.regs->sp(), |
| reinterpret_cast<const uint8_t*>(sample.stack.data()), |
| sample.stack.size()); |
| |
| struct UnwindResult { |
| unwindstack::ErrorCode error_code; |
| uint64_t warnings; |
| std::vector<unwindstack::FrameData> frames; |
| |
| UnwindResult(unwindstack::ErrorCode e, |
| uint64_t w, |
| std::vector<unwindstack::FrameData> f) |
| : error_code(e), warnings(w), frames(std::move(f)) {} |
| UnwindResult(const UnwindResult&) = delete; |
| UnwindResult& operator=(const UnwindResult&) = delete; |
| UnwindResult(UnwindResult&&) __attribute__((unused)) = default; |
| UnwindResult& operator=(UnwindResult&&) = default; |
| }; |
| auto attempt_unwind = [&sample, unwind_state, pid_unwound_before, |
| &overlay_memory]() -> UnwindResult { |
| metatrace::ScopedEvent m(metatrace::TAG_PRODUCER, |
| pid_unwound_before |
| ? metatrace::PROFILER_UNWIND_ATTEMPT |
| : metatrace::PROFILER_UNWIND_INITIAL_ATTEMPT); |
| |
| // Unwindstack clobbers registers, so make a copy in case of retries. |
| auto regs_copy = std::unique_ptr<unwindstack::Regs>{sample.regs->Clone()}; |
| |
| unwindstack::Unwinder unwinder(kUnwindingMaxFrames, &unwind_state->fd_maps, |
| regs_copy.get(), overlay_memory); |
| #if PERFETTO_BUILDFLAG(PERFETTO_ANDROID_BUILD) |
| unwinder.SetJitDebug(unwind_state->GetJitDebug(regs_copy->Arch())); |
| unwinder.SetDexFiles(unwind_state->GetDexFiles(regs_copy->Arch())); |
| #endif |
| unwinder.Unwind(/*initial_map_names_to_skip=*/nullptr, |
| /*map_suffixes_to_ignore=*/nullptr); |
| return {unwinder.LastErrorCode(), unwinder.warnings(), |
| unwinder.ConsumeFrames()}; |
| }; |
| |
| // first unwind attempt |
| UnwindResult unwind = attempt_unwind(); |
| |
| bool should_retry = unwind.error_code == unwindstack::ERROR_INVALID_MAP || |
| unwind.warnings & unwindstack::WARNING_DEX_PC_NOT_IN_MAP; |
| |
| // ERROR_INVALID_MAP means that unwinding reached a point in memory without a |
| // corresponding mapping. This is possible if the parsed /proc/pid/maps is |
| // outdated. Reparse and try again. |
| // |
| // Special case: skip reparsing if the stack sample was (most likely) |
| // truncated. We perform the best-effort unwind of the sampled part, but an |
| // error around the truncated part is not unexpected. |
| // |
| // TODO(rsavitski): consider rate-limiting unwind retries. |
| if (should_retry && sample.stack_maxed) { |
| PERFETTO_DLOG("Skipping reparse/reunwind due to maxed stack for tid [%d]", |
| static_cast<int>(sample.common.tid)); |
| } else if (should_retry) { |
| { |
| PERFETTO_METATRACE_SCOPED(TAG_PRODUCER, PROFILER_MAPS_REPARSE); |
| PERFETTO_DLOG("Reparsing maps for pid [%d]", |
| static_cast<int>(sample.common.pid)); |
| unwind_state->ReparseMaps(); |
| } |
| // reunwind attempt |
| unwind = attempt_unwind(); |
| } |
| |
| ret.build_ids.reserve(kernel_frames_size + unwind.frames.size()); |
| ret.frames.reserve(kernel_frames_size + unwind.frames.size()); |
| for (unwindstack::FrameData& frame : unwind.frames) { |
| ret.build_ids.emplace_back(unwind_state->GetBuildId(frame)); |
| ret.frames.emplace_back(std::move(frame)); |
| } |
| |
| // In case of an unwinding error, add a synthetic error frame (which will |
| // appear as a caller of the partially-unwound fragment), for easier |
| // visualization of errors. |
| if (unwind.error_code != unwindstack::ERROR_NONE) { |
| PERFETTO_DLOG("Unwinding error %" PRIu8, unwind.error_code); |
| unwindstack::FrameData frame_data{}; |
| frame_data.function_name = |
| "ERROR " + StringifyLibUnwindstackError(unwind.error_code); |
| ret.frames.emplace_back(std::move(frame_data)); |
| ret.build_ids.emplace_back(""); |
| ret.unwind_error = unwind.error_code; |
| } |
| |
| PERFETTO_CHECK(ret.build_ids.size() == ret.frames.size()); |
| return ret; |
| } |
| |
| std::vector<unwindstack::FrameData> Unwinder::SymbolizeKernelCallchain( |
| const ParsedSample& sample) { |
| static base::NoDestructor<std::shared_ptr<unwindstack::MapInfo>> |
| kernel_map_info(unwindstack::MapInfo::Create(0, 0, 0, 0, "kernel")); |
| std::vector<unwindstack::FrameData> ret; |
| if (sample.kernel_ips.empty()) |
| return ret; |
| |
| // The list of addresses contains special context marker values (inserted by |
| // the kernel's unwinding) to indicate which section of the callchain belongs |
| // to the kernel/user mode (if the kernel can successfully unwind user |
| // stacks). In our case, we request only the kernel frames. |
| if (sample.kernel_ips[0] != PERF_CONTEXT_KERNEL) { |
| PERFETTO_DFATAL_OR_ELOG( |
| "Unexpected: 0th frame of callchain is not PERF_CONTEXT_KERNEL."); |
| return ret; |
| } |
| |
| auto* kernel_map = kernel_symbolizer_.GetOrCreateKernelSymbolMap(); |
| PERFETTO_DCHECK(kernel_map); |
| ret.reserve(sample.kernel_ips.size()); |
| for (size_t i = 1; i < sample.kernel_ips.size(); i++) { |
| std::string function_name = kernel_map->Lookup(sample.kernel_ips[i]); |
| |
| // Synthesise a partially-valid libunwindstack frame struct for the kernel |
| // frame. We reuse the type for convenience. The kernel frames are marked by |
| // a magical "kernel" MapInfo object as their containing mapping. |
| unwindstack::FrameData frame{}; |
| frame.function_name = std::move(function_name); |
| frame.map_info = kernel_map_info.ref(); |
| ret.emplace_back(std::move(frame)); |
| } |
| return ret; |
| } |
| |
| void Unwinder::PostInitiateDataSourceStop(DataSourceInstanceID ds_id) { |
| task_runner_->PostTask([this, ds_id] { InitiateDataSourceStop(ds_id); }); |
| } |
| |
| void Unwinder::InitiateDataSourceStop(DataSourceInstanceID ds_id) { |
| PERFETTO_DCHECK_THREAD(thread_checker_); |
| PERFETTO_DLOG("Unwinder::InitiateDataSourceStop(%zu)", |
| static_cast<size_t>(ds_id)); |
| |
| auto it = data_sources_.find(ds_id); |
| if (it == data_sources_.end()) |
| return; |
| DataSourceState& ds = it->second; |
| |
| PERFETTO_CHECK(ds.status == DataSourceState::Status::kActive); |
| ds.status = DataSourceState::Status::kShuttingDown; |
| |
| // Make sure that there's an outstanding task to process the unwinding queue, |
| // as it is the point that evaluates the stop condition. |
| PostProcessQueue(); |
| } |
| |
| void Unwinder::FinishDataSourceStop(DataSourceInstanceID ds_id) { |
| PERFETTO_DCHECK_THREAD(thread_checker_); |
| PERFETTO_DLOG("Unwinder::FinishDataSourceStop(%zu)", |
| static_cast<size_t>(ds_id)); |
| |
| auto it = data_sources_.find(ds_id); |
| if (it == data_sources_.end()) |
| return; |
| DataSourceState& ds = it->second; |
| |
| // Drop unwinder's state tied to the source. |
| PERFETTO_CHECK(ds.status == DataSourceState::Status::kShuttingDown); |
| data_sources_.erase(it); |
| |
| // Clean up state if there are no more active sources. |
| if (data_sources_.empty()) { |
| kernel_symbolizer_.Destroy(); |
| ResetAndEnableUnwindstackCache(); |
| } |
| |
| // Inform service thread that the unwinder is done with the source. |
| delegate_->PostFinishDataSourceStop(ds_id); |
| } |
| |
| void Unwinder::PostPurgeDataSource(DataSourceInstanceID ds_id) { |
| task_runner_->PostTask([this, ds_id] { PurgeDataSource(ds_id); }); |
| } |
| |
| void Unwinder::PurgeDataSource(DataSourceInstanceID ds_id) { |
| PERFETTO_DCHECK_THREAD(thread_checker_); |
| PERFETTO_DLOG("Unwinder::PurgeDataSource(%zu)", static_cast<size_t>(ds_id)); |
| |
| auto it = data_sources_.find(ds_id); |
| if (it == data_sources_.end()) |
| return; |
| |
| data_sources_.erase(it); |
| |
| // Clean up state if there are no more active sources. |
| if (data_sources_.empty()) { |
| kernel_symbolizer_.Destroy(); |
| ResetAndEnableUnwindstackCache(); |
| // Also purge scudo on Android, which would normally be done by the service |
| // thread in |FinishDataSourceStop|. This is important as most of the scudo |
| // overhead comes from libunwindstack. |
| base::MaybeReleaseAllocatorMemToOS(); |
| } |
| } |
| |
| void Unwinder::PostClearCachedStatePeriodic(DataSourceInstanceID ds_id, |
| uint32_t period_ms) { |
| task_runner_->PostDelayedTask( |
| [this, ds_id, period_ms] { ClearCachedStatePeriodic(ds_id, period_ms); }, |
| period_ms); |
| } |
| |
| // See header for rationale. |
| void Unwinder::ClearCachedStatePeriodic(DataSourceInstanceID ds_id, |
| uint32_t period_ms) { |
| auto it = data_sources_.find(ds_id); |
| if (it == data_sources_.end()) |
| return; // stop the periodic task |
| |
| DataSourceState& ds = it->second; |
| if (ds.status != DataSourceState::Status::kActive) |
| return; |
| |
| PERFETTO_METATRACE_SCOPED(TAG_PRODUCER, PROFILER_UNWIND_CACHE_CLEAR); |
| PERFETTO_DLOG("Clearing unwinder's cached state."); |
| |
| for (auto& pid_and_process : ds.process_states) { |
| if (pid_and_process.second.status == ProcessState::Status::kFdsResolved) |
| pid_and_process.second.unwind_state->fd_maps.Reset(); |
| } |
| ResetAndEnableUnwindstackCache(); |
| base::MaybeReleaseAllocatorMemToOS(); |
| |
| PostClearCachedStatePeriodic(ds_id, period_ms); // repost |
| } |
| |
| void Unwinder::ResetAndEnableUnwindstackCache() { |
| PERFETTO_DLOG("Resetting unwindstack cache"); |
| // Libunwindstack uses an unsynchronized variable for setting/checking whether |
| // the cache is enabled. Therefore unwinding and cache toggling should stay on |
| // the same thread, but we might be moving unwinding across threads if we're |
| // recreating |Unwinder| instances (during a reconnect to traced). Therefore, |
| // use our own static lock to synchronize the cache toggling. |
| // TODO(rsavitski): consider fixing this in libunwindstack itself. |
| static std::mutex* lock = new std::mutex{}; |
| std::lock_guard<std::mutex> guard{*lock}; |
| unwindstack::Elf::SetCachingEnabled(false); // free any existing state |
| unwindstack::Elf::SetCachingEnabled(true); // reallocate a fresh cache |
| } |
| |
| } // namespace profiling |
| } // namespace perfetto |