Skip to content
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Next Next commit
gh-145458: use self.skip_idle
  • Loading branch information
KowalskiThomas committed Mar 3, 2026
commit 7d031a137ca71af09bc582387dd4b5da6e2c6b10
8 changes: 4 additions & 4 deletions Lib/profiling/sampling/stack_collector.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,9 +19,9 @@ def __init__(self, sample_interval_usec, *, skip_idle=False):
self.sample_interval_usec = sample_interval_usec
self.skip_idle = skip_idle

def collect(self, stack_frames, timestamps_us=None, skip_idle=False):
def collect(self, stack_frames, timestamps_us=None):
weight = len(timestamps_us) if timestamps_us else 1
for frames, thread_id in self._iter_stacks(stack_frames, skip_idle=skip_idle):
for frames, thread_id in self._iter_stacks(stack_frames, skip_idle=self.skip_idle):
self.process_frames(frames, thread_id, weight=weight)

def process_frames(self, frames, thread_id, weight=1):
Expand Down Expand Up @@ -88,7 +88,7 @@ def __init__(self, *args, **kwargs):
# Per-thread statistics
self.per_thread_stats = {} # {thread_id: {has_gil, on_cpu, gil_requested, unknown, has_exception, total, gc_samples}}

def collect(self, stack_frames, timestamps_us=None, skip_idle=False):
def collect(self, stack_frames, timestamps_us=None):
"""Override to track thread status statistics before processing frames."""
# Weight is number of timestamps (samples with identical stack)
weight = len(timestamps_us) if timestamps_us else 1
Expand Down Expand Up @@ -123,7 +123,7 @@ def collect(self, stack_frames, timestamps_us=None, skip_idle=False):
self.per_thread_stats[thread_id][key] += value * weight

# Call parent collect to process frames
super().collect(stack_frames, timestamps_us, skip_idle=skip_idle)
super().collect(stack_frames, timestamps_us)

def set_stats(self, sample_interval_usec, duration_sec, sample_rate,
error_rate=None, missed_samples=None, mode=None):
Expand Down
Loading