PT2 Bug Basher
Debug test failures and runtime errors in the PyTorch 2 compiler stack (Dynamo, Inductor, AOTAutograd, FX graphs).
Workflow Summary
- Reproduce -- Get a consistent reproduction of the failure
- Minimize -- Reduce the repro to the smallest possible standalone case. Strip away unrelated model logic, use minimal tensor shapes, and isolate the specific op or pattern that triggers the bug.
- Add a unit test -- Do this BEFORE diving into code search or root cause investigation. Add a failing test to the codebase that captures the bug. Place it in a specific, topic-appropriate test file (e.g.,
test/dynamo/test_repros.py
, test/inductor/test_torchinductor.py
, test/export/test_export.py
). Avoid — it is already oversized; find a more specific test file that matches the area of the bug. Use torch.testing._internal.common_utils.TestCase
and . The test must fail before the fix and pass after. Having the test first keeps you grounded — you know exactly what "fixed" looks like before you start exploring the codebase.
- Gather logs -- Run with appropriate settings
- Classify -- Use the Error Triage table to identify the category
- Inspect artifacts -- Check FX graphs, IR, and generated code via
- Identify root cause -- Trace from the error back through the compilation pipeline
- Fix -- Apply the fix
- Verify -- Run the new unit test AND nearby related existing tests (e.g., if you changed how works, also run the existing export test). Use to quickly run related tests by name. The task is not complete until all pass.
- Self-review -- Use the skill to review your own changes before presenting them. Fix any issues it flags.
- Celebrate -- Summarize the changes: explain the root cause, what was changed and why, and which tests were added/verified. Then tell the user the bug is squashed. Include a fun, varied motivational message or easter egg to keep spirits high (e.g., a pun, a quote, an ASCII art bug getting squashed). Keep it short and different each time.
Investigation Strategy
Prefer direct tools over meta_codesearch
Use
,
, and
directly for code exploration.
Do not spawn agents — they are slow and expensive. The
Architectural Knowledge and
Key Source Files sections below should give you enough context to know where to look. A targeted
for a function name is always faster.
Know which compilation mode you're in
Before reading implementation code, determine the compilation mode. These share code but diverge in important ways:
- -- Dynamo + Inductor. , no
_compiling_state_context()
.
- (strict) -- ,
_compiling_state_context()
active.
- (non-strict, the default) -- Uses Dynamo via but may differ from strict.
_compiling_state_context()
active. Check torch._export.config.use_new_tracer_experimental
— it changes which code path is used.
Distinguish trace-time vs runtime
Many PT2 bugs come from confusing these two:
- Trace-time: Inside Dynamo's symbolic interpreter. Dynamo intercepts function calls and may constant-fold them (e.g., → ).
- Runtime: Real tensors, real Python calls, module-level flags like
torch.compiler._is_exporting_flag
.
When debugging, add temporary
statements directly in the source file rather than monkey-patching from outside — dispatch chains make monkey-patching unreliable.
Gathering Information
Pick the right diagnostic tool based on the error category:
- Quick overview:
TORCH_LOGS="+dynamo,graph_breaks,recompiles" python your_script.py
- Full debug artifacts:
TORCH_COMPILE_DEBUG=1 python your_script.py
— creates with FX graphs, Inductor IR, and generated code
- Generated code only:
TORCH_LOGS="output_code" python your_script.py
- Structured tracing:
TORCH_TRACE=/path/to/trace python your_script.py
then
- Single-threaded (for pdb):
TORCHINDUCTOR_COMPILE_THREADS=1 python your_script.py
Error Triage
Classify the failure using the error message and traceback:
Debugging by Category
Graph Breaks
Graph breaks split the compiled graph into smaller subgraphs, often causing performance regressions or unexpected behavior.
Diagnosis:
bash
TORCH_LOGS="graph_breaks" python your_script.py
Key files:
- -- exception class
- -- where most graph break decisions happen
Common causes:
- Unsupported Python constructs (data-dependent control flow, unsupported builtins)
- Tensor operations that can't be traced (in-place ops on inputs, unsupported dtypes)
- Calls to non-traceable functions
Fix approach:
- Read the graph break message to identify the unsupported operation
- Check if there's a decomposition or supported alternative
- If the operation genuinely can't be traced, consider
torch._dynamo.allow_in_graph
or restructuring user code
Backend Compiler Failures
means Inductor (or another backend) crashed during compilation.
Diagnosis:
bash
TORCHDYNAMO_REPRO_AFTER=aot TORCHDYNAMO_REPRO_LEVEL=2 python your_script.py
This generates
that isolates the minimal failing graph.
Key files:
torch/_dynamo/repro/after_aot.py
-- repro/minifier for post-AOT failures
- -- the backend itself
Fix approach:
- Run the minifier to get a minimal reproduction
- Inspect the FX graph () to understand what ops are involved
- Check if it's a lowering issue (
torch/_inductor/lowering.py
), scheduling issue, or codegen issue
- Look at the generated output code if the error is in codegen
Recompilation Issues
Excessive recompilation happens when guards are too specific, causing cache misses.
Diagnosis:
bash
TORCH_LOGS="recompiles,recompiles_verbose,guards" python your_script.py
Key config:
torch._dynamo.config.recompile_limit
(default: 8)
torch._dynamo.config.fail_on_recompile_limit_hit
-- set to to get a hard error
Common causes:
- Changing tensor shapes without marking them dynamic
- Python scalar values that change between calls
- Global state mutations between calls
Fix approach:
- Read the recompilation reason from logs
- Identify the failing guard
- Either mark the relevant dimension as dynamic with
torch._dynamo.mark_dynamic()
or fix the source of guard instability
Accuracy Issues
The compiled model produces different numerical results than eager mode.
Diagnosis:
bash
TORCHDYNAMO_REPRO_AFTER=aot TORCHDYNAMO_REPRO_LEVEL=4 python your_script.py
This compares compiled vs. eager with an fp64 reference and dumps a repro if accuracy fails.
Key utilities:
torch/_dynamo/debug_utils.py
-- , ,
torch._dynamo.config.repro_tolerance
(default: 1e-3)
Fix approach:
- Get the minimal failing graph from the minifier
- Compare eager vs. compiled output at fp64 precision
- Binary search through ops to find the diverging operation
- Check for known numerical issues (reduction order, fused kernels, dtype promotions)
Internal Dynamo Errors
indicates a bug in Dynamo itself.
Diagnosis:
bash
TORCHDYNAMO_VERBOSE=1 python your_script.py
# or equivalently:
TORCH_LOGS="+dynamo" python your_script.py
Key files:
torch/_dynamo/symbolic_convert.py
-- bytecode interpreter
- -- variable tracking system
- -- guard generation
Fix approach:
- Get the full stack trace with
- Identify which bytecode instruction or variable type caused the crash
- Create a minimal repro (the error message often includes a minifier path)
- Debug with
TORCHINDUCTOR_COMPILE_THREADS=1
and pdb if needed
Runtime Crashes
Segfaults and CUDA illegal memory access errors during execution of compiled code.
Diagnosis (make crash deterministic):
bash
PYTORCH_NO_CUDA_MEMORY_CACHING=1 CUDA_LAUNCH_BLOCKING=1 python your_script.py
For CUDA IMA, add NaN checks:
bash
TORCHINDUCTOR_NAN_ASSERTS=1 python your_script.py
For Inductor-level sync debugging:
python
torch._inductor.config.triton.debug_sync_kernel = True # sync after every kernel
torch._inductor.config.triton.debug_sync_graph = True # sync before/after graph
Fix approach:
- Make the crash deterministic with
PYTORCH_NO_CUDA_MEMORY_CACHING=1 CUDA_LAUNCH_BLOCKING=1
- Check if it's an input mismatch (shapes, devices, dtypes)
- Inspect the generated kernel code with
- Use
TORCHINDUCTOR_NAN_ASSERTS=1
to find the first kernel producing bad values
- Check for dynamic shapes issues (historically a common source of IMA)
Triton Kernel Failures
Triton assertion failures or index-out-of-bounds in generated kernels.
Diagnosis:
bash
TORCH_LOGS="output_code,schedule" python your_script.py
Key files:
torch/_inductor/codegen/triton.py
-- Triton codegen
torch/_inductor/scheduler.py
-- kernel fusion decisions
Fix approach:
- Get the generated Triton kernel from logs
- Check index computations for off-by-one or wrong stride calculations
- Look at the IR () to trace back to the FX op
- Check if fusion decisions created invalid index combinations
Key Source Files
| File | Purpose |
|---|
| Exception hierarchy and error formatting |
torch/_dynamo/debug_utils.py
| Minifier support, accuracy checking, input serialization |
torch/_dynamo/repro/after_dynamo.py
| Repro/minifier for Dynamo-stage failures |
torch/_dynamo/repro/after_aot.py
| Repro/minifier for post-AOTAutograd failures |
torch/_dynamo/repro/aoti.py
| Repro/minifier for AOTI failures |
| Dynamo config (repro levels, recompile limits) |
torch/_dynamo/variables/torch.py
| Torch function handling, tracing state functions |
torch/_dynamo/variables/higher_order_ops.py
| HOP tracing (cond, map, etc.) |
torch/_dynamo/symbolic_convert.py
| Bytecode interpreter, InstructionTranslator |
torch/_dynamo/convert_frame.py
| Frame compilation, entry point |
torch/_dynamo/functional_export.py
| New export tracer (_dynamo_graph_capture_for_export
) |
torch/_dynamo/eval_frame.py
| , |
| Export pipeline (, , , ) |
| _compiling_state_context()
|
torch/compiler/__init__.py
| , , runtime flags |
torch/_higher_order_ops/cond.py
| implementation and proxy tracing |
torch/_higher_order_ops/utils.py
| for HOP branch tracing |
torch/_inductor/config.py
| Inductor config (debug flags, trace settings) |
| DebugContext, graph visualization, IR logging |
torch/_logging/_registrations.py
| All registered log aliases and artifacts |
Using the Minifier
The minifier reduces a failing graph to the smallest reproduction:
bash
# Step 1: Generate the minifier launcher
TORCHDYNAMO_REPRO_AFTER=aot TORCHDYNAMO_REPRO_LEVEL=2 python your_script.py
# Step 2: Run the minifier
python minifier_launcher.py minify
# Step 3: Run the minimized repro
python minifier_launcher.py run
For accuracy issues, use level 4:
bash
TORCHDYNAMO_REPRO_AFTER=aot TORCHDYNAMO_REPRO_LEVEL=4 python your_script.py