Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 0 additions & 1 deletion .lightning/workflows/transformer-engine.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,6 @@ parametrize:
matrix:
test_file:
- test_transformer_engine_executor.py
- test_transformer_engine_v1_executor.py

run: |
whereis nvidia
Expand Down
5 changes: 0 additions & 5 deletions thunder/benchmarks/benchmark_litgpt.py
Original file line number Diff line number Diff line change
Expand Up @@ -686,11 +686,6 @@ def setup_compile(self, model):

executors.insert(0, torch_compile_ex)

if "transformerengine_v1" in self.compile:
from thunder.executors.transformer_engine_v1ex import transformer_engine_v1_ex

executors.insert(0, transformer_engine_v1_ex)

elif "transformerengine" in self.compile:
from thunder.executors.transformer_engineex import (
transformer_engine_ex,
Expand Down
12 changes: 0 additions & 12 deletions thunder/core/trace.py
Original file line number Diff line number Diff line change
Expand Up @@ -406,18 +406,6 @@ def keyfn(class_or_module: type | ModuleType) -> str:
program.append("")

if include_decorators:
# NOTE: For TransformerEngine executor, we want to wrap the generated
# forward function in fp8_autocast ctx manager.
# In the future, if other executor has similar requirements, we should
# add a new extension point for executors
# NOTE: For TE v1.6 onwards, `fp8_autocast` checks if `torch.is_grad_enabled` for updating
# the FP8 scales/inverses. So this decorator should be applied before `torch.no_grad` (so that
# it is in grad enabled part).
from thunder.executors.transformer_engine_v1ex import _is_te_linear_enabled, _get_te_wrapper_string

if TraceTag.AUGMENTED_FORWARD and _is_te_linear_enabled(import_ctx, object_ctx):
program.append(_get_te_wrapper_string())

# Disable gradients since Thunder takes care of this (for when calling torch operations)
program.append("@torch.no_grad()")
# Disable autocast since we already generated the trace with it in consideration (for when calling torch
Expand Down
Loading
Loading