Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
46 commits
Select commit Hold shift + click to select a range
708dc6e
[CP][BE] Remove _AttentionContextParallel (#162539)
fegin Sep 18, 2025
6f9b4cc
Fix SEMI_STRUCTURED_SUPPORTED_BACKENDS selection on CUDA and ROCm (#1…
dnikolaev-amd Sep 18, 2025
c5e7bb0
[inductor] pdl inductor option (disabled by default) (#160928)
v0i0 Sep 17, 2025
627482a
[torch][cuda][device_limits] Library for querying device hardware lim…
valentinandrei Sep 18, 2025
98ce93d
[DTensor] Add guide for what to do about mixed torch.Tensor and DTens…
zou3519 Sep 17, 2025
cfb8aec
[FSDP2] idempotent reset_sharded_param: no-op if _local_tensor is alr…
weifengpy Sep 18, 2025
c43ccfb
[BE] Remove bottleneck (#163210)
PaliC Sep 17, 2025
14f8d86
Reland #161649, vectorize stored in cat for all dtypes (#162440)
ngimel Sep 18, 2025
e769026
[ROCm] Remove HIPBLASLT_ALLOW_TF32 from codebase (#162998)
xinyazhang Sep 18, 2025
8bc4a46
[ROCm] test_aot_inductor: Enable fp8 tests. (#163050)
jagadish-amd Sep 18, 2025
1330c63
Update Microsoft C++ Redistributable to the latest version (#161430)
saman-amd Sep 18, 2025
1f21f85
fixing graph break for namedtuple._replace (#160139)
morrison-turnansky Sep 18, 2025
3c8b905
support unbacked softmax / logsoftmax (#162216)
ColinPeppler Sep 18, 2025
4908fb5
[testing] Add test owner labels for some ao sparse tests (#163203)
clee2000 Sep 18, 2025
af8c232
[CI] reuse old whl: fix metadata file not getting version replaced (#…
clee2000 Sep 18, 2025
56893ca
Don't register wrong overload to prim decomp (#163138)
tugsbayasgalan Sep 17, 2025
1aeac30
Move prioritized text linker optimization code from setup.py to cmake…
robert-hardwick Sep 18, 2025
4b7aed8
Revert "[torch][cuda][device_limits] Library for querying device hard…
pytorchmergebot Sep 18, 2025
a81a2e5
[submodule] CUTLASS upgrade to 4.2.0 and change cutlass to cutlass_cp…
henrylhtsang Sep 18, 2025
23af32a
[WOQ] Integrate CUDA support for concat linear int8pack_mm woq optimi…
bbeckca Sep 18, 2025
e36a6fc
Massive hack to make autograd shut up about threaded PG mutations (#1…
ezyang Sep 18, 2025
80f8be9
[SymmMem] Fix put_signal + wait_until hang (#163194)
kwen2501 Sep 17, 2025
93964ed
[unit test] correct wrong input shape in test_flop_fx (#163148)
thenumberouscode Sep 18, 2025
8627454
Add local file path to inductor_output_code trace metadata (#160920)
ezyang Sep 18, 2025
62a746f
[ROCm] update ci_expected_accuracy for dynamo benchmarks (#163256)
jeffdaily Sep 18, 2025
159c214
test: ensure editable cached wrapper is respected (#160943)
ezyang Sep 18, 2025
bb7c9a2
[DTensor] Fix DTensor.mean with uneven sharding (#163241)
SherlockNoMad Sep 18, 2025
7dcb568
Turn on capture_scalar_outputs when fullgraph=True (#163121)
bobrenjc93 Sep 18, 2025
ed3438f
Turn on capture_dynamic_output_shape_ops when fullgraph=True (#163123)
bobrenjc93 Sep 18, 2025
f4eca0e
Try updating ET pin in PT/PT (#159664)
mergennachin Sep 18, 2025
4c00707
[dynamic shapes] DynamicInts prototype (#162194)
pianpwk Sep 18, 2025
096d35c
[CP] Remove the need of recording cp_dim in the global var (#162540)
fegin Sep 18, 2025
ddc56f6
[functional] Use the saved device on storage instead for device_custo…
anijain2305 Sep 18, 2025
17c1653
Deprecate Lite Interpreter (#163289)
JacobSzwejbka Sep 18, 2025
04842ac
Change DebugMode record_torchfunction default to False (#163293)
SherlockNoMad Sep 19, 2025
2dcd153
[CI] Move Windows build/tests to Python-3.10 (#162862)
malfet Sep 19, 2025
cd4303a
[CP][BE] Correct the names of some tests (#162541)
fegin Sep 18, 2025
c9b80c4
[ROCm] Bump FBGEMM commit to avoid CK errors (#162590)
pragupta Sep 19, 2025
3c9e220
Refactor PrecompileContext to be considerably more debuggable (#162740)
jamesjwu Sep 19, 2025
6e680ae
add more restriction to fusion with large accumulate reads (#163163)
xuanzhang816 Sep 17, 2025
e134bb3
Update torch-xpu-ops commit pin (#163244)
Chao1Han Sep 19, 2025
ef0ef7d
support high stream
Chao1Han Sep 15, 2025
7b9e09b
fix
Chao1Han Sep 15, 2025
14ff47f
update
Chao1Han Sep 15, 2025
0308253
revert update pin commit
Chao1Han Sep 18, 2025
c06e6f6
commit
Chao1Han Sep 19, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 2 additions & 4 deletions .ci/aarch64_linux/aarch64_ci_build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -31,8 +31,7 @@ pip install -r /pytorch/requirements.txt
pip install auditwheel==6.2.0 wheel
if [ "$DESIRED_CUDA" = "cpu" ]; then
echo "BASE_CUDA_VERSION is not set. Building cpu wheel."
#USE_PRIORITIZED_TEXT_FOR_LD for enable linker script optimization https://github.com/pytorch/pytorch/pull/121975/files
USE_PRIORITIZED_TEXT_FOR_LD=1 python /pytorch/.ci/aarch64_linux/aarch64_wheel_ci_build.py --enable-mkldnn
python /pytorch/.ci/aarch64_linux/aarch64_wheel_ci_build.py --enable-mkldnn
else
echo "BASE_CUDA_VERSION is set to: $DESIRED_CUDA"
export USE_SYSTEM_NCCL=1
Expand All @@ -46,6 +45,5 @@ else
export USE_NVIDIA_PYPI_LIBS=1
fi

#USE_PRIORITIZED_TEXT_FOR_LD for enable linker script optimization https://github.com/pytorch/pytorch/pull/121975/files
USE_PRIORITIZED_TEXT_FOR_LD=1 python /pytorch/.ci/aarch64_linux/aarch64_wheel_ci_build.py --enable-mkldnn --enable-cuda
python /pytorch/.ci/aarch64_linux/aarch64_wheel_ci_build.py --enable-mkldnn --enable-cuda
fi
2 changes: 1 addition & 1 deletion .ci/aarch64_linux/aarch64_wheel_ci_build.py
Original file line number Diff line number Diff line change
Expand Up @@ -317,7 +317,7 @@ def parse_arguments():
).decode()

print("Building PyTorch wheel")
build_vars = "CMAKE_SHARED_LINKER_FLAGS=-Wl,-z,max-page-size=0x10000 "
build_vars = ""
# MAX_JOB=5 is not required for CPU backend (see commit 465d98b)
if enable_cuda:
build_vars += "MAX_JOBS=5 "
Expand Down
2 changes: 1 addition & 1 deletion .ci/docker/ci_commit_pins/executorch.txt
Original file line number Diff line number Diff line change
@@ -1 +1 @@
56392aa978594cc155fa8af48cd949f5b5f1823a
e0dda9059d082537cee36be6c5e4fe3b18c880c0
23 changes: 14 additions & 9 deletions .ci/docker/common/install_executorch.sh
Original file line number Diff line number Diff line change
Expand Up @@ -42,22 +42,27 @@ install_pip_dependencies() {
# A workaround, ExecuTorch has moved to numpy 2.0 which is not compatible with the current
# numba and scipy version used in PyTorch CI
conda_run pip uninstall -y numba scipy
# Yaspin is needed for running CI test (get_benchmark_analysis_data.py)
pip_install yaspin==3.1.0

popd
}

setup_executorch() {
pushd executorch

export PYTHON_EXECUTABLE=python
export CMAKE_ARGS="-DEXECUTORCH_BUILD_PYBIND=ON -DEXECUTORCH_BUILD_XNNPACK=ON -DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON"
export CMAKE_ARGS="-DEXECUTORCH_BUILD_PYBIND=ON -DEXECUTORCH_BUILD_XNNPACK=ON -DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON -DEXECUTORCH_BUILD_TESTS=ON"

as_jenkins .ci/scripts/setup-linux.sh --build-tool cmake || true
popd
}

clone_executorch
install_buck2
install_conda_dependencies
install_pip_dependencies
setup_executorch
if [ $# -eq 0 ]; then
clone_executorch
install_buck2
install_conda_dependencies
install_pip_dependencies
pushd executorch
setup_executorch
popd
else
"$@"
fi
14 changes: 3 additions & 11 deletions .ci/pytorch/test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -1550,14 +1550,10 @@ test_executorch() {
install_torchvision
install_torchaudio

pushd /executorch

export PYTHON_EXECUTABLE=python
export CMAKE_ARGS="-DEXECUTORCH_BUILD_PYBIND=ON -DEXECUTORCH_BUILD_XNNPACK=ON -DEXECUTORCH_BUILD_KERNELS_QUANTIZED=ON"
INSTALL_SCRIPT="$(pwd)/.ci/docker/common/install_executorch.sh"

# NB: We need to rebuild ExecuTorch runner here because it depends on PyTorch
# from the PR
bash .ci/scripts/setup-linux.sh --build-tool cmake
pushd /executorch
"${INSTALL_SCRIPT}" setup_executorch

echo "Run ExecuTorch unit tests"
pytest -v -n auto
Expand All @@ -1571,10 +1567,6 @@ test_executorch() {

popd

# Test torchgen generated code for Executorch.
echo "Testing ExecuTorch op registration"
"$BUILD_BIN_DIR"/test_edge_op_registration

assert_git_not_dirty
}

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,12 +3,12 @@ if "%BUILD_ENVIRONMENT%"=="" (
) else (
set CONDA_PARENT_DIR=C:\Jenkins
)

set CONDA_ROOT_DIR=%CONDA_PARENT_DIR%\Miniconda3

:: Be conservative here when rolling out the new AMI with conda. This will try
:: to install conda as before if it couldn't find the conda installation. This
:: can be removed eventually after we gain enough confidence in the AMI
if not exist %CONDA_PARENT_DIR%\Miniconda3 (
if not exist %CONDA_ROOT_DIR% (
set INSTALL_FRESH_CONDA=1
)

Expand All @@ -17,10 +17,14 @@ if "%INSTALL_FRESH_CONDA%"=="1" (
if errorlevel 1 exit /b
if not errorlevel 0 exit /b

%TMP_DIR_WIN%\Miniconda3-latest-Windows-x86_64.exe /InstallationType=JustMe /RegisterPython=0 /S /AddToPath=0 /D=%CONDA_PARENT_DIR%\Miniconda3
%TMP_DIR_WIN%\Miniconda3-latest-Windows-x86_64.exe /InstallationType=JustMe /RegisterPython=0 /S /AddToPath=0 /D=%CONDA_ROOT_DIR%
if errorlevel 1 exit /b
if not errorlevel 0 exit /b
)

:: Activate conda so that we can use its commands, i.e. conda, python, pip
call %CONDA_PARENT_DIR%\Miniconda3\Scripts\activate.bat %CONDA_PARENT_DIR%\Miniconda3
call %CONDA_ROOT_DIR%\Scripts\activate.bat %CONDA_ROOT_DIR%
:: Activate conda so that we can use its commands, i.e. conda, python, pip
call conda activate py_tmp

call pip install -r requirements.txt
2 changes: 1 addition & 1 deletion .ci/pytorch/win-test-helpers/setup_pytorch_env.bat
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@ if not errorlevel 0 exit /b
:: build\torch. Rather than changing all these references, making a copy of torch folder
:: from conda to the current workspace is easier. The workspace will be cleaned up after
:: the job anyway
xcopy /s %CONDA_PARENT_DIR%\Miniconda3\Lib\site-packages\torch %TMP_DIR_WIN%\build\torch\
xcopy /s %CONDA_ROOT_DIR%\envs\py_tmp\Lib\site-packages\torch %TMP_DIR_WIN%\build\torch\

pushd .
if "%VC_VERSION%" == "" (
Expand Down
11 changes: 7 additions & 4 deletions .ci/pytorch/win-test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,13 @@ if [[ "$BUILD_ENVIRONMENT" == *cuda* ]]; then
fi

# TODO: Move both of them to Windows AMI
python -m pip install pytest-rerunfailures==10.3 pytest-cpp==2.3.0 tensorboard==2.13.0 protobuf==5.29.4 pytest-subtests==0.13.1
python -m pip install tensorboard==2.13.0 protobuf==5.29.4 pytest-subtests==0.13.1

# Copied from https://github.com/pytorch/test-infra/blob/be01a40157c36cd5a48391fdf44a7bc3ebd4c7e3/aws/ami/windows/scripts/Installers/Install-Pip-Dependencies.ps1#L16 with some adjustments
# pytest-rerunfailures==10.3 as 10.2 fails with INTERNALERROR> pluggy._manager.PluginValidationError: unknown hook 'pytest_configure_node'
# scipy from 1.6.3 to 1.10
# expecttest from 0.1.3 to 0.3.0
python -m pip install "future==0.18.2" "hypothesis==5.35.1" "expecttest==0.3.0" "librosa>=0.6.2" "scipy==1.10.1" "psutil==5.9.1" "pynvml==11.4.1" "pillow==9.2.0" "unittest-xml-reporting<=3.2.0,>=2.0.0" "pytest==7.1.3" "pytest-xdist==2.5.0" "pytest-flakefinder==1.1.0" "pytest-rerunfailures==10.3" "pytest-shard==0.1.2" "sympy==1.11.1" "xdoctest==1.0.2" "pygments==2.12.0" "opt-einsum>=3.3" "networkx==2.8.8" "mpmath==1.2.1" "pytest-cpp==2.3.0"

# Install Z3 optional dependency for Windows builds.
python -m pip install z3-solver==4.15.1.0
Expand All @@ -52,9 +58,6 @@ python -m pip install parameterized==0.8.1
# Install pulp for testing ilps under torch\distributed\_tools
python -m pip install pulp==2.9.0

# Install expecttest to merge https://github.com/pytorch/pytorch/pull/155308
python -m pip install expecttest==0.3.0

run_tests() {
# Run nvidia-smi if available
for path in '/c/Program Files/NVIDIA Corporation/NVSMI/nvidia-smi.exe' /c/Windows/System32/nvidia-smi.exe; do
Expand Down
2 changes: 1 addition & 1 deletion .github/actions/reuse-old-whl/reuse_old_whl.py
Original file line number Diff line number Diff line change
Expand Up @@ -264,7 +264,7 @@ def change_content_to_new_version(file: Union[str, Path]) -> None:
change_content_to_new_version(f"artifacts/dist/{old_stem}/torch/version.py")

for file in Path(f"artifacts/dist/{old_stem}").glob(
"*.dist-info/**",
"*.dist-info/*",
):
change_content_to_new_version(file)

Expand Down
16 changes: 14 additions & 2 deletions .github/actions/setup-win/action.yml
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,12 @@ inputs:
cuda-version:
description: which cuda version to install, 'cpu' for none
required: true
python-version:
required: false
type: string
default: "3.10"
description: |
The python version to be used. Will be 3.10 by default

runs:
using: composite
Expand Down Expand Up @@ -38,18 +44,24 @@ runs:
CONDA="C:\Jenkins\Miniconda3\condabin\conda.bat"

{
echo "CONDA=${CONDA}";
echo "CONDA_RUN=${CONDA} run --no-capture-output";
echo "CONDA_BUILD=${CONDA} run conda-build";
echo "CONDA_INSTALL=${CONDA} install";
} >> "${GITHUB_ENV}"

- name: Setup Python3
env:
PYTHON_VERSION: ${{ inputs.python-version }}
shell: bash
run: |
set +e
set -x

PYTHON3=$(${CONDA_RUN} which python3)
# Create new py_tmp env with python-version
${CONDA} create -y -n py_tmp python=${PYTHON_VERSION} intel-openmp

PYTHON3=$(${CONDA_RUN} -n py_tmp which python3)
EXIT_CODE=$?

if [[ "${EXIT_CODE}" == "0" ]]; then
Expand All @@ -62,7 +74,7 @@ runs:
# installation, which is Python 3 based. Its Python is default to Python 3. Further, there
# is also the Miniconda installation that is Python 2 based, and both can be installed if
# needed. In both cases, Python binary is just called python
PYTHON=$(${CONDA_RUN} which python)
PYTHON=$(${CONDA_RUN} -n py_tmp which python)
EXIT_CODE=$?

if [[ "${EXIT_CODE}" == "0" ]]; then
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/_win-build.yml
Original file line number Diff line number Diff line change
Expand Up @@ -151,7 +151,7 @@ jobs:
BUILD_WHEEL: 1
MAX_JOBS: 8
CUDA_VERSION: ${{ inputs.cuda-version }}
PYTHON_VERSION: "3.9"
PYTHON_VERSION: "3.10"
SCCACHE_BUCKET: "ossci-compiler-cache"
SCCACHE_S3_KEY_PREFIX: ${{ github.workflow }}
SCCACHE_REGION: us-east-1
Expand Down
3 changes: 2 additions & 1 deletion .github/workflows/_win-test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -184,7 +184,7 @@ jobs:
env:
USE_CUDA: ${{ inputs.cuda-version != 'cpu' && '1' || '0' }}
INSTALL_WINDOWS_SDK: 1
PYTHON_VERSION: 3.9
PYTHON_VERSION: "3.10"
CONTINUE_THROUGH_ERROR: ${{ steps.keep-going.outputs.keep-going }}
VERBOSE_TEST_LOGS: ${{ steps.keep-going.outputs.ci-verbose-test-logs }}
TEST_SHOWLOCALS: ${{ steps.keep-going.outputs.ci-test-showlocals }}
Expand Down Expand Up @@ -217,6 +217,7 @@ jobs:
PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: ${{ matrix.mem_leak_check && '1' || '0' }}
PYTORCH_TEST_RERUN_DISABLED_TESTS: ${{ matrix.rerun_disabled_tests && '1' || '0' }}
run: |
which python3
pushd "${PYTORCH_FINAL_PACKAGE_DIR}"
# shellcheck disable=SC2046,SC2102
python3 -mpip install $(echo *.whl)[opt-einsum,optree] optree==0.13.0
Expand Down
3 changes: 1 addition & 2 deletions .github/workflows/docker-builds.yml
Original file line number Diff line number Diff line change
Expand Up @@ -71,8 +71,7 @@ jobs:
pytorch-linux-jammy-py3-clang12-onnx,
pytorch-linux-jammy-linter,
pytorch-linux-jammy-cuda12.8-cudnn9-py3.9-linter,
# Executorch pin needs update
# pytorch-linux-jammy-py3-clang12-executorch,
pytorch-linux-jammy-py3-clang12-executorch,
pytorch-linux-jammy-py3.12-triton-cpu,
pytorch-linux-noble-riscv64-py3.12-gcc14
]
Expand Down
26 changes: 0 additions & 26 deletions .github/workflows/pull.yml
Original file line number Diff line number Diff line change
Expand Up @@ -318,32 +318,6 @@ jobs:
]}
secrets: inherit

linux-jammy-py3-clang12-executorch-build:
if: false # Docker build needs pin update
name: linux-jammy-py3-clang12-executorch
uses: ./.github/workflows/_linux-build.yml
needs: get-label-type
with:
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
build-environment: linux-jammy-py3-clang12-executorch
docker-image-name: ci-image:pytorch-linux-jammy-py3-clang12-executorch
test-matrix: |
{ include: [
{ config: "executorch", shard: 1, num_shards: 1, runner: "${{ needs.get-label-type.outputs.label-type }}linux.2xlarge" },
]}
secrets: inherit

linux-jammy-py3-clang12-executorch-test:
name: linux-jammy-py3-clang12-executorch
uses: ./.github/workflows/_linux-test.yml
needs: linux-jammy-py3-clang12-executorch-build
if: false # Has been broken for a while
with:
build-environment: linux-jammy-py3-clang12-executorch
docker-image: ${{ needs.linux-jammy-py3-clang12-executorch-build.outputs.docker-image }}
test-matrix: ${{ needs.linux-jammy-py3-clang12-executorch-build.outputs.test-matrix }}
secrets: inherit

linux-jammy-cuda12_8-py3_10-gcc9-inductor-build:
name: cuda12.8-py3.10-gcc9-sm75
uses: ./.github/workflows/_linux-build.yml
Expand Down
24 changes: 24 additions & 0 deletions .github/workflows/trunk.yml
Original file line number Diff line number Diff line change
Expand Up @@ -259,3 +259,27 @@ jobs:
docker-image: ${{ needs.verify-cachebench-cpu-build.outputs.docker-image }}
test-matrix: ${{ needs.verify-cachebench-cpu-build.outputs.test-matrix }}
secrets: inherit

linux-jammy-py3-clang12-executorch-build:
name: linux-jammy-py3-clang12-executorch
uses: ./.github/workflows/_linux-build.yml
needs: get-label-type
with:
runner_prefix: "${{ needs.get-label-type.outputs.label-type }}"
build-environment: linux-jammy-py3-clang12-executorch
docker-image-name: ci-image:pytorch-linux-jammy-py3-clang12-executorch
test-matrix: |
{ include: [
{ config: "executorch", shard: 1, num_shards: 1, runner: "${{ needs.get-label-type.outputs.label-type }}linux.2xlarge" },
]}
secrets: inherit

linux-jammy-py3-clang12-executorch-test:
name: linux-jammy-py3-clang12-executorch
uses: ./.github/workflows/_linux-test.yml
needs: linux-jammy-py3-clang12-executorch-build
with:
build-environment: linux-jammy-py3-clang12-executorch
docker-image: ${{ needs.linux-jammy-py3-clang12-executorch-build.outputs.docker-image }}
test-matrix: ${{ needs.linux-jammy-py3-clang12-executorch-build.outputs.test-matrix }}
secrets: inherit
3 changes: 3 additions & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -259,6 +259,9 @@ gen
.pytest_cache
aten/build/*

# Linker scripts for prioritized text optimization
cmake/linker_script.ld

# Bram
plsdontbreak

Expand Down
3 changes: 0 additions & 3 deletions .lintrunner.toml
Original file line number Diff line number Diff line change
Expand Up @@ -964,7 +964,6 @@ exclude_patterns = [
'test/jit/**', # should be run through test/test_jit.py
'test/ao/sparsity/**', # should be run through test/test_ao_sparsity.py
'test/fx/**', # should be run through test/test_fx.py
'test/bottleneck_test/**', # excluded by test/run_test.py
'test/package/**', # excluded by test/run_test.py
'test/distributed/argparse_util_test.py',
'test/distributed/bin/test_script.py',
Expand Down Expand Up @@ -1410,8 +1409,6 @@ exclude_patterns = [
'torch/utils/benchmark/utils/timer.py',
'torch/utils/benchmark/utils/valgrind_wrapper/__init__.py',
'torch/utils/benchmark/utils/valgrind_wrapper/timer_interface.py',
'torch/utils/bottleneck/__init__.py',
'torch/utils/bottleneck/__main__.py',
'torch/utils/bundled_inputs.py',
'torch/utils/checkpoint.py',
'torch/utils/collect_env.py',
Expand Down
Loading