Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ cxx_compiler_version:
docker_image:
- quay.io/condaforge/linux-anvil-x86_64:alma9
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-openstack-gpu-2xlarge
is_rc:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ cxx_compiler_version:
docker_image:
- quay.io/condaforge/linux-anvil-x86_64:alma9
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-openstack-gpu-2xlarge
is_rc:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ cxx_compiler_version:
docker_image:
- quay.io/condaforge/linux-anvil-x86_64:alma9
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-openstack-gpu-2xlarge
is_rc:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ cxx_compiler_version:
docker_image:
- quay.io/condaforge/linux-anvil-x86_64:alma9
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-openstack-gpu-2xlarge
is_rc:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ cxx_compiler_version:
docker_image:
- quay.io/condaforge/linux-anvil-x86_64:alma9
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-openstack-gpu-2xlarge
is_rc:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ cxx_compiler_version:
docker_image:
- quay.io/condaforge/linux-anvil-x86_64:alma9
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-openstack-gpu-2xlarge
is_rc:
Expand Down
10 changes: 10 additions & 0 deletions .ci_support/migrations/fmt121_spdlog117.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,10 @@
__migrator:
build_number: 1
commit_message: Rebuild for fmt 12.1 and spdlog 1.17
kind: version
migration_number: 1
migrator_ts: 1767674179.1402016
fmt:
- '12.1'
spdlog:
- '1.17'
10 changes: 0 additions & 10 deletions .ci_support/migrations/fmt12_spdlog116.yaml

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ cxx_compiler:
cxx_compiler_version:
- '19'
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-macos-m4-large
is_rc:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ cxx_compiler:
cxx_compiler_version:
- '19'
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-macos-m4-large
is_rc:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ cxx_compiler:
cxx_compiler_version:
- '19'
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-macos-m4-large
is_rc:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ cudnn:
cxx_compiler:
- vs2022
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-azure-windows-4xlarge
is_rc:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ cudnn:
cxx_compiler:
- vs2022
fmt:
- '12.0'
- '12.1'
github_actions_labels:
- cirun-azure-windows-4xlarge
is_rc:
Expand Down
6 changes: 5 additions & 1 deletion recipe/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -302,7 +302,11 @@ case ${PKG_NAME} in
for f in bin/* lib/* share/* include/*; do
if [[ -e "$PREFIX/$f" ]]; then
rm -rf $f
ln -sf $PREFIX/$f $PWD/$f
# do not symlink include files back
# https://github.com/conda-forge/pytorch-cpu-feedstock/issues/447#issuecomment-3712968499
if [[ ${f} != include/* ]]; then
ln -sf $PREFIX/$f $PWD/$f
fi
fi
done
popd
Expand Down
30 changes: 14 additions & 16 deletions recipe/meta.yaml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
# if you wish to build release candidate number X, append the version string with ".rcX"
{% set version = "2.10.0-rc3" %}
{% set build = 0 %}
{% set build = 1 %}

# Use a higher build number for the CUDA variant, to ensure that it's
# preferred by conda's solver, and it's preferentially
Expand Down Expand Up @@ -365,6 +365,8 @@ outputs:
- cuda-nvrtc-dev # [cuda_compiler_version != "None"]
- nvtx-c # [cuda_compiler_version != "None"]
- pybind11
# not required upstream, but useful
- pytest-timeout
imports:
- torch
- torch._C
Expand Down Expand Up @@ -439,6 +441,12 @@ outputs:
{% set skips = skips ~ " or test_addbmm or test_baddbmm or test_bmm" %} # [aarch64]
# doesn't crash, but gets different result on aarch + CUDA
{% set skips = skips ~ " or illcondition_matrix_input_should_not_crash_cpu" %} # [aarch64 and cuda_compiler_version != "None"]
# may crash spuriously
{% set skips = skips ~ " or (TestAutograd and test_profiler_seq_nr)" %} # [not win]
{% set skips = skips ~ " or (TestAutograd and test_profiler_propagation)" %} # [not win]
# tests that fail due to resource clean-up issues (non-unique temporary libraries), see
# https://github.com/conda-forge/pytorch-cpu-feedstock/pull/318#issuecomment-2620080859
{% set skips = skips ~ " or test_mutable_custom_op_fixed_layout" %} # [not win]
# minor inaccuracy on aarch64 (emulation?)
{% set skips = skips ~ " or (TestNN and test_upsampling_bfloat16)" %} # [aarch64]
# may crash spuriously
Expand Down Expand Up @@ -479,23 +487,13 @@ outputs:
# non-MKL problems
{% set skips = skips ~ " or test_gather_scatter_cpu or test_index_put2_cpu" %} # [linux and blas_impl != "mkl" and cuda_compiler_version != "None"]
# these tests are failing with low -n values
{% set skips = skips ~ " or test_base_does_not_require_grad_mode_nothing" %}
{% set skips = skips ~ " or test_base_does_not_require_grad_mode_warn" %}
{% set skips = skips ~ " or test_composite_registered_to_cpu_mode_nothing" %}
{% set skips = skips ~ " or test_base_does_not_require_grad_mode_nothing" %} # [not win]
{% set skips = skips ~ " or test_base_does_not_require_grad_mode_warn" %} # [not win]
{% set skips = skips ~ " or test_composite_registered_to_cpu_mode_nothing" %} # [not win]
# these tests are failing on windows
{% set skips = skips ~ " or (TestMkldnnCPU and test_batch_norm_2d_cpu)" %} # [win]
{% set skips = skips ~ " or (TestNN and test_Conv1d_dilated)" %} # [win]
{% set skips = skips ~ " or (TestNN and test_Conv1d_pad_same_dilated)" %} # [win]
{% set skips = skips ~ " or (TestNN and test_Conv2d_pad_same_dilated)" %} # [win]
{% set skips = skips ~ " or (TestNN and test_Conv2d_padding)" %} # [win]
{% set skips = skips ~ " or (TestNN and test_Conv2d_strided)" %} # [win]
{% set skips = skips ~ " or (TestNN and test_Conv3d_dilated)" %} # [win]
{% set skips = skips ~ " or (TestNN and test_Conv3d_dilated_strided)" %} # [win]
{% set skips = skips ~ " or (TestNN and test_Conv3d_pad_same_dilated)" %} # [win]
{% set skips = skips ~ " or (TestNN and test_Conv3d_stride)" %} # [win]
{% set skips = skips ~ " or (TestNN and test_Conv3d_stride_padding)" %} # [win]
# flaky test, fragile to GC behavior
{% set skips = skips ~ " or (TestTorch and test_tensor_cycle_via_slots)" %}
{% set skips = skips ~ " or (TestTorch and test_tensor_cycle_via_slots)" %} # [not win]
# unexpected success
{% set skips = skips ~ " or test_forward_nn_Bilinear_mps_float16" %} # [osx and arm64]
# "quantized engine NoQEngine is not supported"
Expand All @@ -514,7 +512,7 @@ outputs:
{% set jobs = "-n 1" %} # [linux64 and cuda_compiler_version != "None"]
# test only one python version on aarch because emulation is super-slow;
# disable hypothesis because it randomly yields health check errors
- pytest {{ jobs }} {{ tests }} -k "not ({{ skips }})" -m "not hypothesis" --durations=50 --disable-warnings # [not aarch64 or py==312]
- pytest {{ jobs }} {{ tests }} -k "not ({{ skips }})" -m "not hypothesis" --durations=50 --timeout=1200 --disable-warnings # [not aarch64 or py==312]

# regression test for https://github.com/conda-forge/pytorch-cpu-feedstock/issues/329, where we picked up
# duplicate `.pyc` files due to newest py-ver (3.13) in the build environment not matching the one in host;
Expand Down
Loading