Skip to content

Add Dockerfile for Blackwell GPU support (RTX 50xx)#1371

Open
TroyHernandez wants to merge 1 commit intodeepbeepmeep:mainfrom
cornball-ai:main
Open

Add Dockerfile for Blackwell GPU support (RTX 50xx)#1371
TroyHernandez wants to merge 1 commit intodeepbeepmeep:mainfrom
cornball-ai:main

Conversation

@TroyHernandez
Copy link

Summary

  • Adds Dockerfile.blackwell for NVIDIA Blackwell architecture (RTX 5060 Ti, 5070, 5080, 5090)

Why this is needed

Blackwell GPUs (sm_120) require:

  • CUDA 12.8+ - CUDA 12.4 doesn't include sm_120 support
  • PyTorch 2.7.0+ - First PyTorch version with Blackwell support

The existing Dockerfile uses CUDA 12.4 and PyTorch 2.6, which won't recognize Blackwell GPUs.

Changes

Component Main Dockerfile Blackwell Dockerfile
CUDA 12.4.1 12.8.1
PyTorch 2.6.0+cu124 2.7.1+cu128
SM Arch 8.0/8.6 12.0

Testing

Tested on RTX 5060 Ti (16GB VRAM) with Profile 4:

Model Resolution Video Length Generation Time
LTX-2 19B Distilled 448x832 5.4s ~2.5 min
LTX-2 19B Distilled 832x832 5.4s ~2.5 min
VACE 1.3B 480x832 5.1s ~10 min
Hunyuan Avatar 13B 480x832 5s ~40 min (too slow)

All models load and run correctly with --profile 4.

Build instructions

docker build -f Dockerfile.blackwell -t wan2gp:blackwell .

🤖 Generated with Claude Code

Blackwell architecture (sm_120) requires:
- CUDA 12.8+ (12.4 doesn't support sm_120)
- PyTorch 2.7.0+ (first version with Blackwell support)

Key changes from main Dockerfile:
- Base image: nvidia/cuda:12.8.1-cudnn-devel-ubuntu22.04
- PyTorch: 2.7.1+cu128
- CUDA arch: 12.0 (sm_120)

Tested on RTX 5060 Ti 16GB with:
- LTX-2 19B Distilled: ~3.5 min per video
- VACE 1.3B: ~10 min per video
- Profile 4 (low VRAM mode) works correctly

Co-Authored-By: Claude Opus 4.5 <[email protected]>
@fu8765
Copy link

fu8765 commented Jan 19, 2026

Need to make it cuda 13, to make sure it's possible to take advantage of nv fp4, also need sage 2.2+ rather than 1

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants