Update qa_l3_test.yml #5
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| name: QA L3 Tests | |
| on: | |
| push: | |
| branches: | |
| - test-ye | |
| pull_request: | |
| branches: | |
| - main | |
| concurrency: | |
| group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}-${{ github.actor }} | |
| cancel-in-progress: true | |
| jobs: | |
| qa-l3-test: | |
| runs-on: [ self-hosted, TE_FL ] | |
| defaults: | |
| run: | |
| shell: bash | |
| container: | |
| image: localhost:5000/flagscale-cuda12.8.1-torch2.7.1-python3.10-te2.9:20260209 | |
| ports: | |
| - 80:80 | |
| options: >- | |
| --gpus all | |
| --shm-size=500g | |
| --privileged | |
| --ipc=host | |
| --ulimit memlock=-1 | |
| --ulimit stack=67108864 | |
| --ulimit nofile=65535:65535 | |
| --user root | |
| --pull always | |
| steps: | |
| - name: Checkout Code | |
| uses: actions/[email protected] | |
| with: | |
| repository: ${{ github.event.pull_request.head.repo.full_name }} | |
| ref: ${{ github.event.pull_request.head.ref }} | |
| ssh-strict: true | |
| ssh-user: git | |
| persist-credentials: true | |
| clean: true | |
| sparse-checkout-cone-mode: true | |
| fetch-tags: false | |
| show-progress: true | |
| lfs: false | |
| submodules: recursive | |
| set-safe-directory: true | |
| - name: Install dependencies and build transformer_engine | |
| # timeout-minutes: 30 | |
| env: | |
| NVTE_FRAMEWORK: pytorch | |
| TE_WITH_NCCL: 1 | |
| run: | | |
| source /opt/miniconda3/etc/profile.d/conda.sh | |
| conda activate flagscale-train | |
| # Install MPI | |
| apt update | |
| apt install -y libopenmpi-dev openmpi-bin openmpi-common | |
| apt install -y libmpich-dev mpich | |
| # Verify the MPI header file | |
| mpicxx -show | awk '{for(i=1;i<=NF;i++) if($i ~ /-I/) print substr($i,3)}' | |
| # Verify whether the MPI C++ environment is ready | |
| # 1. Verify whether the MPI C++ compiler (mpicxx) exists | |
| mpicxx --version | |
| # 2. Verify if the MPI library file exists | |
| ls /usr/lib/x86_64-linux-gnu/libmpi_cxx.so | |
| # Install dependencies | |
| pip install optree looseversion opt_einsum lightning_utilities | |
| # Clone lightning-thunder | |
| git clone --recurse-submodules https://github.com/Lightning-AI/lightning-thunder.git | |
| echo "Install transformer_engine" | |
| pip install --no-build-isolation -vvv . --no-deps | |
| # Verify installation | |
| python3 tests/pytorch/test_sanity_import.py | |
| - name: GPU Usage Check / Verification | |
| run: | | |
| source .github/workflows/scripts/gpu_check.sh | |
| wait_for_gpu | |
| - name: QA L3 Pytorch FA Versions Test | |
| id: L3_pytorch_FA_versions_test | |
| timeout-minutes: 30 | |
| env: | |
| XML_LOG_DIR: "/logs/pytorch/attention" | |
| TE_PATH: . | |
| MAX_JOBS: 32 | |
| run: | | |
| source /opt/miniconda3/etc/profile.d/conda.sh | |
| conda activate flagscale-train | |
| mkdir -p "$XML_LOG_DIR" | |
| pip3 install pytest==8.2.1 | |
| git clone https://github.com/Dao-AILab/flash-attention.git | |
| cd flash-attention/ && git checkout 27f501d && cd hopper/ && python setup.py install | |
| python_path=`python -c "import site; print(site.getsitepackages()[0])"` | |
| mkdir -p $python_path/flash_attn_3 | |
| wget -P $python_path/flash_attn_3 https://raw.githubusercontent.com/Dao-AILab/flash-attention/27f501dbe011f4371bff938fe7e09311ab3002fa/hopper/flash_attn_interface.py | |
| cd ../../ | |
| # Run tests | |
| NVTE_TORCH_COMPILE=0 python3 -m pytest -v -s --junitxml=$XML_LOG_DIR/pytest.xml $TE_PATH/tests/pytorch/attention/test_attention.py | |