File tree Expand file tree Collapse file tree 2 files changed +5
-3
lines changed Expand file tree Collapse file tree 2 files changed +5
-3
lines changed Original file line number Diff line number Diff line change @@ -205,7 +205,7 @@ jobs:
205205 - list-test-files
206206 if : needs.list-test-files.outputs.torch-files != '[]' || needs.list-test-files.outputs.transformers-files != '[]'
207207 container :
208- image : ${{ needs.check-vm.outputs.ip }}:5000/nvidia/cuda:${{ needs.check-vm.outputs.cuda_version }}-ubuntu22.04
208+ image : ${{ needs.check-vm.outputs.ip }}:5000/nvidia/cuda:${{ needs.check-vm.outputs.cuda_version }}-ubuntu22.04_0822
209209 options : --device /dev/dri --ipc=host --runtime=nvidia --gpus all
210210 volumes :
211211 - /dev/dri/by-path:/dev/dri/by-path
@@ -327,7 +327,7 @@ jobs:
327327 - check-vm
328328 runs-on : [ self-hosted, xeon5 ]
329329 container :
330- image : ${{ needs.check-vm.outputs.ip }}:5000/nvidia/cuda:${{ needs.check-vm.outputs.cuda_version }}-ubuntu22.04
330+ image : ${{ needs.check-vm.outputs.ip }}:5000/nvidia/cuda:${{ needs.check-vm.outputs.cuda_version }}-ubuntu22.04_0822
331331 options : --device /dev/dri --ipc=host --runtime=nvidia --gpus all
332332 volumes :
333333 - /dev/dri/by-path:/dev/dri/by-path
Original file line number Diff line number Diff line change 88from gptqmodel .nn_modules .qlinear .exllama import ExllamaQuantLinear # noqa: E402
99from gptqmodel .nn_modules .qlinear .exllama_eora import ExllamaEoraQuantLinear
1010from gptqmodel .nn_modules .qlinear .exllamav2 import ExllamaV2QuantLinear # noqa: E402
11- from gptqmodel .nn_modules .qlinear .ipex import IPEXQuantLinear # noqa: E402
11+ from gptqmodel .nn_modules .qlinear .ipex import IPEXQuantLinear , HAS_IPEX # noqa: E402
1212from gptqmodel .nn_modules .qlinear .marlin import MarlinQuantLinear # noqa: E402
1313from gptqmodel .nn_modules .qlinear .torch import TorchQuantLinear # noqa: E402
1414from gptqmodel .nn_modules .qlinear .tritonv2 import TritonV2QuantLinear # noqa: E402
@@ -59,6 +59,8 @@ def setUpClass(cls):
5959 ]
6060 )
6161 def test_post_init (self , backend : BACKEND , equal : Dict [str , bool ]):
62+ if backend == BACKEND .IPEX and not HAS_IPEX :
63+ self .skipTest ("IPEX is not available" )
6264 model = GPTQModel .load (self .model_id , backend = backend , device_map = "auto" )
6365 model = convert_gptq_v2_to_v1_format (model , model .quantize_config , self .QLINEAR_DICT [backend ])
6466
You can’t perform that action at this time.
0 commit comments