Skip to content

Commit bafafeb

Browse files
authored
[FEAT] Add Ipex-llm support for Intel CPU, iGPU and GPU (#7)
# Added * Documentations: * README.md * docs\ipex_models.md * docs\onnxruntime_models.md * Ipex-llm support --------- Co-authored-by: tjtanaa <[email protected]>
1 parent 8b69018 commit bafafeb

20 files changed

+1342
-594
lines changed

.gitignore

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,4 +10,5 @@ test_phi3*
1010
scripts/*.ps1
1111
scripts/*.sh
1212
**/dist
13-
**/build
13+
**/build
14+
*.log

README.md

Lines changed: 74 additions & 51 deletions
Original file line numberDiff line numberDiff line change
@@ -1,14 +1,13 @@
11
# EmbeddedLLM
22

3-
Run local LLMs on iGPU, APU and CPU (AMD , Intel, and Qualcomm (Coming Soon)).
4-
Easiest way to launch OpenAI API Compatible Server on Windows, Linux and MacOS
3+
Run local LLMs on iGPU, APU and CPU (AMD , Intel, and Qualcomm (Coming Soon)). Easiest way to launch OpenAI API Compatible Server on Windows, Linux and MacOS
54

65
| Support matrix | Supported now | Under Development | On the roadmap |
76
| --------------------- | --------------------------------------------------- | ----------------- | -------------- |
87
| Model architectures | Gemma <br/> Llama \* <br/> Mistral + <br/>Phi <br/> | | |
98
| Platform | Linux <br/> Windows | | |
109
| Architecture | x86 <br/> x64 <br/> | Arm64 | |
11-
| Hardware Acceleration | CUDA<br/>DirectML<br/> | QNN <br/> ROCm | OpenVINO |
10+
| Hardware Acceleration | CUDA<br/>DirectML<br/>IpexLLM | QNN <br/> ROCm | OpenVINO |
1211

1312
\* The Llama model architecture supports similar model families such as CodeLlama, Vicuna, Yi, and more.
1413

@@ -19,6 +18,19 @@ Easiest way to launch OpenAI API Compatible Server on Windows, Linux and MacOS
1918
- [2024/06] Support Phi-3 (mini, small, medium), Phi-3-Vision-Mini, Llama-2, Llama-3, Gemma (v1), Mistral v0.3, Starling-LM, Yi-1.5.
2019
- [2024/06] Support vision/chat inference on iGPU, APU, CPU and CUDA.
2120

21+
## Table Content
22+
23+
- [Supported Models](#supported-models-quick-start)
24+
- [Onnxruntime Models](./docs/model/onnxruntime_models.md)
25+
- [Ipex-LLM Models](./docs/model/ipex_models.md)
26+
- [Getting Started](#getting-started)
27+
- [Installation From Source](#installation)
28+
- [Launch OpenAI API Compatible Server](#launch-openai-api-compatible-server)
29+
- [Launch Chatbot Web UI](#launch-chatbot-web-ui)
30+
- [Launch Model Management UI](#launch-model-management-ui)
31+
- [Compile OpenAI-API Compatible Server into Windows Executable](#compile-openai-api-compatible-server-into-windows-executable)
32+
- [Acknowledgements](#acknowledgements)
33+
2234
## Supported Models (Quick Start)
2335

2436
| Models | Parameters | Context Length | Link |
@@ -35,83 +47,94 @@ Easiest way to launch OpenAI API Compatible Server on Windows, Linux and MacOS
3547
| Phi3-medium-128k-instruct | 17B | 128k | [microsoft/Phi-3-medium-128k-instruct-onnx-directml](https://huggingface.co/microsoft/Phi-3-medium-128k-instruct-onnx-directml) |
3648
| Openchat-3.6-8b | 8B | 8192 | [EmbeddedLLM/openchat-3.6-8b-20240522-onnx](https://huggingface.co/EmbeddedLLM/openchat-3.6-8b-20240522-onnx) |
3749
| Yi-1.5-6b-chat | 6B | 32k | [EmbeddedLLM/01-ai_Yi-1.5-6B-Chat-onnx](https://huggingface.co/EmbeddedLLM/01-ai_Yi-1.5-6B-Chat-onnx) |
38-
| Phi-3-vision-128k-instruct | | 128k | [EmbeddedLLM/Phi-3-vision-128k-instruct-onnx](https://huggingface.co/EmbeddedLLM/Phi-3-vision-128k-instruct-onnx/tree/main/onnx/cpu_and_mobile/cpu-int4-rtn-block-32-acc-level-4) |
39-
50+
| Phi-3-vision-128k-instruct | | 128k | [EmbeddedLLM/Phi-3-vision-128k-instruct-onnx](https://huggingface.co/EmbeddedLLM/Phi-3-vision-128k-instruct-onnx/tree/main/onnx/cpu_and_mobile/cpu-int4-rtn-block-32-acc-level-4) |
4051

4152
## Getting Started
4253

4354
### Installation
4455

4556
#### From Source
4657

47-
**Windows**
58+
- **Windows**
59+
60+
1. Custom Setup:
61+
62+
- **XPU**: Requires anaconda environment. `conda create -n ellm python=3.10 libuv; conda activate llm`.
63+
- **DirectML**: If you are using Conda Environment. Install additional dependencies: `conda install conda-forge::vs2015_runtime`.
64+
65+
2. Install embeddedllm package. `$env:ELLM_TARGET_DEVICE='directml'; pip install -e .`. Note: currently support `cpu`, `directml` and `cuda`.
4866

49-
1. Install embeddedllm package. `$env:ELLM_TARGET_DEVICE='directml'; pip install -e .`. Note: currently support `cpu`, `directml` and `cuda`.
50-
- **DirectML:** `$env:ELLM_TARGET_DEVICE='directml'; pip install -e .[directml]`
51-
- **CPU:** `$env:ELLM_TARGET_DEVICE='cpu'; pip install -e .[cpu]`
52-
- **CUDA:** `$env:ELLM_TARGET_DEVICE='cuda'; pip install -e .[cuda]`
53-
- **With Web UI**:
54-
- **DirectML:** `$env:ELLM_TARGET_DEVICE='directml'; pip install -e .[directml, webui]`
55-
- **CPU:** `$env:ELLM_TARGET_DEVICE='cpu'; pip install -e .[cpu, webui]`
56-
- **CUDA:** `$env:ELLM_TARGET_DEVICE='cuda'; pip install -e .[cuda, webui]`
67+
- **DirectML:** `$env:ELLM_TARGET_DEVICE='directml'; pip install -e .[directml]`
68+
- **CPU:** `$env:ELLM_TARGET_DEVICE='cpu'; pip install -e .[cpu]`
69+
- **CUDA:** `$env:ELLM_TARGET_DEVICE='cuda'; pip install -e .[cuda]`
70+
- **XPU:** `$env:ELLM_TARGET_DEVICE='xpu'; pip install -e .[xpu]`
71+
- **With Web UI**:
72+
- **DirectML:** `$env:ELLM_TARGET_DEVICE='directml'; pip install -e .[directml,webui]`
73+
- **CPU:** `$env:ELLM_TARGET_DEVICE='cpu'; pip install -e .[cpu,webui]`
74+
- **CUDA:** `$env:ELLM_TARGET_DEVICE='cuda'; pip install -e .[cuda,webui]`
75+
- **XPU:** `$env:ELLM_TARGET_DEVICE='xpu'; pip install -e .[xpu,webui]`
5776

58-
**Linux**
77+
- **Linux**
5978

60-
1. Install embeddedllm package. `ELLM_TARGET_DEVICE='directml' pip install -e .`. Note: currently support `cpu`, `directml` and `cuda`.
61-
- **DirectML:** `ELLM_TARGET_DEVICE='directml' pip install -e .[directml]`
62-
- **CPU:** `ELLM_TARGET_DEVICE='cpu' pip install -e .[cpu]`
63-
- **CUDA:** `ELLM_TARGET_DEVICE='cuda' pip install -e .[cuda]`
64-
- **With Web UI**:
65-
- **DirectML:** `ELLM_TARGET_DEVICE='directml' pip install -e .[directml, webui]`
66-
- **CPU:** `ELLM_TARGET_DEVICE='cpu' pip install -e .[cpu, webui]`
67-
- **CUDA:** `ELLM_TARGET_DEVICE='cuda' pip install -e .[cuda, webui]`
79+
1. Custom Setup:
6880

69-
**Note**
70-
1. If you are using Conda Environment. Install additional dependencies: `conda install conda-forge::vs2015_runtime`.
81+
- **XPU**: Requires anaconda environment. `conda create -n ellm python=3.10 libuv; conda activate llm`.
82+
- **DirectML**: If you are using Conda Environment. Install additional dependencies: `conda install conda-forge::vs2015_runtime`.
83+
84+
2. Install embeddedllm package. `ELLM_TARGET_DEVICE='directml' pip install -e .`. Note: currently support `cpu`, `directml` and `cuda`.
85+
86+
- **DirectML:** `ELLM_TARGET_DEVICE='directml' pip install -e .[directml]`
87+
- **CPU:** `ELLM_TARGET_DEVICE='cpu' pip install -e .[cpu]`
88+
- **CUDA:** `ELLM_TARGET_DEVICE='cuda' pip install -e .[cuda]`
89+
- **XPU:** `ELLM_TARGET_DEVICE='xpu' pip install -e .[xpu]`
90+
- **With Web UI**:
91+
- **DirectML:** `ELLM_TARGET_DEVICE='directml' pip install -e .[directml,webui]`
92+
- **CPU:** `ELLM_TARGET_DEVICE='cpu' pip install -e .[cpu,webui]`
93+
- **CUDA:** `ELLM_TARGET_DEVICE='cuda' pip install -e .[cuda,webui]`
94+
- **XPU:** `ELLM_TARGET_DEVICE='xpu' pip install -e .[xpu,webui]`
7195

7296
### Launch OpenAI API Compatible Server
7397

74-
```
75-
usage: ellm_server.exe [-h] [--port int] [--host str] [--response_role str] [--uvicorn_log_level str]
76-
[--served_model_name str] [--model_path str] [--vision bool]
98+
1. Custom Setup:
99+
100+
- **Ipex**
77101

78-
options:
79-
-h, --help show this help message and exit
80-
--port int Server port. (default: 6979)
81-
--host str Server host. (default: 0.0.0.0)
82-
--response_role str Server response role. (default: assistant)
83-
--uvicorn_log_level str
84-
Uvicorn logging level. `debug`, `info`, `trace`, `warning`, `critical` (default: info)
85-
--served_model_name str
86-
Model name. (default: phi3-mini-int4)
87-
--model_path str Path to model weights. (required)
88-
--vision bool Enable vision capability, only if model supports vision input. (default: False)
89-
```
102+
- For **Intel iGPU**:
90103

91-
1. `ellm_server --model_path <path/to/model/weight>`.
92-
2. Example code to connect to the api server can be found in `scripts/python`.
104+
```cmd
105+
set SYCL_CACHE_PERSISTENT=1
106+
set BIGDL_LLM_XMX_DISABLED=1
107+
```
93108
94-
## Launch Chatbot Web UI
109+
- For **Intel Arc™ A-Series Graphics**:
110+
```cmd
111+
set SYCL_CACHE_PERSISTENT=1
112+
```
95113
96-
1. `ellm_chatbot --port 7788 --host localhost --server_port <ellm_server_port> --server_host localhost`.
114+
2. `ellm_server --model_path <path/to/model/weight>`.
115+
3. Example code to connect to the api server can be found in `scripts/python`. **Note:** To find out more of the supported arguments. `ellm_server --help`.
97116
98-
![Chatbot Web UI](asset/ellm_chatbot_vid.webp)
117+
### Launch Chatbot Web UI
99118
100-
## Launch Model Management UI
101-
It is an interface that allows you to download and deploy OpenAI API compatible server.
102-
You can find out the disk space required to download the model in the UI.
119+
1. `ellm_chatbot --port 7788 --host localhost --server_port <ellm_server_port> --server_host localhost`. **Note:** To find out more of the supported arguments. `ellm_chatbot --help`.
103120
104-
1. `ellm_modelui --port 6678`
121+
![Chatbot Web UI](asset/ellm_chatbot_vid.webp)
105122
106-
![Model Management UI](asset/ellm_modelui.png)
123+
### Launch Model Management UI
107124
125+
It is an interface that allows you to download and deploy OpenAI API compatible server. You can find out the disk space required to download the model in the UI.
126+
127+
1. `ellm_modelui --port 6678`. **Note:** To find out more of the supported arguments. `ellm_modelui --help`.
128+
129+
![Model Management UI](asset/ellm_modelui.png)
108130
109131
## Compile OpenAI-API Compatible Server into Windows Executable
132+
110133
1. Install `embeddedllm`.
111134
2. Install PyInstaller: `pip install pyinstaller`.
112135
3. Compile Windows Executable: `pyinstaller .\ellm_api_server.spec`.
113136
4. You can find the executable in the `dist\ellm_api_server`.
114137
115138
## Acknowledgements
116139
117-
- Excellent open-source projects: [vLLM](https://github.com/vllm-project/vllm.git), [onnxruntime-genai](https://github.com/microsoft/onnxruntime-genai.git) and many others.
140+
- Excellent open-source projects: [vLLM](https://github.com/vllm-project/vllm.git), [onnxruntime-genai](https://github.com/microsoft/onnxruntime-genai.git), [Ipex-LLM](https://github.com/intel-analytics/ipex-llm/tree/main) and many others.

docs/model/ipex_models.md

Lines changed: 65 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,65 @@
1+
# Model Powered by Ipex-LLM
2+
3+
## Verified Models
4+
| Model | Model Link |
5+
| --- | --- |
6+
| Phi-3 | [link](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) |
7+
8+
## Supported Models by Ipex-LLM
9+
10+
| Model | Model Link |
11+
| --- | --- |
12+
| LLaMA _(such as Vicuna, Guanaco, Koala, Baize, WizardLM, etc.)_ | |
13+
| LLaMA 2 | [link1](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf), [link2](https://huggingface.co/meta-llama/Llama-2-13b-chat-hf) |
14+
| LLaMA 3 | [link](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) |
15+
| ChatGLM | |
16+
| ChatGLM2 | [link](https://huggingface.co/THUDM/chatglm2-6b) |
17+
| ChatGLM3 | [link](https://huggingface.co/THUDM/chatglm3-6b) |
18+
| GLM-4 | [link](https://huggingface.co/THUDM/glm-4-9b-chat) |
19+
| Mistral | [link](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) |
20+
| Mixtral | [link](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) |
21+
| Falcon | [link](https://huggingface.co/tiiuae/falcon-7b-instruct) |
22+
| MPT | [link](https://huggingface.co/mosaicml/mpt-7b-chat) |
23+
| Dolly-v1 | [link](https://huggingface.co/databricks/dolly-v1-6b) |
24+
| Dolly-v2 | [link](https://huggingface.co/databricks/dolly-v2-12b) |
25+
| Replit Code | [link](https://huggingface.co/replit/replit-code-v1-3b) |
26+
| RedPajama | [link](https://huggingface.co/togethercomputer/RedPajama-INCITE-7B-Chat) |
27+
| Phoenix | [link](https://huggingface.co/FreedomIntelligence/phoenix-inst-chat-7b) |
28+
| StarCoder | [link](https://huggingface.co/bigcode/starcoder) |
29+
| Baichuan | [link](https://huggingface.co/baichuan-inc/Baichuan-13B-Chat) |
30+
| Baichuan2 | [link](https://huggingface.co/baichuan-inc/Baichuan2-13B-Chat) |
31+
| InternLM | [link](https://huggingface.co/internlm/internlm-chat-7b) |
32+
| InternLM2 | [link](https://huggingface.co/internlm/internlm2-chat-7b) |
33+
| Qwen | [link](https://huggingface.co/Qwen/Qwen-7B-Chat) |
34+
| Qwen1.5 | [link](https://huggingface.co/Qwen/Qwen1.5-7B-Chat) |
35+
| Qwen2 | [link](https://huggingface.co/Qwen/Qwen2-7B-Instruct) |
36+
| Aquila | [link](https://huggingface.co/BAAI/AquilaChat-7B) |
37+
| Aquila2 | [link](https://huggingface.co/BAAI/AquilaChat2-7B) |
38+
| Phi-1_5 | [link](https://huggingface.co/microsoft/phi-1_5) |
39+
| Flan-t5 | [link](https://huggingface.co/google/flan-t5-xxl) |
40+
| CodeLlama | [link](https://huggingface.co/codellama/CodeLlama-7b-hf) |
41+
| Skywork | [link](https://huggingface.co/Skywork/Skywork-13B-base) |
42+
| InternLM-XComposer | [link](https://huggingface.co/internlm/internlm-xcomposer-vl-7b) |
43+
| CodeShell | [link](https://huggingface.co/WisdomShell/CodeShell-7B) |
44+
| Yi | [link](https://huggingface.co/01-ai/Yi-6B) |
45+
| BlueLM | [link](https://huggingface.co/vivo-ai/BlueLM-7B-Chat) |
46+
| Mamba | [link1](https://huggingface.co/state-spaces/mamba-1.4b), [link2](https://huggingface.co/state-spaces/mamba-2.8b) |
47+
| SOLAR | [link](https://huggingface.co/upstage/SOLAR-10.7B-Instruct-v1.0) |
48+
| Phixtral | [link](https://huggingface.co/mlabonne/phixtral-4x2_8) |
49+
| RWKV4 | |
50+
| RWKV5 | |
51+
| DeepSeek-MoE | [link](https://huggingface.co/deepseek-ai/deepseek-moe-16b-chat) |
52+
| Ziya-Coding-34B-v1.0 | [link](https://huggingface.co/IDEA-CCNL/Ziya-Coding-34B-v1.0) |
53+
| Phi-2 | [link](https://huggingface.co/microsoft/phi-2) |
54+
| Phi-3 | [link](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) |
55+
| Yuan2 | [link](https://huggingface.co/IEITYuan/Yuan2-2B-hf) |
56+
| Gemma | [link1](https://huggingface.co/google/gemma-2b-it), [link2](https://huggingface.co/google/gemma-7b-it) |
57+
| DeciLM-7B | [link](https://huggingface.co/Deci/DeciLM-7B-instruct) |
58+
| Deepseek | [link](phttps://huggingface.co/deepseek-ai/deepseek-coder-6.7b-instruct) |
59+
| StableLM | [link](https://huggingface.co/stabilityai/stablelm-zephyr-3b) |
60+
| CodeGemma | [link](https://huggingface.co/google/codegemma-7b-it) |
61+
| Command-R/cohere | [link](https://huggingface.co/CohereForAI/c4ai-command-r-v01) |
62+
| CodeGeeX2 | [link](https://huggingface.co/THUDM/codegeex2-6b) |
63+
| MiniCPM | [link](https://huggingface.co/openbmb/MiniCPM-2B-sft-bf16) |
64+
65+
Resources from: https://github.com/intel-analytics/ipex-llm/

docs/model/onnxruntime_models.md

Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,19 @@
1+
# Model Powered by Onnxruntime GenAI
2+
3+
## Supported Models
4+
5+
| Models | Parameters | Context Length | Link |
6+
| --- | --- | --- | --- |
7+
| Gemma-2b-Instruct v1 | 2B | 8192 | [EmbeddedLLM/gemma-2b-it-onnx](https://huggingface.co/EmbeddedLLM/gemma-2b-it-onnx) |
8+
| Llama-2-7b-chat | 7B | 4096 | [EmbeddedLLM/llama-2-7b-chat-int4-onnx-directml](https://huggingface.co/EmbeddedLLM/llama-2-7b-chat-int4-onnx-directml) |
9+
| Llama-2-13b-chat | 13B | 4096 | [EmbeddedLLM/llama-2-13b-chat-int4-onnx-directml](https://huggingface.co/EmbeddedLLM/llama-2-13b-chat-int4-onnx-directml) |
10+
| Llama-3-8b-chat | 8B | 8192 | [EmbeddedLLM/mistral-7b-instruct-v0.3-onnx](https://huggingface.co/EmbeddedLLM/mistral-7b-instruct-v0.3-onnx) |
11+
| Mistral-7b-v0.3-instruct | 7B | 32768 | [EmbeddedLLM/mistral-7b-instruct-v0.3-onnx](https://huggingface.co/EmbeddedLLM/mistral-7b-instruct-v0.3-onnx) |
12+
| Phi-3-mini-4k-instruct-062024 | 3.8B | 4096 | [EmbeddedLLM/Phi-3-mini-4k-instruct-062024-onnx](https://huggingface.co/EmbeddedLLM/Phi-3-mini-4k-instruct-062024-onnx/tree/main/onnx/directml/Phi-3-mini-4k-instruct-062024-int4) |
13+
| Phi3-mini-4k-instruct | 3.8B | 4096 | [microsoft/Phi-3-mini-4k-instruct-onnx](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-onnx) |
14+
| Phi3-mini-128k-instruct | 3.8B | 128k | [microsoft/Phi-3-mini-128k-instruct-onnx](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct-onnx) |
15+
| Phi3-medium-4k-instruct | 17B | 4096 | [microsoft/Phi-3-medium-4k-instruct-onnx-directml](https://huggingface.co/microsoft/Phi-3-medium-4k-instruct-onnx-directml) |
16+
| Phi3-medium-128k-instruct | 17B | 128k | [microsoft/Phi-3-medium-128k-instruct-onnx-directml](https://huggingface.co/microsoft/Phi-3-medium-128k-instruct-onnx-directml) |
17+
| Openchat-3.6-8b | 8B | 8192 | [EmbeddedLLM/openchat-3.6-8b-20240522-onnx](https://huggingface.co/EmbeddedLLM/openchat-3.6-8b-20240522-onnx) |
18+
| Yi-1.5-6b-chat | 6B | 32k | [EmbeddedLLM/01-ai_Yi-1.5-6B-Chat-onnx](https://huggingface.co/EmbeddedLLM/01-ai_Yi-1.5-6B-Chat-onnx) |
19+
| Phi-3-vision-128k-instruct | | 128k | [EmbeddedLLM/Phi-3-vision-128k-instruct-onnx](https://huggingface.co/EmbeddedLLM/Phi-3-vision-128k-instruct-onnx/tree/main/onnx/cpu_and_mobile/cpu-int4-rtn-block-32-acc-level-4) |

requirements-build.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# Should be mirrored in pyproject.toml
22
packaging
33
setuptools>=49.4.0
4-
torch==2.3.1
4+
torch
55
wheel

requirements-common.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,4 +13,4 @@ transformers
1313
uvicorn
1414
filetype~=1.2.0
1515
Pillow~=10.3.0
16-
torchvision~=0.18.1
16+
torchvision

requirements-cpu.txt

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,2 +1,4 @@
1+
torch==2.3.1
2+
torchvision~=0.18.1
13
onnxruntime
24
onnxruntime-genai==0.3.0rc2

requirements-cuda.txt

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,2 +1,4 @@
1+
torch==2.3.1
2+
torchvision~=0.18.1
13
onnxruntime-gpu~=1.18.0
24
onnxruntime-genai-cuda~=0.3.0rc2

requirements-directml.txt

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,2 +1,4 @@
1+
torch==2.3.1
2+
torchvision~=0.18.1
13
onnxruntime-directml~=1.18.0
24
onnxruntime-genai-directml~=0.3.0

requirements-xpu.txt

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,4 @@
1+
torch==2.1.0
2+
torchvision
3+
trl
4+
transformers~=4.42.3

0 commit comments

Comments
 (0)