@@ -28,6 +28,7 @@ Easiest way to launch OpenAI API Compatible Server on Windows, Linux and MacOS
28
28
| Llama-2-13b-chat | 13B | 4096 | [ EmbeddedLLM/llama-2-13b-chat-int4-onnx-directml] ( https://huggingface.co/EmbeddedLLM/llama-2-13b-chat-int4-onnx-directml ) |
29
29
| Llama-3-8b-chat | 8B | 8192 | [ EmbeddedLLM/mistral-7b-instruct-v0.3-onnx] ( https://huggingface.co/EmbeddedLLM/mistral-7b-instruct-v0.3-onnx ) |
30
30
| Mistral-7b-v0.3-instruct | 7B | 32768 | [ EmbeddedLLM/mistral-7b-instruct-v0.3-onnx] ( https://huggingface.co/EmbeddedLLM/mistral-7b-instruct-v0.3-onnx ) |
31
+ | Phi-3-mini-4k-instruct-062024 | 3.8B | 4096 | [ EmbeddedLLM/Phi-3-mini-4k-instruct-062024-onnx] ( https://huggingface.co/EmbeddedLLM/Phi-3-mini-4k-instruct-062024-onnx/tree/main/onnx/directml/Phi-3-mini-4k-instruct-062024-int4 ) |
31
32
| Phi3-mini-4k-instruct | 3.8B | 4096 | [ microsoft/Phi-3-mini-4k-instruct-onnx] ( https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-onnx ) |
32
33
| Phi3-mini-128k-instruct | 3.8B | 128k | [ microsoft/Phi-3-mini-128k-instruct-onnx] ( https://huggingface.co/microsoft/Phi-3-mini-128k-instruct-onnx ) |
33
34
| Phi3-medium-4k-instruct | 17B | 4096 | [ microsoft/Phi-3-medium-4k-instruct-onnx-directml] ( https://huggingface.co/microsoft/Phi-3-medium-4k-instruct-onnx-directml ) |
@@ -65,6 +66,9 @@ Easiest way to launch OpenAI API Compatible Server on Windows, Linux and MacOS
65
66
- ** CPU:** ` ELLM_TARGET_DEVICE='cpu' pip install -e .[cpu, webui] `
66
67
- ** CUDA:** ` ELLM_TARGET_DEVICE='cuda' pip install -e .[cuda, webui] `
67
68
69
+ ** Note**
70
+ 1 . If you are using Conda Environment. Install additional dependencies: ` conda install conda-forge::vs2015_runtime ` .
71
+
68
72
### Launch OpenAI API Compatible Server
69
73
70
74
```
@@ -102,6 +106,11 @@ You can find out the disk space required to download the model in the UI.
102
106
![ Model Management UI] ( asset/ellm_modelui.png )
103
107
104
108
109
+ ## Compile OpenAI-API Compatible Server into Windows Executable
110
+ 1 . Install ` embeddedllm ` .
111
+ 2 . Install PyInstaller: ` pip install pyinstaller ` .
112
+ 3 . Compile Windows Executable: ` pyinstaller .\ellm_api_server.spec ` .
113
+ 4 . You can find the executable in the ` dist\ellm_api_server ` .
105
114
106
115
## Acknowledgements
107
116
0 commit comments