@@ -28,6 +28,7 @@ Easiest way to launch OpenAI API Compatible Server on Windows, Linux and MacOS
2828| Llama-2-13b-chat | 13B | 4096 | [ EmbeddedLLM/llama-2-13b-chat-int4-onnx-directml] ( https://huggingface.co/EmbeddedLLM/llama-2-13b-chat-int4-onnx-directml ) |
2929| Llama-3-8b-chat | 8B | 8192 | [ EmbeddedLLM/mistral-7b-instruct-v0.3-onnx] ( https://huggingface.co/EmbeddedLLM/mistral-7b-instruct-v0.3-onnx ) |
3030| Mistral-7b-v0.3-instruct | 7B | 32768 | [ EmbeddedLLM/mistral-7b-instruct-v0.3-onnx] ( https://huggingface.co/EmbeddedLLM/mistral-7b-instruct-v0.3-onnx ) |
31+ | Phi-3-mini-4k-instruct-062024 | 3.8B | 4096 | [ EmbeddedLLM/Phi-3-mini-4k-instruct-062024-onnx] ( https://huggingface.co/EmbeddedLLM/Phi-3-mini-4k-instruct-062024-onnx/tree/main/onnx/directml/Phi-3-mini-4k-instruct-062024-int4 ) |
3132| Phi3-mini-4k-instruct | 3.8B | 4096 | [ microsoft/Phi-3-mini-4k-instruct-onnx] ( https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-onnx ) |
3233| Phi3-mini-128k-instruct | 3.8B | 128k | [ microsoft/Phi-3-mini-128k-instruct-onnx] ( https://huggingface.co/microsoft/Phi-3-mini-128k-instruct-onnx ) |
3334| Phi3-medium-4k-instruct | 17B | 4096 | [ microsoft/Phi-3-medium-4k-instruct-onnx-directml] ( https://huggingface.co/microsoft/Phi-3-medium-4k-instruct-onnx-directml ) |
@@ -65,6 +66,9 @@ Easiest way to launch OpenAI API Compatible Server on Windows, Linux and MacOS
6566 - ** CPU:** ` ELLM_TARGET_DEVICE='cpu' pip install -e .[cpu, webui] `
6667 - ** CUDA:** ` ELLM_TARGET_DEVICE='cuda' pip install -e .[cuda, webui] `
6768
69+ ** Note**
70+ 1 . If you are using Conda Environment. Install additional dependencies: ` conda install conda-forge::vs2015_runtime ` .
71+
6872### Launch OpenAI API Compatible Server
6973
7074```
@@ -102,6 +106,11 @@ You can find out the disk space required to download the model in the UI.
102106 ![ Model Management UI] ( asset/ellm_modelui.png )
103107
104108
109+ ## Compile OpenAI-API Compatible Server into Windows Executable
110+ 1 . Install ` embeddedllm ` .
111+ 2 . Install PyInstaller: ` pip install pyinstaller ` .
112+ 3 . Compile Windows Executable: ` pyinstaller .\ellm_api_server.spec ` .
113+ 4 . You can find the executable in the ` dist\ellm_api_server ` .
105114
106115## Acknowledgements
107116
0 commit comments