|
1 | 1 | # EmbeddedLLM
|
2 | 2 |
|
3 |
| -Run local LLMs on iGPU, APU and CPU (AMD , Intel, and Qualcomm (Coming Soon)) |
| 3 | +Run local LLMs on iGPU, APU and CPU (AMD , Intel, and Qualcomm (Coming Soon)). |
4 | 4 | Easiest way to launch OpenAI API Compatible Server on Windows, Linux and MacOS
|
5 | 5 |
|
6 | 6 | | Support matrix | Supported now | Under Development | On the roadmap |
|
@@ -32,6 +32,10 @@ Easiest way to launch OpenAI API Compatible Server on Windows, Linux and MacOS
|
32 | 32 | | Phi3-mini-128k-instruct | 3.8B | 128k | [microsoft/Phi-3-mini-128k-instruct-onnx](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct-onnx) |
|
33 | 33 | | Phi3-medium-4k-instruct | 17B | 4096 | [microsoft/Phi-3-medium-4k-instruct-onnx-directml](https://huggingface.co/microsoft/Phi-3-medium-4k-instruct-onnx-directml) |
|
34 | 34 | | Phi3-medium-128k-instruct | 17B | 128k | [microsoft/Phi-3-medium-128k-instruct-onnx-directml](https://huggingface.co/microsoft/Phi-3-medium-128k-instruct-onnx-directml) |
|
| 35 | +| Openchat-3.6-8b | 8B | 8192 | [EmbeddedLLM/openchat-3.6-8b-20240522-onnx](https://huggingface.co/EmbeddedLLM/openchat-3.6-8b-20240522-onnx) | |
| 36 | +| Yi-1.5-6b-chat | 6B | 32k | [EmbeddedLLM/01-ai_Yi-1.5-6B-Chat-onnx](https://huggingface.co/EmbeddedLLM/01-ai_Yi-1.5-6B-Chat-onnx) | |
| 37 | +| Phi-3-vision-128k-instruct | | 128k | [EmbeddedLLM/Phi-3-vision-128k-instruct-onnx](https://huggingface.co/EmbeddedLLM/Phi-3-vision-128k-instruct-onnx/tree/main/onnx/cpu_and_mobile/cpu-int4-rtn-block-32-acc-level-4) | |
| 38 | + |
35 | 39 |
|
36 | 40 | ## Getting Started
|
37 | 41 |
|
@@ -87,6 +91,18 @@ options:
|
87 | 91 |
|
88 | 92 | 1. `ellm_chatbot --port 7788 --host localhost --server_port <ellm_server_port> --server_host localhost`.
|
89 | 93 |
|
| 94 | +  |
| 95 | + |
| 96 | +## Launch Model Management UI |
| 97 | +It is an interface that allows you to download and deploy OpenAI API compatible server. |
| 98 | +You can find out the disk space required to download the model in the UI. |
| 99 | + |
| 100 | +1. `ellm_modelui --port 6678` |
| 101 | + |
| 102 | +  |
| 103 | + |
| 104 | + |
| 105 | + |
90 | 106 | ## Acknowledgements
|
91 | 107 |
|
92 |
| -- Excellent open-source projects: [vLLM](https://github.com/vllm-project/vllm.git), [onnxruntime-genai](https://github.com/microsoft/onnxruntime-genai.git) and many others. |
| 108 | +- Excellent open-source projects: [vLLM](https://github.com/vllm-project/vllm.git), [onnxruntime-genai](https://github.com/microsoft/onnxruntime-genai.git) and many others. |
0 commit comments