Add llama.cpp to the examples
#3
by
b-a-s-e-d
- opened
README.md
CHANGED
@@ -107,6 +107,7 @@ The model can also be deployed with the following libraries:
|
|
107 |
- [`mistral-inference`](https://github.com/mistralai/mistral-inference): See [here](#mistral-inference)
|
108 |
- [`transformers`](https://github.com/huggingface/transformers): See [here](#transformers)
|
109 |
- [`LMStudio`](https://lmstudio.ai/): See [here](#lmstudio)
|
|
|
110 |
- [`ollama`](https://github.com/ollama/ollama): See [here](#ollama)
|
111 |
|
112 |
|
@@ -394,6 +395,23 @@ docker run -it --rm --pull=always \
|
|
394 |
Click “see advanced setting” on the second line.
|
395 |
In the new tab, toggle advanced to on. Set the custom model to be mistral/devstralq4_k_m and Base URL the api address we get from the last step in LM Studio. Set API Key to dummy. Click save changes.
|
396 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
397 |
|
398 |
### Ollama
|
399 |
|
|
|
107 |
- [`mistral-inference`](https://github.com/mistralai/mistral-inference): See [here](#mistral-inference)
|
108 |
- [`transformers`](https://github.com/huggingface/transformers): See [here](#transformers)
|
109 |
- [`LMStudio`](https://lmstudio.ai/): See [here](#lmstudio)
|
110 |
+
- [`llama.cpp`](https://github.com/ggml-org/llama.cpp): See [here](#llama.cpp)
|
111 |
- [`ollama`](https://github.com/ollama/ollama): See [here](#ollama)
|
112 |
|
113 |
|
|
|
395 |
Click “see advanced setting” on the second line.
|
396 |
In the new tab, toggle advanced to on. Set the custom model to be mistral/devstralq4_k_m and Base URL the api address we get from the last step in LM Studio. Set API Key to dummy. Click save changes.
|
397 |
|
398 |
+
### llama.cpp
|
399 |
+
|
400 |
+
Download the weights from huggingface:
|
401 |
+
|
402 |
+
```
|
403 |
+
pip install -U "huggingface_hub[cli]"
|
404 |
+
huggingface-cli download \
|
405 |
+
"mistralai/Devstral-Small-2505_gguf" \
|
406 |
+
--include "devstralQ4_K_M.gguf" \
|
407 |
+
--local-dir "mistralai/Devstral-Small-2505_gguf/"
|
408 |
+
```
|
409 |
+
|
410 |
+
Then run Devstral using the llama.cpp CLI.
|
411 |
+
|
412 |
+
```bash
|
413 |
+
./llama-cli -m Devstral-Small-2505_gguf/devstralQ4_K_M.gguf -cnv
|
414 |
+
```
|
415 |
|
416 |
### Ollama
|
417 |
|