Add llama.cpp to the examples

#3
by b-a-s-e-d - opened
Files changed (1) hide show
  1. README.md +18 -0
README.md CHANGED
@@ -107,6 +107,7 @@ The model can also be deployed with the following libraries:
107
  - [`mistral-inference`](https://github.com/mistralai/mistral-inference): See [here](#mistral-inference)
108
  - [`transformers`](https://github.com/huggingface/transformers): See [here](#transformers)
109
  - [`LMStudio`](https://lmstudio.ai/): See [here](#lmstudio)
 
110
  - [`ollama`](https://github.com/ollama/ollama): See [here](#ollama)
111
 
112
 
@@ -394,6 +395,23 @@ docker run -it --rm --pull=always \
394
  Click “see advanced setting” on the second line.
395
  In the new tab, toggle advanced to on. Set the custom model to be mistral/devstralq4_k_m and Base URL the api address we get from the last step in LM Studio. Set API Key to dummy. Click save changes.
396
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
397
 
398
  ### Ollama
399
 
 
107
  - [`mistral-inference`](https://github.com/mistralai/mistral-inference): See [here](#mistral-inference)
108
  - [`transformers`](https://github.com/huggingface/transformers): See [here](#transformers)
109
  - [`LMStudio`](https://lmstudio.ai/): See [here](#lmstudio)
110
+ - [`llama.cpp`](https://github.com/ggml-org/llama.cpp): See [here](#llama.cpp)
111
  - [`ollama`](https://github.com/ollama/ollama): See [here](#ollama)
112
 
113
 
 
395
  Click “see advanced setting” on the second line.
396
  In the new tab, toggle advanced to on. Set the custom model to be mistral/devstralq4_k_m and Base URL the api address we get from the last step in LM Studio. Set API Key to dummy. Click save changes.
397
 
398
+ ### llama.cpp
399
+
400
+ Download the weights from huggingface:
401
+
402
+ ```
403
+ pip install -U "huggingface_hub[cli]"
404
+ huggingface-cli download \
405
+ "mistralai/Devstral-Small-2505_gguf" \
406
+ --include "devstralQ4_K_M.gguf" \
407
+ --local-dir "mistralai/Devstral-Small-2505_gguf/"
408
+ ```
409
+
410
+ Then run Devstral using the llama.cpp CLI.
411
+
412
+ ```bash
413
+ ./llama-cli -m Devstral-Small-2505_gguf/devstralQ4_K_M.gguf -cnv
414
+ ```
415
 
416
  ### Ollama
417