Cran-May commited on
Commit
0ec5e10
·
verified ·
1 Parent(s): 10e0c15

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +27 -12
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  language:
3
  - zh
4
  - en
@@ -9,42 +10,56 @@ language:
9
  - it
10
  - ru
11
  - fi
12
- license: apache-2.0
13
  library_name: transformers
 
 
14
  tags:
15
  - mixtral
16
  - llama-cpp
17
  - gguf-my-repo
18
- pipeline_tag: text-generation
19
  inference: false
20
  ---
21
 
22
  # Cran-May/openbuddy-mistral-22b-v21.1-32k-Q4_K_M-GGUF
23
  This model was converted to GGUF format from [`OpenBuddy/openbuddy-mistral-22b-v21.1-32k`](https://huggingface.co/OpenBuddy/openbuddy-mistral-22b-v21.1-32k) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
24
  Refer to the [original model card](https://huggingface.co/OpenBuddy/openbuddy-mistral-22b-v21.1-32k) for more details on the model.
25
- ## Use with llama.cpp
26
 
27
- Install llama.cpp through brew.
 
28
 
29
  ```bash
30
- brew install ggerganov/ggerganov/llama.cpp
 
31
  ```
32
  Invoke the llama.cpp server or the CLI.
33
 
34
- CLI:
35
-
36
  ```bash
37
- llama-cli --hf-repo Cran-May/openbuddy-mistral-22b-v21.1-32k-Q4_K_M-GGUF --model openbuddy-mistral-22b-v21.1-32k.Q4_K_M.gguf -p "The meaning to life and the universe is"
38
  ```
39
 
40
- Server:
41
-
42
  ```bash
43
- llama-server --hf-repo Cran-May/openbuddy-mistral-22b-v21.1-32k-Q4_K_M-GGUF --model openbuddy-mistral-22b-v21.1-32k.Q4_K_M.gguf -c 2048
44
  ```
45
 
46
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
47
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
  ```
49
- git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m openbuddy-mistral-22b-v21.1-32k.Q4_K_M.gguf -n 128
50
  ```
 
1
  ---
2
+ base_model: OpenBuddy/openbuddy-mistral-22b-v21.1-32k
3
  language:
4
  - zh
5
  - en
 
10
  - it
11
  - ru
12
  - fi
 
13
  library_name: transformers
14
+ license: apache-2.0
15
+ pipeline_tag: text-generation
16
  tags:
17
  - mixtral
18
  - llama-cpp
19
  - gguf-my-repo
 
20
  inference: false
21
  ---
22
 
23
  # Cran-May/openbuddy-mistral-22b-v21.1-32k-Q4_K_M-GGUF
24
  This model was converted to GGUF format from [`OpenBuddy/openbuddy-mistral-22b-v21.1-32k`](https://huggingface.co/OpenBuddy/openbuddy-mistral-22b-v21.1-32k) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
25
  Refer to the [original model card](https://huggingface.co/OpenBuddy/openbuddy-mistral-22b-v21.1-32k) for more details on the model.
 
26
 
27
+ ## Use with llama.cpp
28
+ Install llama.cpp through brew (works on Mac and Linux)
29
 
30
  ```bash
31
+ brew install llama.cpp
32
+
33
  ```
34
  Invoke the llama.cpp server or the CLI.
35
 
36
+ ### CLI:
 
37
  ```bash
38
+ llama-cli --hf-repo Cran-May/openbuddy-mistral-22b-v21.1-32k-Q4_K_M-GGUF --hf-file openbuddy-mistral-22b-v21.1-32k-q4_k_m.gguf -p "The meaning to life and the universe is"
39
  ```
40
 
41
+ ### Server:
 
42
  ```bash
43
+ llama-server --hf-repo Cran-May/openbuddy-mistral-22b-v21.1-32k-Q4_K_M-GGUF --hf-file openbuddy-mistral-22b-v21.1-32k-q4_k_m.gguf -c 2048
44
  ```
45
 
46
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
47
 
48
+ Step 1: Clone llama.cpp from GitHub.
49
+ ```
50
+ git clone https://github.com/ggerganov/llama.cpp
51
+ ```
52
+
53
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
54
+ ```
55
+ cd llama.cpp && LLAMA_CURL=1 make
56
+ ```
57
+
58
+ Step 3: Run inference through the main binary.
59
+ ```
60
+ ./llama-cli --hf-repo Cran-May/openbuddy-mistral-22b-v21.1-32k-Q4_K_M-GGUF --hf-file openbuddy-mistral-22b-v21.1-32k-q4_k_m.gguf -p "The meaning to life and the universe is"
61
+ ```
62
+ or
63
  ```
64
+ ./llama-server --hf-repo Cran-May/openbuddy-mistral-22b-v21.1-32k-Q4_K_M-GGUF --hf-file openbuddy-mistral-22b-v21.1-32k-q4_k_m.gguf -c 2048
65
  ```