Update README.md
Browse files
README.md
CHANGED
@@ -17,6 +17,32 @@ tags:
|
|
17 |
|
18 |
This repo contains the [IIC/RigoChat-7b-v2](https://huggingface.co/IIC/RigoChat-7b-v2) model in the GGUF Format, with the original weights and quantized to different precisions.
|
19 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
## How to Get Started with the Model
|
21 |
|
22 |
## Evaluation
|
|
|
17 |
|
18 |
This repo contains the [IIC/RigoChat-7b-v2](https://huggingface.co/IIC/RigoChat-7b-v2) model in the GGUF Format, with the original weights and quantized to different precisions.
|
19 |
|
20 |
+
The [llama.cpp](https://github.com/ggerganov/llama.cpp) library has been used to transform the parameters into GGUF format, as well as to perform the quantizations. Specifically, the following command has been used to obtain the model in full precision:
|
21 |
+
|
22 |
+
1. To download the weights:
|
23 |
+
|
24 |
+
```python
|
25 |
+
from huggingface_hub import snapshot_download
|
26 |
+
import os
|
27 |
+
|
28 |
+
model_id="IIC/RigoChat-7b-v2"
|
29 |
+
|
30 |
+
os.environ["MODEL_DIR"] = snapshot_download(
|
31 |
+
repo_id=model_id,
|
32 |
+
local_dir="model",
|
33 |
+
local_dir_use_symlinks=False,
|
34 |
+
revision="main",
|
35 |
+
)
|
36 |
+
```
|
37 |
+
|
38 |
+
2. To transform to `FP16`:
|
39 |
+
|
40 |
+
```shell
|
41 |
+
python ../llama.cpp/convert_hf_to_gguf.py $MODEL_DIR --outfile rigochat-7b-v2-F16.gguf --outtype f16
|
42 |
+
```
|
43 |
+
|
44 |
+
Yo can download this weights [here](https://huggingface.co/IIC/RigoChat-7b-v2-GGUF/blob/main/rigochat-7b-v2-F16.gguf).
|
45 |
+
|
46 |
## How to Get Started with the Model
|
47 |
|
48 |
## Evaluation
|