IIC
/

Text Generation
Transformers
GGUF
Spanish
chat
conversational
gonzalo-santamaria-iic commited on
Commit
0e19e06
verified
1 Parent(s): 702726a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -0
README.md CHANGED
@@ -17,6 +17,32 @@ tags:
17
 
18
  This repo contains the [IIC/RigoChat-7b-v2](https://huggingface.co/IIC/RigoChat-7b-v2) model in the GGUF Format, with the original weights and quantized to different precisions.
19
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  ## How to Get Started with the Model
21
 
22
  ## Evaluation
 
17
 
18
  This repo contains the [IIC/RigoChat-7b-v2](https://huggingface.co/IIC/RigoChat-7b-v2) model in the GGUF Format, with the original weights and quantized to different precisions.
19
 
20
+ The [llama.cpp](https://github.com/ggerganov/llama.cpp) library has been used to transform the parameters into GGUF format, as well as to perform the quantizations. Specifically, the following command has been used to obtain the model in full precision:
21
+
22
+ 1. To download the weights:
23
+
24
+ ```python
25
+ from huggingface_hub import snapshot_download
26
+ import os
27
+
28
+ model_id="IIC/RigoChat-7b-v2"
29
+
30
+ os.environ["MODEL_DIR"] = snapshot_download(
31
+ repo_id=model_id,
32
+ local_dir="model",
33
+ local_dir_use_symlinks=False,
34
+ revision="main",
35
+ )
36
+ ```
37
+
38
+ 2. To transform to `FP16`:
39
+
40
+ ```shell
41
+ python ../llama.cpp/convert_hf_to_gguf.py $MODEL_DIR --outfile rigochat-7b-v2-F16.gguf --outtype f16
42
+ ```
43
+
44
+ Yo can download this weights [here](https://huggingface.co/IIC/RigoChat-7b-v2-GGUF/blob/main/rigochat-7b-v2-F16.gguf).
45
+
46
  ## How to Get Started with the Model
47
 
48
  ## Evaluation