Oleg Shulyakov commited on
Commit
c6c7289
·
1 Parent(s): 4700c13

Rename shell

Browse files
Files changed (1) hide show
  1. hf-quantize.sh → ggufy.sh +8 -8
hf-quantize.sh → ggufy.sh RENAMED
@@ -3,15 +3,15 @@
3
  # Shortcut for quantizing HF models using named parameters and short options
4
  #
5
  # Usage with long options:
6
- # ./hf-quantize.sh --model meta-llama/Llama-2-7b --quant-method Q4_K_M
7
- # ./hf-quantize.sh --model meta-llama/Llama-2-7b --quant-method Q4_K_M --use-imatrix
8
- # ./hf-quantize.sh --model meta-llama/Llama-2-7b --quant-method Q4_K_M --use-imatrix --output-filename Llama-2-7b-Q4_K_M.gguf
9
- # ./hf-quantize.sh --model meta-llama/Llama-2-7b --quant-method Q4_K_M --use-imatrix --output-filename Llama-2-7b-Q4_K_M.gguf --split-model --split-max-tensors 256 --split-max-size 4G
10
  #
11
- # ./hf-quantize.sh -m meta-llama/Llama-2-7b -q Q4_K_M
12
- # ./hf-quantize.sh -m meta-llama/Llama-2-7b -q Q4_K_M -imatrix
13
- # ./hf-quantize.sh -m meta-llama/Llama-2-7b -q Q4_K_M -imatrix -o Llama-2-7b-Q4_K_M.gguf
14
- # ./hf-quantize.sh -m meta-llama/Llama-2-7b -q Q4_K_M -imatrix -o Llama-2-7b-Q4_K_M.gguf -split --split-max-tensors 256 --split-max-size 4G
15
  #
16
 
17
  # --- Configuration ---
 
3
  # Shortcut for quantizing HF models using named parameters and short options
4
  #
5
  # Usage with long options:
6
+ # ./ggufy.sh --model meta-llama/Llama-2-7b --quant-method Q4_K_M
7
+ # ./ggufy.sh --model meta-llama/Llama-2-7b --quant-method Q4_K_M --use-imatrix
8
+ # ./ggufy.sh --model meta-llama/Llama-2-7b --quant-method Q4_K_M --use-imatrix --output-filename Llama-2-7b-Q4_K_M.gguf
9
+ # ./ggufy.sh --model meta-llama/Llama-2-7b --quant-method Q4_K_M --use-imatrix --output-filename Llama-2-7b-Q4_K_M.gguf --split-model --split-max-tensors 256 --split-max-size 4G
10
  #
11
+ # ./ggufy.sh -m meta-llama/Llama-2-7b -q Q4_K_M
12
+ # ./ggufy.sh -m meta-llama/Llama-2-7b -q Q4_K_M -imatrix
13
+ # ./ggufy.sh -m meta-llama/Llama-2-7b -q Q4_K_M -imatrix -o Llama-2-7b-Q4_K_M.gguf
14
+ # ./ggufy.sh -m meta-llama/Llama-2-7b -q Q4_K_M -imatrix -o Llama-2-7b-Q4_K_M.gguf -split --split-max-tensors 256 --split-max-size 4G
15
  #
16
 
17
  # --- Configuration ---