cgus commited on
Commit
29f62cb
·
verified ·
1 Parent(s): 17eb406

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -2
README.md CHANGED
@@ -3,9 +3,9 @@ license: apache-2.0
3
  language:
4
  - en
5
  base_model:
6
- - prithivMLmods/Lacerta-Opus-14B-Elite8
7
  pipeline_tag: text-generation
8
- library_name: transformers
9
  tags:
10
  - text-generation-inference
11
  - math
@@ -112,6 +112,22 @@ model-index:
112
  url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FGalactic-Qwen-14B-Exp2
113
  name: Open LLM Leaderboard
114
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
115
  ![gsdfgsdfrgs.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/pWwijAiXQlRWE93wfvCcb.png)
116
 
117
  # **Galactic-Qwen-14B-Exp2**
 
3
  language:
4
  - en
5
  base_model:
6
+ - prithivMLmods/Galactic-Qwen-14B-Exp2
7
  pipeline_tag: text-generation
8
+ library_name: exllamav2
9
  tags:
10
  - text-generation-inference
11
  - math
 
112
  url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FGalactic-Qwen-14B-Exp2
113
  name: Open LLM Leaderboard
114
  ---
115
+ # Galactic-Qwen-14B-Exp2-exl2
116
+ Original model: [Galactic-Qwen-14B-Exp2](https://huggingface.co/prithivMLmods/Galactic-Qwen-14B-Exp2) by [prithivMLmods](https://huggingface.co/prithivMLmods)
117
+ Based on: [Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct) by [Qwen](https://huggingface.co/Qwen)
118
+ ## Quants
119
+ [4bpw h6 (main)](https://huggingface.co/cgus/Galactic-Qwen-14B-Exp2-exl2/tree/main)
120
+ [4.5bpw h6](https://huggingface.co/cgus/Galactic-Qwen-14B-Exp2-exl2/tree/4.5bpw-h6)
121
+ [5bpw h6](https://huggingface.co/cgus/Galactic-Qwen-14B-Exp2-exl2/tree/5bpw-h6)
122
+ [6bpw h6](https://huggingface.co/cgus/Galactic-Qwen-14B-Exp2-exl2/tree/6bpw-h6)
123
+ [8bpw h8](https://huggingface.co/cgus/Galactic-Qwen-14B-Exp2-exl2/tree/8bpw-h8)
124
+
125
+ ## Quantization notes
126
+ Made with Exllamav2 0.2.8 with default calibration dataset.
127
+ Exl2 models can be used with TabbyAPI, Text-Generation-WebUI and some other apps.
128
+ It requires to use Nvidia RTX GPU on Windows or Nvidia RTX/AMD ROCm on Linux.
129
+ Model has to fit VRAM to be used properly. If it's bigger than your GPU can handle, please use GGUF quants for llama.cpp apps instead.
130
+ # Original model card
131
  ![gsdfgsdfrgs.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/pWwijAiXQlRWE93wfvCcb.png)
132
 
133
  # **Galactic-Qwen-14B-Exp2**