Update README.md
Browse files
README.md
CHANGED
@@ -3,9 +3,9 @@ license: apache-2.0
|
|
3 |
language:
|
4 |
- en
|
5 |
base_model:
|
6 |
-
- prithivMLmods/
|
7 |
pipeline_tag: text-generation
|
8 |
-
library_name:
|
9 |
tags:
|
10 |
- text-generation-inference
|
11 |
- math
|
@@ -112,6 +112,22 @@ model-index:
|
|
112 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FGalactic-Qwen-14B-Exp2
|
113 |
name: Open LLM Leaderboard
|
114 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
115 |

|
116 |
|
117 |
# **Galactic-Qwen-14B-Exp2**
|
|
|
3 |
language:
|
4 |
- en
|
5 |
base_model:
|
6 |
+
- prithivMLmods/Galactic-Qwen-14B-Exp2
|
7 |
pipeline_tag: text-generation
|
8 |
+
library_name: exllamav2
|
9 |
tags:
|
10 |
- text-generation-inference
|
11 |
- math
|
|
|
112 |
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FGalactic-Qwen-14B-Exp2
|
113 |
name: Open LLM Leaderboard
|
114 |
---
|
115 |
+
# Galactic-Qwen-14B-Exp2-exl2
|
116 |
+
Original model: [Galactic-Qwen-14B-Exp2](https://huggingface.co/prithivMLmods/Galactic-Qwen-14B-Exp2) by [prithivMLmods](https://huggingface.co/prithivMLmods)
|
117 |
+
Based on: [Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct) by [Qwen](https://huggingface.co/Qwen)
|
118 |
+
## Quants
|
119 |
+
[4bpw h6 (main)](https://huggingface.co/cgus/Galactic-Qwen-14B-Exp2-exl2/tree/main)
|
120 |
+
[4.5bpw h6](https://huggingface.co/cgus/Galactic-Qwen-14B-Exp2-exl2/tree/4.5bpw-h6)
|
121 |
+
[5bpw h6](https://huggingface.co/cgus/Galactic-Qwen-14B-Exp2-exl2/tree/5bpw-h6)
|
122 |
+
[6bpw h6](https://huggingface.co/cgus/Galactic-Qwen-14B-Exp2-exl2/tree/6bpw-h6)
|
123 |
+
[8bpw h8](https://huggingface.co/cgus/Galactic-Qwen-14B-Exp2-exl2/tree/8bpw-h8)
|
124 |
+
|
125 |
+
## Quantization notes
|
126 |
+
Made with Exllamav2 0.2.8 with default calibration dataset.
|
127 |
+
Exl2 models can be used with TabbyAPI, Text-Generation-WebUI and some other apps.
|
128 |
+
It requires to use Nvidia RTX GPU on Windows or Nvidia RTX/AMD ROCm on Linux.
|
129 |
+
Model has to fit VRAM to be used properly. If it's bigger than your GPU can handle, please use GGUF quants for llama.cpp apps instead.
|
130 |
+
# Original model card
|
131 |

|
132 |
|
133 |
# **Galactic-Qwen-14B-Exp2**
|