Nondzu commited on
Commit
54ad099
·
verified ·
1 Parent(s): fa44227

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +61 -3
README.md CHANGED
@@ -1,3 +1,61 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - pl
5
+ base_model:
6
+ - CYFRAGOVPL/PLLuM-8x7B-chat
7
+ ---
8
+ # PLLuM-8x7B-chat GGUF Quantizations by Nondzu
9
+
10
+ DISCLAIMER: This is state of the art quantized model. I am not the author of the original model. I am only hosting the quantized models. I do not take any responsibility for the models.
11
+
12
+ This repository contains GGUF quantized versions of the [PLLuM-8x7B-chat](https://huggingface.co/CYFRAGOVPL/PLLuM-8x7B-chat) model. All quantizations were performed using the [llama.cpp](https://github.com/ggerganov/llama.cpp) (release [b4765](https://github.com/ggml-org/llama.cpp/releases/tag/b4765)). These quantized models can be run in [LM Studio](https://lmstudio.ai/) or any other llama.cpp–based project.
13
+
14
+ ## Prompt Format
15
+
16
+ Use the following prompt structure:
17
+ ```
18
+ ???
19
+ ```
20
+
21
+ ## Available Files
22
+
23
+ Below is a list of available quantized model files along with their quantization type, file size, and a short description.
24
+
25
+ | Filename | Quant Type | File Size | Description |
26
+ | ------------------------------------------------------------------------------------- | ---------- | --------- | --------------------------------------------------------------------------------------------- |
27
+ | [PLLuM-8x7B-chat-Q2_K.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q2_K | 17 GB | Very low quality but surprisingly usable. |
28
+ | [PLLuM-8x7B-chat-Q3_K.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q3_K | 21 GB | Low quality, suitable for setups with very limited RAM. |
29
+ | [PLLuM-8x7B-chat-Q3_K_L.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q3_K_L | 23 GB | High quality; recommended for quality-focused usage. |
30
+ | [PLLuM-8x7B-chat-Q3_K_M.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q3_K_M | 21 GB | Very high quality, near perfect output – recommended. |
31
+ | [PLLuM-8x7B-chat-Q3_K_S.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q3_K_S | 20 GB | Moderate quality with improved space efficiency. |
32
+ | [PLLuM-8x7B-chat-Q4_K.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q4_K | 27 GB | Good quality for standard use. |
33
+ | [PLLuM-8x7B-chat-Q4_K_M.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q4_K_M | 27 GB | Default quality for most use cases – recommended. |
34
+ | [PLLuM-8x7B-chat-Q4_K_S.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q4_K_S | 25 GB | Slightly lower quality with enhanced space savings – recommended when size is a priority. |
35
+ | [PLLuM-8x7B-chat-Q5_0.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q5_0 | 31 GB | Extremely high quality – the maximum quant available. |
36
+ | [PLLuM-8x7B-chat-Q5_K.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q5_K | 31 GB | Very high quality – recommended for demanding use cases. |
37
+ | [PLLuM-8x7B-chat-Q5_K_M.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q5_K_M | 31 GB | High quality – recommended. |
38
+ | [PLLuM-8x7B-chat-Q5_K_S.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q5_K_S | 31 GB | High quality, offered as an alternative with minimal quality loss. |
39
+ | [PLLuM-8x7B-chat-Q6_K.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q6_K | 36 GB | Very high quality with quantized embed/output weights. |
40
+ | [PLLuM-8x7B-chat-Q8_0.gguf](https://huggingface.co/Nondzu/PLLuM-8x7B-chat-GGUF/tree/main) | Q8_0 | 47 GB | Maximum quality quantization. |
41
+
42
+ ## Downloading Using Hugging Face CLI
43
+
44
+ <details>
45
+ <summary>Click to view download instructions</summary>
46
+
47
+ First, ensure you have the Hugging Face CLI installed:
48
+
49
+ ```bash
50
+ pip install -U "huggingface_hub[cli]"
51
+ ```
52
+
53
+ Then, target a specific file to download:
54
+
55
+ ```bash
56
+ huggingface-cli download Nondzu/PLLuM-8x7B-chat-GGUF --include "PLLuM-8x7B-chat-Q4_K_M.gguf" --local-dir ./
57
+ ```
58
+
59
+ For larger files, you can specify a new local directory (e.g., `PLLuM-8x7B-chat-Q8_0`) or download them directly into the current directory (`./`).
60
+
61
+ </details>