ArtusDev commited on
Commit
d24d683
·
verified ·
1 Parent(s): a691da8

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,41 +1,101 @@
1
  ---
2
- base_model: TheDrummer/Behemoth-X-123B-v2
3
- base_model_relation: quantized
4
- quantized_by: ArtusDev
5
- tags:
6
- - exl3
7
  ---
 
 
 
 
 
 
 
 
 
 
8
 
9
- ## EXL3 Quants of TheDrummer/Behemoth-X-123B-v2
10
 
11
- EXL3 quants of [TheDrummer/Behemoth-X-123B-v2](https://huggingface.co/TheDrummer/Behemoth-X-123B-v2) using <a href="https://github.com/turboderp-org/exllamav3/">exllamav3</a> for quantization.
 
 
 
 
 
12
 
13
- ### Quants
14
- | Quant(Revision) | Bits per Weight | Head Bits |
15
- | -------- | ---------- | --------- |
16
- | [2.5_H6](https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3/tree/2.5bpw_H6) | 2.5 | 6 |
17
- | [3.0_H6](https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3/tree/3.0bpw_H6) | 3.0 | 6 |
18
- | [3.5_H6](https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3/tree/3.5bpw_H6) | 3.5 | 6 |
19
- | [4.0_H6](https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3/tree/4.0bpw_H6) | 4.0 | 6 |
20
- | [4.25_H6](https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3/tree/4.25bpw_H6) | 4.25 | 6 |
21
- | [5.0_H6](https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3/tree/5.0bpw_H6) | 5.0 | 6 |
22
- | [6.0_H6](https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3/tree/6.0bpw_H6) | 6.0 | 6 |
23
- | [8.0_H8](https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3/tree/8.0bpw_H8) | 8.0 | 8 |
24
 
25
- ### Downloading quants with huggingface-cli
 
 
 
 
26
 
27
  <details>
28
- <summary>Click to view download instructions</summary>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
 
30
- Install hugginface-cli:
31
 
32
- ```bash
33
- pip install -U "huggingface_hub[cli]"
34
- ```
35
 
36
- Download quant by targeting the specific quant revision (branch):
 
 
 
 
37
 
38
- ```
39
- huggingface-cli download ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3 --revision "5.0bpw_H6" --local-dir ./
40
- ```
41
- </details>
 
1
  ---
2
+ base_model:
3
+ - mistralai/Mistral-Large-Instruct-2411
 
 
 
4
  ---
5
+ # Join our Discord! https://discord.gg/BeaverAI
6
+ ## More than 7000 members strong 💪 A hub for users and makers alike!
7
+ ---
8
+ ## Drummer is open for work / employment (I'm a Software Engineer). Contact me through any of these channels: https://linktr.ee/thelocaldrummer
9
+ ### Thank you to everyone who subscribed through [Patreon](https://www.patreon.com/TheDrummer). Your support helps me chug along in this brave new world.
10
+
11
+ ### FAQ for those out-of-the-loop
12
+
13
+ <details>
14
+ <summary>🐶 Who is Drummer?</summary>
15
 
16
+ Hi! I'm Drummer. I'm a Software Engineer with experience in JavaScript, Golang, Python, and generally engineering the crap out of things.
17
 
18
+ Why I'm in the AI space:
19
+
20
+ - **Exploration:** Everyone is trying to figure out how AI works and what it's capable of. I am too - just not in creating the smartest, safest model at all costs.
21
+ - **Upskill:** The world is headed towards AI. It is here to stay. This has been my way of brushing up in this new form of computing.
22
+ - **Value:** I yearn to create value. I feel satisfaction and fulfillment in providing something meaningful for others.
23
+ - **Fun:** It's just fun using and making models. It's also fun coming up with theories and realizing them in practice (training AI).
24
 
25
+ I started my tuning venture back in mid-2024 when I wanted to improve its literary capabilities.
26
+ I've come a long way since then and I have branched out and specialized.
 
 
 
 
 
 
 
 
 
27
 
28
+ Foundational models today are optimized for non-creative uses, and I believe there is a place for AI in creativity and entertainment.
29
+
30
+ I am here to take *the road less traveled by*.
31
+
32
+ </details>
33
 
34
  <details>
35
+ <summary>❓ What are my models like?</summary>
36
+
37
+ **Bottomline:** My models are usually geared towards creativity (writing & interactivity), usability (no censorship & overalignment), and entertainment!
38
+
39
+ While intelligence, correctness, and problem solving are not my priority, they are still one of many qualities I want in my models.
40
+
41
+ The primary goal is to enhance the experience for users looking to use models for creative uses.
42
+
43
+ The *subqualities* I look for are the following:
44
+
45
+ - **Alignment:** How far off is it from being a goody, syncopathic assistant-writer?
46
+ - **Writing:** Does it string together words and sentences in a pleasant & effective way like a wordsmith?
47
+ - **Dynamism:** How good is the AI at being compelling and intriguing in its storytelling?
48
+ - **Adherence:** Is it sticking to the prompt?
49
+ - **Knowledge:** Does it know about the world in both its truths and stories told?
50
+ - **Intelligence:** Can it handle nuance, complexity, and logic?
51
+ - **Formatting:** Is it biased towards writing paragraphs / lists or can it come up with its own?
52
+
53
+ If it doesn't excel in one of these qualities, or if it's overall mediocre for its size, then I would most likely reiterate until I get something right.
54
+
55
+ </details>
56
+
57
+ <details>
58
+ <summary>🦫 What is BeaverAI?</summary>
59
+
60
+ BeaverAI is a Discord server.
61
+ </details>
62
+
63
+ ---
64
+
65
+ [Drummer](https://huggingface.co/TheDrummer) proudly presents...
66
+
67
+ # Behemoth X 123B v2 🦣
68
+
69
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65f2fd1c25b848bd061b5c2e/XiBlPtMdK3kvqw2AcVbCa.png)
70
+
71
+ ## Usage
72
+
73
+ - Mistral v7 (Non-Tekken) | (i.e., Mistral v3 + `[SYSTEM_PROMPT] `)
74
+ - Non-reasoning model
75
+
76
+ ## Description
77
+
78
+ Behemoth X is a finetune Mistral Large 2411.
79
+ X denotes a significant update in training (compared to its predecessors), pushing its capabilities further with updated knowledge and techniques.
80
+
81
+ > Seems to pass a secrecy test in a few gens. Successfully tracked who share a secret among 6 characters and who don't. Really liking the prose. Logic is mostly decent.
82
+
83
+ > so far its 5 out of 5. made me cry. would let it stab me in the feels again.
84
+
85
+ > This is one of the moments where I really enjoyed reading the generation
86
+
87
+ > Recall is fantastic in v2b. I had a response just now that pulled in like 20 minor details. It was nuts. I'm at ~100 gens in an RP now and v2b has been perfect throughout so far. Maybe you hit on some magic like midnight miqu, I dunno.
88
+
89
+ > genuinely impressed with this model, gave it a starting prompt and it's been generating a coherent story for 19k tokens straight, and it doesn't feel like there's much slop
90
 
91
+ > Language choice is better than OG too. That's what Monstral gave that I liked so much. My card is an evil character and she is keeping it hidden so so well. Laying the trap slowly, gaining trust. It's just amazing to watch. If this keeps up this might be your best model ever imo
92
 
93
+ > I mostly do mixed co-narrator/mc RP content on my phone. I'm so deep into one storyline it takes a minute of furious scrolling to reach back at the top, and it's still going coherent. No templates, just a lightweight sys prompt. Great model, I'd hate to go back to 24B or even 70B from here. It recognized the video game Katawa Shoujo and incorporated character flaws seamlessly into my story.
 
 
94
 
95
+ ## Links
96
+ - Original: https://huggingface.co/TheDrummer/Behemoth-X-123B-v2
97
+ - GGUF: https://huggingface.co/TheDrummer/Behemoth-X-123B-v2-GGUF
98
+ - iMatrix (recommended): https://huggingface.co/bartowski/TheDrummer_Behemoth-X-123B-v2-GGUF
99
+ - EXL3: https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3
100
 
101
+ `config-v2b`
 
 
 
chat_template.jinja ADDED
@@ -0,0 +1 @@
 
 
1
+ {{ bos_token }}{% for message in messages %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + '[/INST]' }}{% elif message['role'] == 'system' %}{{ '[SYSTEM_PROMPT] ' + message['content'] + '[/SYSTEM_PROMPT]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + eos_token }}{% else %}{{ raise_exception('Only user, system and assistant roles are supported!') }}{% endif %}{% endfor %}
config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MistralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "head_dim": 128,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 12288,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 28672,
13
+ "max_position_embeddings": 131072,
14
+ "model_type": "mistral",
15
+ "num_attention_heads": 96,
16
+ "num_hidden_layers": 88,
17
+ "num_key_value_heads": 8,
18
+ "rms_norm_eps": 1e-05,
19
+ "rope_theta": 1000000.0,
20
+ "sliding_window": null,
21
+ "tie_word_embeddings": false,
22
+ "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.55.2",
24
+ "use_cache": true,
25
+ "vocab_size": 32768,
26
+ "quantization_config": {
27
+ "quant_method": "exl3",
28
+ "version": "0.0.7",
29
+ "bits": 3.75,
30
+ "head_bits": 6,
31
+ "calibration": {
32
+ "rows": 100,
33
+ "cols": 2048
34
+ },
35
+ "out_scales": "auto"
36
+ }
37
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "transformers_version": "4.55.2"
7
+ }
model-00001-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e129a1f604bbf7c670191e8158a060f007c063333d8cb30ba2b9253f4fe395f
3
+ size 8586933904
model-00002-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fbac2043fbb6a1a2a73e5c107717adc371c11ea65f32d891d9057dca49ed569
3
+ size 8439532896
model-00003-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d1c64f77f8b4c1043f0b0fa1fcee37b1dec90c202709ba02baf54aa3caa232f
3
+ size 8439532896
model-00004-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:408497b285b7243706fb9410918d014849c1f789448b534381432ab70d38a686
3
+ size 8433241440
model-00005-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f5cfc1655630f74924c55f00eced34c0b7d79dc06da9dde3e7a18c111bc40e1
3
+ size 8439532904
model-00006-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e97c448832b753eda787f5b9ce2d0ac4585bb35bb9cede0d4a296f00e610ad9
3
+ size 8439532904
model-00007-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:865b6401b2f576a51b92005de1c15d9a7a976365315987237fdd87aea1175fa5
3
+ size 7452564320
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
quantization_config.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b968b8dc352f42192367337c78ccc61e1eaddc6d641a579372d4f20694beb7a
3
+ size 587562
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff