ArtusDev commited on
Commit
32a61af
·
verified ·
1 Parent(s): aa39ef6

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - mistralai/Mistral-Large-Instruct-2411
4
+ ---
5
+ # Join our Discord! https://discord.gg/BeaverAI
6
+ ## More than 7000 members strong 💪 A hub for users and makers alike!
7
+ ---
8
+ ## Drummer is open for work / employment (I'm a Software Engineer). Contact me through any of these channels: https://linktr.ee/thelocaldrummer
9
+ ### Thank you to everyone who subscribed through [Patreon](https://www.patreon.com/TheDrummer). Your suppprt helps me chug along in this brave new world.
10
+ ---
11
+
12
+ [Drummer](https://huggingface.co/TheDrummer) proudly presents...
13
+
14
+ # Behemoth X 123B v2 🦣
15
+
16
+ ## Usage
17
+
18
+ - Mistral v7 (Non-Tekken) | (i.e., Mistral v3 + `[SYSTEM_PROMPT] `)
19
+ - Non-reasoning model
20
+
21
+ ## Description
22
+
23
+ > Seems to pass a secrecy test in a few gens. Successfully tracked who share a secret among 6 characters and who don't. Really liking the prose. Logic is mostly decent.
24
+
25
+ > so far its 5 out of 5. made me cry. would let it stab me in the feels again.
26
+
27
+ > This is one of the moments where I really enjoyed reading the generation
28
+
29
+ > Recall is fantastic in v2b. I had a response just now that pulled in like 20 minor details. It was nuts. I'm at ~100 gens in an RP now and v2b has been perfect throughout so far. Maybe you hit on some magic like midnight miqu, I dunno.
30
+
31
+ > Language choice is better than OG too. That's what Monstral gave that I liked so much. My card is an evil character and she is keeping it hidden so so well. Laying the trap slowly, gaining trust. It's just amazing to watch. If this keeps up this might be your best model ever imo
32
+
33
+ > I mostly do mixed co-narrator/mc RP content on my phone. I'm so deep into one storyline it takes a minute of furious scrolling to reach back at the top, and it's still going coherent. No templates, just a lightweight sys prompt. Great model, I'd hate to go back to 24B or even 70B from here. It recognized the video game Katawa Shoujo and incorporated character flaws seamlessly into my story.
34
+
35
+ ## Links
36
+ - Original: https://huggingface.co/TheDrummer/Behemoth-X-123B-v2
37
+ - GGUF: https://huggingface.co/TheDrummer/Behemoth-X-123B-v2-GGUF
38
+ - iMatrix (recommended): https://huggingface.co/bartowski/TheDrummer_Behemoth-X-123B-v2-GGUF
39
+ - EXL3: https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3
40
+
41
+ `config-v2b`
chat_template.jinja ADDED
@@ -0,0 +1 @@
 
 
1
+ {{ bos_token }}{% for message in messages %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + '[/INST]' }}{% elif message['role'] == 'system' %}{{ '[SYSTEM_PROMPT] ' + message['content'] + '[/SYSTEM_PROMPT]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + eos_token }}{% else %}{{ raise_exception('Only user, system and assistant roles are supported!') }}{% endif %}{% endfor %}
config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MistralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "head_dim": 128,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 12288,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 28672,
13
+ "max_position_embeddings": 131072,
14
+ "model_type": "mistral",
15
+ "num_attention_heads": 96,
16
+ "num_hidden_layers": 88,
17
+ "num_key_value_heads": 8,
18
+ "rms_norm_eps": 1e-05,
19
+ "rope_theta": 1000000.0,
20
+ "sliding_window": null,
21
+ "tie_word_embeddings": false,
22
+ "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.55.2",
24
+ "use_cache": true,
25
+ "vocab_size": 32768,
26
+ "quantization_config": {
27
+ "quant_method": "exl3",
28
+ "version": "0.0.6",
29
+ "bits": 3.5,
30
+ "head_bits": 6,
31
+ "calibration": {
32
+ "rows": 100,
33
+ "cols": 2048
34
+ },
35
+ "out_scales": "auto"
36
+ }
37
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "transformers_version": "4.55.2"
7
+ }
model-00001-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afbfd09670290f8cd71d092b36faa0e327b07cd68a7fdfad54d2bbf5b550720a
3
+ size 8066315920
model-00002-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46426c65f9f1fb6c5c86413905d92c5cc784ae89857f3ca88d3a74182784de32
3
+ size 8476157712
model-00003-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85445e31f6b19e0a93993503bbe7935513cbedceedf9ebef3e50712bbc3d9f9b
3
+ size 8502896400
model-00004-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f645fe1146c04118281d50300b764aa6250cce02b64e49386dbc739ccd5bc012
3
+ size 8482449168
model-00005-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:832b795dfb66b505aa0ab4b1dfd3f367ce1ba4780a4abfc2d515c7722677fc47
3
+ size 8482449168
model-00006-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe45c2d609f3c8c2d55294e6feb4857be6a67fcf12557bf7974ab67ea9c48e5e
3
+ size 8476157712
model-00007-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac28d35e3f69aea76f85296f5ca5a89fff85fcae5c21c6516a6bac9d2880a816
3
+ size 3945978632
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
quantization_config.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b968b8dc352f42192367337c78ccc61e1eaddc6d641a579372d4f20694beb7a
3
+ size 587562
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff