Upload folder using huggingface_hub
Browse files- README.md +41 -0
- chat_template.jinja +1 -0
- config.json +37 -0
- generation_config.json +7 -0
- model-00001-of-00007.safetensors +3 -0
- model-00002-of-00007.safetensors +3 -0
- model-00003-of-00007.safetensors +3 -0
- model-00004-of-00007.safetensors +3 -0
- model-00005-of-00007.safetensors +3 -0
- model-00006-of-00007.safetensors +3 -0
- model-00007-of-00007.safetensors +3 -0
- model.safetensors.index.json +0 -0
- quantization_config.json +0 -0
- special_tokens_map.json +24 -0
- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer_config.json +0 -0
README.md
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model:
|
3 |
+
- mistralai/Mistral-Large-Instruct-2411
|
4 |
+
---
|
5 |
+
# Join our Discord! https://discord.gg/BeaverAI
|
6 |
+
## More than 7000 members strong 💪 A hub for users and makers alike!
|
7 |
+
---
|
8 |
+
## Drummer is open for work / employment (I'm a Software Engineer). Contact me through any of these channels: https://linktr.ee/thelocaldrummer
|
9 |
+
### Thank you to everyone who subscribed through [Patreon](https://www.patreon.com/TheDrummer). Your suppprt helps me chug along in this brave new world.
|
10 |
+
---
|
11 |
+
|
12 |
+
[Drummer](https://huggingface.co/TheDrummer) proudly presents...
|
13 |
+
|
14 |
+
# Behemoth X 123B v2 🦣
|
15 |
+
|
16 |
+
## Usage
|
17 |
+
|
18 |
+
- Mistral v7 (Non-Tekken) | (i.e., Mistral v3 + `[SYSTEM_PROMPT] `)
|
19 |
+
- Non-reasoning model
|
20 |
+
|
21 |
+
## Description
|
22 |
+
|
23 |
+
> Seems to pass a secrecy test in a few gens. Successfully tracked who share a secret among 6 characters and who don't. Really liking the prose. Logic is mostly decent.
|
24 |
+
|
25 |
+
> so far its 5 out of 5. made me cry. would let it stab me in the feels again.
|
26 |
+
|
27 |
+
> This is one of the moments where I really enjoyed reading the generation
|
28 |
+
|
29 |
+
> Recall is fantastic in v2b. I had a response just now that pulled in like 20 minor details. It was nuts. I'm at ~100 gens in an RP now and v2b has been perfect throughout so far. Maybe you hit on some magic like midnight miqu, I dunno.
|
30 |
+
|
31 |
+
> Language choice is better than OG too. That's what Monstral gave that I liked so much. My card is an evil character and she is keeping it hidden so so well. Laying the trap slowly, gaining trust. It's just amazing to watch. If this keeps up this might be your best model ever imo
|
32 |
+
|
33 |
+
> I mostly do mixed co-narrator/mc RP content on my phone. I'm so deep into one storyline it takes a minute of furious scrolling to reach back at the top, and it's still going coherent. No templates, just a lightweight sys prompt. Great model, I'd hate to go back to 24B or even 70B from here. It recognized the video game Katawa Shoujo and incorporated character flaws seamlessly into my story.
|
34 |
+
|
35 |
+
## Links
|
36 |
+
- Original: https://huggingface.co/TheDrummer/Behemoth-X-123B-v2
|
37 |
+
- GGUF: https://huggingface.co/TheDrummer/Behemoth-X-123B-v2-GGUF
|
38 |
+
- iMatrix (recommended): https://huggingface.co/bartowski/TheDrummer_Behemoth-X-123B-v2-GGUF
|
39 |
+
- EXL3: https://huggingface.co/ArtusDev/TheDrummer_Behemoth-X-123B-v2-EXL3
|
40 |
+
|
41 |
+
`config-v2b`
|
chat_template.jinja
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{{ bos_token }}{% for message in messages %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + '[/INST]' }}{% elif message['role'] == 'system' %}{{ '[SYSTEM_PROMPT] ' + message['content'] + '[/SYSTEM_PROMPT]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + eos_token }}{% else %}{{ raise_exception('Only user, system and assistant roles are supported!') }}{% endif %}{% endfor %}
|
config.json
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"MistralForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_dropout": 0.0,
|
6 |
+
"bos_token_id": 1,
|
7 |
+
"eos_token_id": 2,
|
8 |
+
"head_dim": 128,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 12288,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 28672,
|
13 |
+
"max_position_embeddings": 131072,
|
14 |
+
"model_type": "mistral",
|
15 |
+
"num_attention_heads": 96,
|
16 |
+
"num_hidden_layers": 88,
|
17 |
+
"num_key_value_heads": 8,
|
18 |
+
"rms_norm_eps": 1e-05,
|
19 |
+
"rope_theta": 1000000.0,
|
20 |
+
"sliding_window": null,
|
21 |
+
"tie_word_embeddings": false,
|
22 |
+
"torch_dtype": "bfloat16",
|
23 |
+
"transformers_version": "4.55.2",
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 32768,
|
26 |
+
"quantization_config": {
|
27 |
+
"quant_method": "exl3",
|
28 |
+
"version": "0.0.6",
|
29 |
+
"bits": 3.5,
|
30 |
+
"head_bits": 6,
|
31 |
+
"calibration": {
|
32 |
+
"rows": 100,
|
33 |
+
"cols": 2048
|
34 |
+
},
|
35 |
+
"out_scales": "auto"
|
36 |
+
}
|
37 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 1,
|
4 |
+
"do_sample": true,
|
5 |
+
"eos_token_id": 2,
|
6 |
+
"transformers_version": "4.55.2"
|
7 |
+
}
|
model-00001-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:afbfd09670290f8cd71d092b36faa0e327b07cd68a7fdfad54d2bbf5b550720a
|
3 |
+
size 8066315920
|
model-00002-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46426c65f9f1fb6c5c86413905d92c5cc784ae89857f3ca88d3a74182784de32
|
3 |
+
size 8476157712
|
model-00003-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:85445e31f6b19e0a93993503bbe7935513cbedceedf9ebef3e50712bbc3d9f9b
|
3 |
+
size 8502896400
|
model-00004-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f645fe1146c04118281d50300b764aa6250cce02b64e49386dbc739ccd5bc012
|
3 |
+
size 8482449168
|
model-00005-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:832b795dfb66b505aa0ab4b1dfd3f367ce1ba4780a4abfc2d515c7722677fc47
|
3 |
+
size 8482449168
|
model-00006-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe45c2d609f3c8c2d55294e6feb4857be6a67fcf12557bf7974ab67ea9c48e5e
|
3 |
+
size 8476157712
|
model-00007-of-00007.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ac28d35e3f69aea76f85296f5ca5a89fff85fcae5c21c6516a6bac9d2880a816
|
3 |
+
size 3945978632
|
model.safetensors.index.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
quantization_config.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "</s>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<unk>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b968b8dc352f42192367337c78ccc61e1eaddc6d641a579372d4f20694beb7a
|
3 |
+
size 587562
|
tokenizer_config.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|