GodSlayer-12B-ABYSS / README.md
redrix's picture
Upload README.md
b602ebc verified
---
base_model:
- LatitudeGames/Wayfarer-12B
- ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.2
- PocketDoc/Dans-PersonalityEngine-V1.1.0-12b
- HumanLLMs/Human-Like-Mistral-Nemo-Instruct-2407
- TheDrummer/UnslopNemo-12B-v4
- romaingrx/red-teamer-mistral-nemo
- DavidAU/MN-GRAND-Gutenberg-Lyra4-Lyra-12B-DARKNESS
- rAIfle/Questionable-MN-bf16
- allura-org/MN-12b-RP-Ink
- IntervitensInc/Mistral-Nemo-Base-2407-chatml
library_name: transformers
tags:
- mergekit
- merge
- 12b
- chat
- roleplay
- creative-writing
- DELLA-linear
---
# GodSlayer-12B-ABYSS
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
The goal of this model is to remain fairly stable and coherent, while counteracting positivity-bias and improving realism and diverse responses.
Model #12
## Merge Details
### Merge Method
This model was merged using the NuSLERP merge method using [IntervitensInc/Mistral-Nemo-Base-2407-chatml](https://huggingface.co/IntervitensInc/Mistral-Nemo-Base-2407-chatml) as a base.
### Models Merged
The following models were included in the merge:
* [LatitudeGames/Wayfarer-12B](https://huggingface.co/LatitudeGames/Wayfarer-12B)
* [ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.2](https://huggingface.co/ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.2)
* [PocketDoc/Dans-PersonalityEngine-V1.1.0-12b](https://huggingface.co/PocketDoc/Dans-PersonalityEngine-V1.1.0-12b)
* [HumanLLMs/Human-Like-Mistral-Nemo-Instruct-2407](https://huggingface.co/HumanLLMs/Human-Like-Mistral-Nemo-Instruct-2407)
* [romaingrx/red-teamer-mistral-nemo](https://huggingface.co/romaingrx/red-teamer-mistral-nemo)
* [DavidAU/MN-GRAND-Gutenberg-Lyra4-Lyra-12B-DARKNESS](https://huggingface.co/DavidAU/MN-GRAND-Gutenberg-Lyra4-Lyra-12B-DARKNESS)
* [rAIfle/Questionable-MN-bf16](https://huggingface.co/rAIfle/Questionable-MN-bf16)
* [allura-org/MN-12b-RP-Ink](https://huggingface.co/allura-org/MN-12b-RP-Ink)
### Configuration
The following YAML configurations were used to produce this model:
```yaml
# P1:
models:
- model: PocketDoc/Dans-PersonalityEngine-V1.1.0-12b
parameters:
weight:
- filter: self_attn
value: 0.2
- filter: mlp
value: 0.2
- value: 0.2
density: 0.6
- model: ArliAI/Mistral-Nemo-12B-ArliAI-RPMax-v1.2
parameters:
weight:
- filter: self_attn
value: 0.15
- filter: mlp
value: 0.15
- value: 0.15
density: 0.55
- model: HumanLLMs/Human-Like-Mistral-Nemo-Instruct-2407
parameters:
weight:
- filter: self_attn
value: 0.1
- filter: mlp
value: 0.1
- value: 0.1
density: 0.5
- model: LatitudeGames/Wayfarer-12B
parameters:
weight:
- filter: self_attn
value: 0.25
- filter: mlp
value: 0.25
- value: 0.25
density: 0.65
base_model: TheDrummer/UnslopNemo-12B-v4
merge_method: della_linear
dtype: bfloat16
chat_template: "chatml"
tokenizer_source: union
parameters:
normalize: true
int8_mask: true
epsilon: 0.1
lambda: 1
```
```yaml
# P2:
models:
- model: rAIfle/Questionable-MN-bf16
parameters:
weight:
- filter: self_attn
value: 0.2
- filter: mlp
value: 0.2
- value: 0.2
density: 0.6
- model: DavidAU/MN-GRAND-Gutenberg-Lyra4-Lyra-12B-DARKNESS
parameters:
weight:
- filter: self_attn
value: 0.3
- filter: mlp
value: 0.3
- value: 0.3
density: 0.7
- model: allura-org/MN-12b-RP-Ink
parameters:
weight:
- filter: self_attn
value: 0.35
- filter: mlp
value: 0.35
- value: 0.35
density: 0.75
- model: romaingrx/red-teamer-mistral-nemo
parameters:
weight:
- filter: self_attn
value: 0.25
- filter: mlp
value: 0.25
- value: 0.25
density: 0.65
base_model: TheDrummer/UnslopNemo-12B-v4
merge_method: della_linear
dtype: bfloat16
chat_template: "chatml"
tokenizer_source: union
parameters:
normalize: true
int8_mask: true
epsilon: 0.1
lambda: 1
```
```yaml
# Final:
models:
- model: P1
parameters:
weight: 0.5
- model: P2
parameters:
weight: 0.5
base_model: IntervitensInc/Mistral-Nemo-Base-2407-chatml
merge_method: nuslerp
dtype: bfloat16
chat_template: "chatml"
tokenizer:
source: union
parameters:
normalize: true
int8_mask: true
```