QuantFactory/sillyrp-7b-GGUF
This is quantized version of nonetrix/sillyrp-7b created using llama.cpp
Original Model Card
Silly RP 7B
First time with merges really, basically experimenting and seeing what works. Seems solid in my tests so far, but no guarantees on quality โ give it a shot and share your feedback! Eager to hear how others like it. Honestly, I'm still learning the ropes, like the benefits of different merge methods :P
Extra info
- Chat template: ChatML
base_model:
- tavtav/eros-7b-test
- NousResearch/Nous-Hermes-2-Mistral-7B-DPO
- maywell/Synatra-7B-v0.3-RP
- NeverSleep/Noromaid-7B-0.4-DPO
- cogbuji/Mr-Grammatology-clinical-problems-Mistral-7B-0.5 library_name: transformers tags:
- mergekit
- merge
output
This is a merge of pre-trained language models created using mergekit.
Merge Details
Merge Method
This model was merged using the task arithmetic merge method using NeverSleep/Noromaid-7B-0.4-DPO as a base.
Models Merged
The following models were included in the merge:
- tavtav/eros-7b-test
- NousResearch/Nous-Hermes-2-Mistral-7B-DPO
- maywell/Synatra-7B-v0.3-RP
- cogbuji/Mr-Grammatology-clinical-problems-Mistral-7B-0.5
Configuration
The following YAML configuration was used to produce this model:
base_model: NeverSleep/Noromaid-7B-0.4-DPO
models:
- model: maywell/Synatra-7B-v0.3-RP
parameters:
weight: 0.2
- model: tavtav/eros-7b-test
parameters:
weight: 0.2
- model: cogbuji/Mr-Grammatology-clinical-problems-Mistral-7B-0.5
parameters:
weight: 0.2
- model: NousResearch/Nous-Hermes-2-Mistral-7B-DPO
parameters:
weight: 0.2
merge_method: task_arithmetic
parameters:
weight: 0.17
dtype: float16
random_seed: 694201337567099116663322537
- Downloads last month
- 22
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit