--- base_model: [] library_name: transformers tags: - mergekit - merge --- # EVA-Tissint-14B This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). Combining [Tissint](https://huggingface.co/Ttimofeyka/Tissint-14B-128k-RP) with [EVA v0.2](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2) seemed to yield nice results in RP, I've uploaded the model so others can try it as well. Recommended samplers are 0.1 Min P, 1.02 Repititon penalty, and DRY's range set to 2048. If you'd like to use XTC, I recommend a threshold of 0.2. Lower thresholds seem to adversely affect the coherency. # Quantisations Static: https://huggingface.co/mradermacher/EVA-Tissint-14B-GGUF Imatrix: https://huggingface.co/mradermacher/EVA-Tissint-14B-i1-GGUF ### Merge Method This model was merged using the della_linear merge method using EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2 as a base. ### Models Merged The following models were included in the merge: * Ttimofeyka/Tissint-14B-128k-RP ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: Ttimofeyka/Tissint-14B-128k-RP parameters: density: 0.4 weight: 0.3 - model: EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2 parameters: density: 0.6 weight: 0.7 merge_method: della_linear base_model: EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2 parameters: epsilon: 0.05 lambda: 1 dtype: bfloat16 ```