--- base_model: - nbeerbower/Llama3.1-Gutenberg-Doppel-70B - Nexesenex/Llama_3.x_70b_SmarTricks_V1.01 - Nexesenex/Llama_3.1_70b_HighPriestess_R1_V1 - Steelskull/L3.3-Electra-R1-70b - Nexesenex/Llama_3.3_70b_DarkHorse - Nexesenex/Llama_3.1_70b_FLDx2-Tess3_abliterated_fusion_norm library_name: transformers tags: - mergekit - merge --- # about V3.0 changes (relatively minor update) : - DarkHorse replace DoppelGangerR1 to add a bit of Negative Llama at the expense of a bit of Fallen Llama R1. - A bit of Fallen Llama is recovered by using Smartricks instead of Smartracks as a base. - Priestess is upgraded with Lumitron Lorablated. - Tess is merged with Hitachi FLDx2 in the perplexity dropper model. Electra R1 and GutenbergDoppel are kept as they were. If you have V2 already, this model is quite similar, and the difference might not be worth a download. --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Nexesenex/Llama_3.x_70b_SmarTricks_V1.01](https://huggingface.co/Nexesenex/Llama_3.x_70b_SmarTricks_V1.01) as a base. ### Models Merged The following models were included in the merge: * [nbeerbower/Llama3.1-Gutenberg-Doppel-70B](https://huggingface.co/nbeerbower/Llama3.1-Gutenberg-Doppel-70B) * [Nexesenex/Llama_3.1_70b_HighPriestess_R1_V1](https://huggingface.co/Nexesenex/Llama_3.1_70b_HighPriestess_R1_V1) * [Steelskull/L3.3-Electra-R1-70b](https://huggingface.co/Steelskull/L3.3-Electra-R1-70b) * [Nexesenex/Llama_3.3_70b_DarkHorse](https://huggingface.co/Nexesenex/Llama_3.3_70b_DarkHorse) * [Nexesenex/Llama_3.1_70b_FLDx2-Tess3_abliterated_fusion_norm](https://huggingface.co/Nexesenex/Llama_3.1_70b_FLDx2-Tess3_abliterated_fusion_norm) ### Configuration The following YAML configuration was used to produce this model: ```yaml merge_method: model_stock models: - model: Nexesenex/Llama_3.1_70b_FLDx2-Tess3_abliterated_fusion_norm parameters: weight: 1.0 - model: nbeerbower/Llama3.1-Gutenberg-Doppel-70B parameters: weight: 1.0 - model: Nexesenex/Llama_3.1_70b_HighPriestess_R1_V1 parameters: weight: 1.0 - model: Steelskull/L3.3-Electra-R1-70b parameters: weight: 1.0 - model: Nexesenex/Llama_3.3_70b_DarkHorse parameters: weight: 1.0 base_model: Nexesenex/Llama_3.x_70b_SmarTricks_V1.01 dtype: bfloat16 out_dtype: bfloat16 parameters: int8_mask: true normalize: true rescale: false chat_template: auto tokenizer: source: union ```