--- base_model: - WhiteRabbitNeo/Llama-3.1-WhiteRabbitNeo-2-70B - meta-llama/Llama-3.3-70B-Instruct - schonsense/70B_unstruct library_name: transformers tags: - mergekit - merge --- # 70B_unstructWR ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317d4867690c5b55e61ce3d/ewMDrxExnEE61Cunrc4D8.png) This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DELLA](https://arxiv.org/abs/2406.11617) merge method using [meta-llama/Llama-3.3-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct) as a base. ### Models Merged The following models were included in the merge: * [WhiteRabbitNeo/Llama-3.1-WhiteRabbitNeo-2-70B](https://huggingface.co/WhiteRabbitNeo/Llama-3.1-WhiteRabbitNeo-2-70B) * [schonsense/70B_unstruct](https://huggingface.co/schonsense/70B_unstruct) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: schonsense/70B_unstruct parameters: density: 0.7 epsilon: 0.2 weight: 0.9 - model: WhiteRabbitNeo/Llama-3.1-WhiteRabbitNeo-2-70B parameters: density: 0.9 epsilon: 0.05 weight: 0.1 - model: meta-llama/Llama-3.3-70B-Instruct merge_method: della base_model: meta-llama/Llama-3.3-70B-Instruct tokenizer_source: meta-llama/Llama-3.3-70B-Instruct parameters: normalize: false int8_mask: false lambda: 1.0 dtype: float32 out_dtype: bfloat16 ```