--- base_model: - Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2 - NousResearch/Hermes-3-Llama-3.1-8B - arcee-ai/Llama-3.1-SuperNova-Lite - akjindal53244/Llama-3.1-Storm-8B - mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated library_name: transformers tags: - mergekit - merge license: llama3.1 model-index: - name: Llama3.1-LexiHermes-SuperStorm results: - task: type: text-generation name: Text Generation dataset: name: IFEval (0-Shot) type: wis-k/instruction-following-eval split: train args: num_few_shot: 0 metrics: - type: inst_level_strict_acc and prompt_level_strict_acc value: 78.35 name: averaged accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FLlama3.1-LexiHermes-SuperStorm name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: BBH (3-Shot) type: SaylorTwift/bbh split: test args: num_few_shot: 3 metrics: - type: acc_norm value: 32.55 name: normalized accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FLlama3.1-LexiHermes-SuperStorm name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MATH Lvl 5 (4-Shot) type: lighteval/MATH-Hard split: test args: num_few_shot: 4 metrics: - type: exact_match value: 16.16 name: exact match source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FLlama3.1-LexiHermes-SuperStorm name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: GPQA (0-shot) type: Idavidrein/gpqa split: train args: num_few_shot: 0 metrics: - type: acc_norm value: 9.73 name: acc_norm source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FLlama3.1-LexiHermes-SuperStorm name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MuSR (0-shot) type: TAUR-Lab/MuSR args: num_few_shot: 0 metrics: - type: acc_norm value: 8.2 name: acc_norm source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FLlama3.1-LexiHermes-SuperStorm name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MMLU-PRO (5-shot) type: TIGER-Lab/MMLU-Pro config: main split: test args: num_few_shot: 5 metrics: - type: acc value: 31.6 name: accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=agentlans%2FLlama3.1-LexiHermes-SuperStorm name: Open LLM Leaderboard --- # Llama3.1-LexiHermes-SuperStorm A mix of four high-performing models using the new [SCE](https://arxiv.org/abs/2408.07990) merging technique. This model was created using [mergekit](https://github.com/cg123/mergekit). Models Merged - [Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2](https://huggingface.co/Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2) - [NousResearch/Hermes-3-Llama-3.1-8B](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B) - [arcee-ai/Llama-3.1-SuperNova-Lite](https://huggingface.co/arcee-ai/Llama-3.1-SuperNova-Lite) - [akjindal53244/Llama-3.1-Storm-8B](https://huggingface.co/akjindal53244/Llama-3.1-Storm-8B) Base model: [mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated](https://huggingface.co/mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated) ## Features - Solid and consistent style even without finetuning - Abliterated to avoid outright request refusals - Suitable for roleplaying (possibly due to Hermes component) - Can replace Llama 3.1 8B Instruct for general tasks ## Limitations - Function calling and other languages not tested - May struggle with math and logic like other language models - Potential for factual errors, especially in specialized fields - Not intended for public deployment without additional safeguards ## Merge Configuration The following YAML configuration was used: ```yaml models: - model: Hermes-3-Llama-3.1-8B - model: Llama-3.1-8B-Lexi-Uncensored-V2 - model: Llama-3.1-SuperNova-Lite - model: Llama-3.1-Storm-8B merge_method: sce base_model: Meta-Llama-3.1-8B-Instruct-abliterated parameters: select_topk: 1.5 dtype: bfloat16 ``` # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/agentlans__Llama3.1-LexiHermes-SuperStorm-details)! Summarized results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/contents/viewer/default/train?q=agentlans%2FLlama3.1-LexiHermes-SuperStorm&sort[column]=Average%20%E2%AC%86%EF%B8%8F&sort[direction]=desc)! | Metric |Value (%)| |-------------------|--------:| |**Average** | 29.43| |IFEval (0-Shot) | 78.35| |BBH (3-Shot) | 32.55| |MATH Lvl 5 (4-Shot)| 16.16| |GPQA (0-shot) | 9.73| |MuSR (0-shot) | 8.20| |MMLU-PRO (5-shot) | 31.60|