ExllamaV2 version of the model created by Steelskull!

Original Model https://huggingface.co/Steelskull/Umbra-v3-MoE-4x11b

calibration dataset here.

Requires ExllamaV2, which is being developed by turboderp https://github.com/turboderp/exllamav2 under an MIT license.

Test using 4096 measurement length and rp dataset. Perplexity came out to an 8 vs the Wiki which was at a 6. Haven't tested enough to tell if there is much difference in practice between the two.

Branch is 8b8h using wikitext at 4096 length


Umbra-v3-MoE-4x11b Data Card

Umbra-v3-MoE-4x11b

Creator: SteelSkull

About Umbra-v3-MoE-4x11b: A Mixture of Experts model designed for general assistance with a special knack for storytelling and RP/ERP

Integrates models from notable sources for enhanced performance in diverse tasks.

Source Models:

Update-Log:

The [Umbra Series] keeps rolling out from the [Lumosia Series] garage, aiming to be your digital Alfred with a side of Shakespeare for those RP/ERP nights.

What's Fresh in v3?

Didnโ€™t reinvent the wheel, just slapped on some fancier rims. Upgraded the models and tweaked the prompts a bit. Now, Umbra's not just a general use LLM; it's also focused on spinning stories and "Stories".

Negative Prompt Minimalism

Got the prompts to do a bit of a diet and gym routineโ€”more beef on the positives, trimming down the negatives as usual with a dash of my midnight musings.

Still Guessing, Arenโ€™t We?

Just so we're clear, "v3" is not the messiah of updates. Itโ€™s another experiment in the saga.

Dive into Umbra v3 and toss your two cents my way. Your feedback is the caffeine in my code marathon.

# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Steelskull__Umbra-v3-MoE-4x11b)
Metric Value
Avg. 73.09
AI2 Reasoning Challenge (25-Shot) 68.43
HellaSwag (10-Shot) 87.83
MMLU (5-Shot) 65.99
TruthfulQA (0-shot) 69.30
Winogrande (5-shot) 83.90
GSM8k (5-shot) 63.08
Downloads last month
35
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for AzureBlack/Umbra-v3-MoE-4x11b-exl2-8bpw-8h-rpcal

Evaluation results