Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 76.10
AI2 Reasoning Challenge (25-Shot) 72.27
HellaSwag (10-Shot) 89.31
MMLU (5-Shot) 66.55
TruthfulQA (0-shot) 78.55
Winogrande (5-shot) 83.82
GSM8k (5-shot) 66.11

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 21.29
IFEval (0-Shot) 52.12
BBH (3-Shot) 33.99
MATH Lvl 5 (4-Shot) 4.76
GPQA (0-shot) 6.60
MuSR (0-shot) 4.11
MMLU-PRO (5-shot) 26.19
Downloads last month
13,886
Safetensors
Model size
10.7B params
Tensor type
FP16
Β·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for vicgalle/CarbonBeagle-11B-truthy

Merges
1 model
Quantizations
4 models

Dataset used to train vicgalle/CarbonBeagle-11B-truthy

Spaces using vicgalle/CarbonBeagle-11B-truthy 5

Evaluation results