Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 73.85
AI2 Reasoning Challenge (25-Shot) 70.48
HellaSwag (10-Shot) 88.76
MMLU (5-Shot) 66.94
TruthfulQA (0-shot) 67.01
Winogrande (5-shot) 83.50
GSM8k (5-shot) 66.41
Downloads last month
2
Safetensors
Model size
10.7B params
Tensor type
FP16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for vicgalle/OpenBeagle-11B

Quantizations
1 model

Dataset used to train vicgalle/OpenBeagle-11B

Evaluation results