Text Generation
Transformers
Safetensors
English
mistral
code
Eval Results
text-generation-inference
leaderboard-pr-bot commited on
Commit
55970a4
·
verified ·
1 Parent(s): 12afe40

Adding Evaluation Results

Browse files

This is an automated PR created with https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr

The purpose of this PR is to add evaluation results from the Open LLM Leaderboard to your model card.

If you encounter any issues, please report them to https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr/discussions

Files changed (1) hide show
  1. README.md +121 -7
README.md CHANGED
@@ -1,28 +1,128 @@
1
  ---
2
  language:
3
  - en
 
4
  library_name: transformers
5
- pipeline_tag: text-generation
 
6
  datasets:
7
  - teknium/OpenHermes-2.5
8
  - TokenBender/python_eval_instruct_51k
9
  - codefuse-ai/Evol-instruction-66k
10
- tags:
11
- - code
12
- license: apache-2.0
13
  model-index:
14
  - name: SpeechlessCoder
15
  results:
16
  - task:
17
  type: text-generation
18
  dataset:
19
- type: openai_humaneval
20
  name: HumanEval
 
21
  metrics:
22
- - name: pass@1
23
- type: pass@1
24
  value: 0.0
 
25
  verified: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  ---
27
 
28
  <p><h1> speechless-mistral-hermes-code-7b </h1></p>
@@ -82,3 +182,17 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
82
  | Winogrande (5-shot) | |
83
  | GSM8K (5-shot) | |
84
  | DROP (3-shot) | |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  language:
3
  - en
4
+ license: apache-2.0
5
  library_name: transformers
6
+ tags:
7
+ - code
8
  datasets:
9
  - teknium/OpenHermes-2.5
10
  - TokenBender/python_eval_instruct_51k
11
  - codefuse-ai/Evol-instruction-66k
12
+ pipeline_tag: text-generation
 
 
13
  model-index:
14
  - name: SpeechlessCoder
15
  results:
16
  - task:
17
  type: text-generation
18
  dataset:
 
19
  name: HumanEval
20
+ type: openai_humaneval
21
  metrics:
22
+ - type: pass@1
 
23
  value: 0.0
24
+ name: pass@1
25
  verified: false
26
+ - task:
27
+ type: text-generation
28
+ name: Text Generation
29
+ dataset:
30
+ name: AI2 Reasoning Challenge (25-Shot)
31
+ type: ai2_arc
32
+ config: ARC-Challenge
33
+ split: test
34
+ args:
35
+ num_few_shot: 25
36
+ metrics:
37
+ - type: acc_norm
38
+ value: 59.39
39
+ name: normalized accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-hermes-code-7b
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: HellaSwag (10-Shot)
48
+ type: hellaswag
49
+ split: validation
50
+ args:
51
+ num_few_shot: 10
52
+ metrics:
53
+ - type: acc_norm
54
+ value: 78.55
55
+ name: normalized accuracy
56
+ source:
57
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-hermes-code-7b
58
+ name: Open LLM Leaderboard
59
+ - task:
60
+ type: text-generation
61
+ name: Text Generation
62
+ dataset:
63
+ name: MMLU (5-Shot)
64
+ type: cais/mmlu
65
+ config: all
66
+ split: test
67
+ args:
68
+ num_few_shot: 5
69
+ metrics:
70
+ - type: acc
71
+ value: 59.88
72
+ name: accuracy
73
+ source:
74
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-hermes-code-7b
75
+ name: Open LLM Leaderboard
76
+ - task:
77
+ type: text-generation
78
+ name: Text Generation
79
+ dataset:
80
+ name: TruthfulQA (0-shot)
81
+ type: truthful_qa
82
+ config: multiple_choice
83
+ split: validation
84
+ args:
85
+ num_few_shot: 0
86
+ metrics:
87
+ - type: mc2
88
+ value: 51.26
89
+ source:
90
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-hermes-code-7b
91
+ name: Open LLM Leaderboard
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: Winogrande (5-shot)
97
+ type: winogrande
98
+ config: winogrande_xl
99
+ split: validation
100
+ args:
101
+ num_few_shot: 5
102
+ metrics:
103
+ - type: acc
104
+ value: 77.27
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-hermes-code-7b
108
+ name: Open LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: GSM8k (5-shot)
114
+ type: gsm8k
115
+ config: main
116
+ split: test
117
+ args:
118
+ num_few_shot: 5
119
+ metrics:
120
+ - type: acc
121
+ value: 40.64
122
+ name: accuracy
123
+ source:
124
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-hermes-code-7b
125
+ name: Open LLM Leaderboard
126
  ---
127
 
128
  <p><h1> speechless-mistral-hermes-code-7b </h1></p>
 
182
  | Winogrande (5-shot) | |
183
  | GSM8K (5-shot) | |
184
  | DROP (3-shot) | |
185
+
186
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
187
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_uukuguy__speechless-mistral-hermes-code-7b)
188
+
189
+ | Metric |Value|
190
+ |---------------------------------|----:|
191
+ |Avg. |61.16|
192
+ |AI2 Reasoning Challenge (25-Shot)|59.39|
193
+ |HellaSwag (10-Shot) |78.55|
194
+ |MMLU (5-Shot) |59.88|
195
+ |TruthfulQA (0-shot) |51.26|
196
+ |Winogrande (5-shot) |77.27|
197
+ |GSM8k (5-shot) |40.64|
198
+