Update README.md
Browse files
README.md
CHANGED
@@ -78,18 +78,18 @@ The evaluation was done zero-shot with Indonesian prompts and only a sample of 1
|
|
78 |
| TA | Mistral-7B-Instruct-v0.3 | 20.70% | 13.85 | 0.00 | 0.37 | 0.52 | 5.31 | 14.20 | - | 0.80 | - |
|
79 |
|
80 |
#### Instruction-following Capabilities (IFEval)
|
81 |
-
|
|
82 |
-
|
83 |
-
| **Model**
|
84 |
-
| gemma-2-9b-it
|
85 |
-
| Meta-Llama-3.1-8B-Instruct
|
86 |
-
| Qwen2-7B-Instruct
|
87 |
-
| llama3-8b-cpt-sealionv2-instruct
|
88 |
-
| aya-23-8B
|
89 |
-
| SeaLLMs-v3-7B-Chat
|
90 |
-
| Mistral-7B-Instruct-v0.3
|
91 |
-
| Meta-Llama-3-8B-Instruct
|
92 |
-
| Sailor-7B-Chat
|
93 |
|
94 |
#### Multi-turn Capatbilities (MT-Bench)
|
95 |
| | **Indonesian** | **Vietnamese** | **English** |
|
|
|
78 |
| TA | Mistral-7B-Instruct-v0.3 | 20.70% | 13.85 | 0.00 | 0.37 | 0.52 | 5.31 | 14.20 | - | 0.80 | - |
|
79 |
|
80 |
#### Instruction-following Capabilities (IFEval)
|
81 |
+
| | **Indonesian** | **Vietnamese** | **English** |
|
82 |
+
|---|:---:|:---:|:---:|
|
83 |
+
| **Model** | **Lang normalised score** | **Lang normalised score** | **Lang normalised score** |
|
84 |
+
| gemma-2-9b-it | 0.88 | 0.77 | 0.85 |
|
85 |
+
| Meta-Llama-3.1-8B-Instruct | 0.68 | 0.68 | 0.85 |
|
86 |
+
| Qwen2-7B-Instruct | 0.63 | 0.65 | 0.70 |
|
87 |
+
| llama3-8b-cpt-sealionv2-instruct | 0.61 | 0.66 | 0.70 |
|
88 |
+
| aya-23-8B | 0.58 | 0.56 | 0.67 |
|
89 |
+
| SeaLLMs-v3-7B-Chat | 0.55 | 0.52 | 0.67 |
|
90 |
+
| Mistral-7B-Instruct-v0.3 | 0.43 | 0.39 | 0.70 |
|
91 |
+
| Meta-Llama-3-8B-Instruct | 0.27 | 0.21 | 0.80 |
|
92 |
+
| Sailor-7B-Chat | 0.26 | 0.25 | 0.42 |
|
93 |
|
94 |
#### Multi-turn Capatbilities (MT-Bench)
|
95 |
| | **Indonesian** | **Vietnamese** | **English** |
|