Remek commited on
Commit
b2ed531
·
verified ·
1 Parent(s): 6b91924

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -52,7 +52,7 @@ To align the model with user preferences we tested many different techniques: DP
52
 
53
  Bielik instruct models have been trained with the use of an original open source framework called [ALLaMo](https://github.com/chrisociepa/allamo) implemented by [Krzysztof Ociepa](https://www.linkedin.com/in/krzysztof-ociepa-44886550/). This framework allows users to train language models with architecture similar to LLaMA and Mistral in fast and efficient way.
54
 
55
- Bielik-11B-v2.3-Instruct is a linear merge of the [Bielik-11B-v2.0-Instruct](https://huggingface.co/speakleash/Bielik-11B-v2.0-Instruct), [Bielik-11B-v2.1-Instruct](https://huggingface.co/speakleash/Bielik-11B-v2.1-Instruct), and [Bielik-11B-v2.2-Instruct](https://huggingface.co/speakleash/Bielik-11B-v2.2-Instruct) models, each contributing equally with a weight of 1.0. The merge was performed in float16 precision by [Remigiusz Kinas](https://www.linkedin.com/in/remigiusz-kinas/) using [mergekit](https://github.com/cg123/mergekit).
56
 
57
  ### Model description:
58
 
@@ -316,7 +316,7 @@ This benchmark provides a robust and time-efficient method for assessing LLM per
316
  | Bielik-11B-v2.0-Instruct | 72.10 | 40.20 |
317
  | Mistral-7B-Instruct-v0.2 | 70.00 | 36.20 |
318
 
319
- The results show that Bielik-11B-v2.3-Instruct performs well on the MixEval benchmark, achieving a score of ~~74.55~~ on the standard MixEval and ~~45.00~~ on MixEval-Hard. Notably, Bielik-11B-v2.3-Instruct significantly outperforms Mistral-7B-Instruct-v0.2 on both metrics, demonstrating its improved capabilities despite being based on a similar architecture.
320
 
321
 
322
  ## Limitations and Biases
 
52
 
53
  Bielik instruct models have been trained with the use of an original open source framework called [ALLaMo](https://github.com/chrisociepa/allamo) implemented by [Krzysztof Ociepa](https://www.linkedin.com/in/krzysztof-ociepa-44886550/). This framework allows users to train language models with architecture similar to LLaMA and Mistral in fast and efficient way.
54
 
55
+ Bielik-11B-v2.3-Instruct is a merge of the [Bielik-11B-v2.0-Instruct](https://huggingface.co/speakleash/Bielik-11B-v2.0-Instruct), [Bielik-11B-v2.1-Instruct](https://huggingface.co/speakleash/Bielik-11B-v2.1-Instruct), and [Bielik-11B-v2.2-Instruct](https://huggingface.co/speakleash/Bielik-11B-v2.2-Instruct) models. The merge was performed in float16 precision by [Remigiusz Kinas](https://www.linkedin.com/in/remigiusz-kinas/) using [mergekit](https://github.com/cg123/mergekit).
56
 
57
  ### Model description:
58
 
 
316
  | Bielik-11B-v2.0-Instruct | 72.10 | 40.20 |
317
  | Mistral-7B-Instruct-v0.2 | 70.00 | 36.20 |
318
 
319
+ The results show that Bielik-11B-v2.3-Instruct performs well on the MixEval benchmark, achieving a score of 72.95 on the standard MixEval and 43.20 on MixEval-Hard. Notably, Bielik-11B-v2.3-Instruct significantly outperforms Mistral-7B-Instruct-v0.2 on both metrics, demonstrating its improved capabilities despite being based on a similar architecture.
320
 
321
 
322
  ## Limitations and Biases