Update README.md
Browse files
    	
        README.md
    CHANGED
    
    | @@ -13,6 +13,8 @@ pipeline_tag: fill-mask | |
| 13 | 
             
            <!-- Provide a quick summary of what the model is/does. -->
         | 
| 14 | 
             
            LiBERTa Large is a BERT-like model pre-trained from scratch exclusively for Ukrainian. It was presented during the [UNLP](https://unlp.org.ua/) @ [LREC-COLING 2024](https://lrec-coling-2024.org/). Further details are in the [LiBERTa: Advancing Ukrainian Language Modeling through Pre-training from Scratch](https://aclanthology.org/2024.unlp-1.14/) paper.
         | 
| 15 |  | 
|  | |
|  | |
| 16 |  | 
| 17 | 
             
            ## Evaluation
         | 
| 18 |  | 
| @@ -32,6 +34,16 @@ Read the [paper](https://aclanthology.org/2024.unlp-1.14/) for more detailed tas | |
| 32 | 
             
            | [liberta-large](https://huggingface.co/Goader/liberta-large)                                                            | 91.27 (1.22)             | 92.50 (0.07)       | 98.62 (0.08)                   | 95.44 (0.04)                             |
         | 
| 33 | 
             
            | [liberta-large-v2](https://huggingface.co/Goader/liberta-large-v2)                                                      | __91.73 (1.81)__         | __93.22 (0.14)__   | __98.79 (0.06)__               | 95.67 (0.12)                             |
         | 
| 34 |  | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 35 |  | 
| 36 |  | 
| 37 | 
             
            ## How to Get Started with the Model
         | 
|  | |
| 13 | 
             
            <!-- Provide a quick summary of what the model is/does. -->
         | 
| 14 | 
             
            LiBERTa Large is a BERT-like model pre-trained from scratch exclusively for Ukrainian. It was presented during the [UNLP](https://unlp.org.ua/) @ [LREC-COLING 2024](https://lrec-coling-2024.org/). Further details are in the [LiBERTa: Advancing Ukrainian Language Modeling through Pre-training from Scratch](https://aclanthology.org/2024.unlp-1.14/) paper.
         | 
| 15 |  | 
| 16 | 
            +
            All the code is available in the [Goader/ukr-lm](https://github.com/Goader/ukr-lm) repository.
         | 
| 17 | 
            +
             | 
| 18 |  | 
| 19 | 
             
            ## Evaluation
         | 
| 20 |  | 
|  | |
| 34 | 
             
            | [liberta-large](https://huggingface.co/Goader/liberta-large)                                                            | 91.27 (1.22)             | 92.50 (0.07)       | 98.62 (0.08)                   | 95.44 (0.04)                             |
         | 
| 35 | 
             
            | [liberta-large-v2](https://huggingface.co/Goader/liberta-large-v2)                                                      | __91.73 (1.81)__         | __93.22 (0.14)__   | __98.79 (0.06)__               | 95.67 (0.12)                             |
         | 
| 36 |  | 
| 37 | 
            +
            ## Fine-Tuning Hyperparameters
         | 
| 38 | 
            +
             | 
| 39 | 
            +
            | Hyperparameter | Value |
         | 
| 40 | 
            +
            |:---------------|:-----:|
         | 
| 41 | 
            +
            | Peak Learning Rate  | 3e-5   |
         | 
| 42 | 
            +
            | Warm-up Ratio       | 0.05   |
         | 
| 43 | 
            +
            | Learning Rate Decay | Linear |
         | 
| 44 | 
            +
            | Batch Size          | 16     |
         | 
| 45 | 
            +
            | Epochs              | 10     |
         | 
| 46 | 
            +
            | Weight Decay        | 0.05   |
         | 
| 47 |  | 
| 48 |  | 
| 49 | 
             
            ## How to Get Started with the Model
         | 
