amansaini commited on
Commit
5deb3d8
1 Parent(s): 16e7c0a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -2
README.md CHANGED
@@ -4,7 +4,7 @@ license: cc-by-nc-4.0
4
 
5
  # Model Card for Spivavtor-Large
6
 
7
- This model was obtained by fine-tuning the corresponding `bigscience/mt0-large` model on the Spivavtor dataset. All deetails of the dataset used and the fine tuned model can be found in our paper and repository.
8
 
9
  **Paper:** Spivavtor: An Instruction Tuned Ukrainian Text Editing Model
10
 
@@ -14,5 +14,38 @@ This model was obtained by fine-tuning the corresponding `bigscience/mt0-large`
14
 
15
  ### Model Description
16
 
17
- - **Language(s) (NLP)**: Ukrainian
18
  - **Finetuned from model:** bigscience/mt0-large
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
 
5
  # Model Card for Spivavtor-Large
6
 
7
+ This model was obtained by fine-tuning the corresponding `bigscience/mt0-large` model on the Spivavtor dataset. All details of the dataset and fine tuning process can be found in our paper and repository.
8
 
9
  **Paper:** Spivavtor: An Instruction Tuned Ukrainian Text Editing Model
10
 
 
14
 
15
  ### Model Description
16
 
17
+ - **Language**: Ukrainian
18
  - **Finetuned from model:** bigscience/mt0-large
19
+
20
+ ## How to use
21
+ We make available the following models presented in our paper.
22
+
23
+ <table>
24
+ <tr>
25
+ <th>Model</th>
26
+ <th>Number of parameters</th>
27
+ <th>Reference name in Paper</th>
28
+ </tr>
29
+ <tr>
30
+ <td>Spivavtor-large</td>
31
+ <td>1.2B</td>
32
+ <td>Spivavtor-mt0-large</td>
33
+ </tr>
34
+ <tr>
35
+ <td>Spivavtor-xxl</td>
36
+ <td>11B</td>
37
+ <td>Spivavtor-aya-101</td>
38
+ </tr>
39
+ </table>
40
+
41
+ ## Usage
42
+ ```python
43
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
44
+ tokenizer = AutoTokenizer.from_pretrained("grammarly/spivavtor-large")
45
+ model = AutoModelForSeq2SeqLM.from_pretrained("grammarly/spivavtor-large")
46
+
47
+ input_text = 'Виправте граматику в цьому реченнi: Дякую за iнформацiю! ми з Надiєю саме вийшли з дому'
48
+
49
+ input_ids = tokenizer(input_text, return_tensors="pt").input_ids
50
+ outputs = model.generate(input_ids, max_length=256)
51
+ output_text = tokenizer.decode(outputs[0], skip_special_tokens=True)