javi8979 commited on
Commit
b84f4c9
·
verified ·
1 Parent(s): d83f0d8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -1
README.md CHANGED
@@ -16,4 +16,22 @@ metrics:
16
  pipeline_tag: translation
17
  ---
18
 
19
- # Model Card for Plume32k
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  pipeline_tag: translation
17
  ---
18
 
19
+ # Plume32k
20
+
21
+ ## Table of Contents
22
+ <details>
23
+ <summary>Click to expand</summary>
24
+
25
+ - [Model description](#model-description)
26
+ - [Intended uses and limitations](#intended-uses-and-limitations)
27
+ - [How to use](#how-to-use)
28
+ - [Training](#training)
29
+ - [Evaluation](#evaluation)
30
+ - [Citation](#citation)
31
+ - [Additional information](#additional-information)
32
+
33
+ </details>
34
+
35
+ ## Summary
36
+
37
+ Plume is the first LLM trained for Neural Machine Translation with only parallel Catalan-Centric data from scratch. It is a language model with the same architecture as Gemma 2B. The model is trained for general translation tasks at sentence level. For more information about training, architecture and interpretability of the model check out the paper; "Investigating the translation capabilities of Large Language Models trained on parallel data only". The preprint is available on [arXiv]().