Update README.md
Browse files
README.md
CHANGED
@@ -16,4 +16,22 @@ metrics:
|
|
16 |
pipeline_tag: translation
|
17 |
---
|
18 |
|
19 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
pipeline_tag: translation
|
17 |
---
|
18 |
|
19 |
+
# Plume32k
|
20 |
+
|
21 |
+
## Table of Contents
|
22 |
+
<details>
|
23 |
+
<summary>Click to expand</summary>
|
24 |
+
|
25 |
+
- [Model description](#model-description)
|
26 |
+
- [Intended uses and limitations](#intended-uses-and-limitations)
|
27 |
+
- [How to use](#how-to-use)
|
28 |
+
- [Training](#training)
|
29 |
+
- [Evaluation](#evaluation)
|
30 |
+
- [Citation](#citation)
|
31 |
+
- [Additional information](#additional-information)
|
32 |
+
|
33 |
+
</details>
|
34 |
+
|
35 |
+
## Summary
|
36 |
+
|
37 |
+
Plume is the first LLM trained for Neural Machine Translation with only parallel Catalan-Centric data from scratch. It is a language model with the same architecture as Gemma 2B. The model is trained for general translation tasks at sentence level. For more information about training, architecture and interpretability of the model check out the paper; "Investigating the translation capabilities of Large Language Models trained on parallel data only". The preprint is available on [arXiv]().
|