pinzhenchen
commited on
Commit
•
0bfc907
1
Parent(s):
b2fcd84
add model in HF format
Browse files
README.md
CHANGED
@@ -9,7 +9,7 @@ license: cc-by-4.0
|
|
9 |
|
10 |
### HPLT MT release v1.0
|
11 |
|
12 |
-
This repository contains the translation model for
|
13 |
|
14 |
### Model Info
|
15 |
|
@@ -18,26 +18,51 @@ This repository contains the translation model for zh_hant-en trained with HPLT
|
|
18 |
* Data: HPLT data only
|
19 |
* Model architecture: Transformer-base
|
20 |
* Tokenizer: SentencePiece (Unigram)
|
21 |
-
* Cleaning: We used OpusCleaner with a set of basic rules. Details can be found in the filter files
|
22 |
|
23 |
-
You can
|
24 |
|
25 |
### Usage
|
26 |
**Note** that for quality considerations, we recommend using `[HPLT/translate-zh_hant-en-v1.0-hplt_opus](https://huggingface.co/HPLT/translate-zh_hant-en-v1.0-hplt_opus)` instead of this model.
|
27 |
|
28 |
-
The model has been trained with
|
29 |
|
30 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
31 |
|
32 |
## Benchmarks
|
33 |
|
34 |
-
|
|
|
|
|
35 |
| -------------------------------------- | ---- | ----- | ----- |
|
36 |
-
|
|
37 |
-
|
|
38 |
|
39 |
### Acknowledgements
|
40 |
|
41 |
This project has received funding from the European Union's Horizon Europe research and innovation programme under grant agreement No 101070350 and from UK Research and Innovation (UKRI) under the UK government's Horizon Europe funding guarantee [grant number 10052546]
|
42 |
|
43 |
-
Brought to you by researchers from the University of Edinburgh
|
|
|
9 |
|
10 |
### HPLT MT release v1.0
|
11 |
|
12 |
+
This repository contains the translation model for Traditional Chinese-English trained with HPLT data only. The model is available in both Marian and Hugging Face formats.
|
13 |
|
14 |
### Model Info
|
15 |
|
|
|
18 |
* Data: HPLT data only
|
19 |
* Model architecture: Transformer-base
|
20 |
* Tokenizer: SentencePiece (Unigram)
|
21 |
+
* Cleaning: We used [OpusCleaner](https://github.com/hplt-project/OpusCleaner) with a set of basic rules. Details can be found in the filter files [here](https://github.com/hplt-project/HPLT-MT-Models/tree/main/v1.0/data/en-zh_hant/raw/v0).
|
22 |
|
23 |
+
You can check out our [deliverable report](https://hplt-project.org/HPLT_D5_1___Translation_models_for_select_language_pairs.pdf), [GitHub repository](https://github.com/hplt-project/HPLT-MT-Models/tree/main/v1.0), and [website](https://hplt-project.org) for more details.
|
24 |
|
25 |
### Usage
|
26 |
**Note** that for quality considerations, we recommend using `[HPLT/translate-zh_hant-en-v1.0-hplt_opus](https://huggingface.co/HPLT/translate-zh_hant-en-v1.0-hplt_opus)` instead of this model.
|
27 |
|
28 |
+
The model has been trained with [MarianNMT](https://github.com/marian-nmt/marian) and the weights are in the Marian format. We have also converted the model into the Hugging Face format so it is compatible with `transformers`.
|
29 |
|
30 |
+
#### Using Marian
|
31 |
+
To run inference with MarianNMT, refer to the [Inference/Decoding/Translation](https://github.com/hplt-project/HPLT-MT-Models/tree/main/v1.0#inferencedecodingtranslation) section of our GitHub repository. You will need the model file `model.npz.best-chrf.npz` and the vocabulary file `model.zh_hant-en.spm` from this repository.
|
32 |
+
|
33 |
+
#### Using transformers
|
34 |
+
We have also converted this model to the Hugging Face format and you can get started with the script below. **Note** that due a [known issue](https://github.com/huggingface/transformers/issues/26216) in weight conversion, the checkpoint cannot work with transformer versions <4.26 or >4.30. We tested and suggest `pip install transformers==4.28`.
|
35 |
+
|
36 |
+
```
|
37 |
+
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
38 |
+
|
39 |
+
tokenizer = AutoTokenizer.from_pretrained("HPLT/translate-zh_hant-en-v1.0-hplt")
|
40 |
+
model = AutoModelForSeq2SeqLM.from_pretrained("HPLT/translate-zh_hant-en-v1.0-hplt")
|
41 |
+
|
42 |
+
inputs = ["Input goes here.", "Make sure the language is right."]
|
43 |
+
batch_tokenized = tokenizer(inputs, return_tensors="pt", padding=True)
|
44 |
+
model_output = model.generate(
|
45 |
+
**batch_tokenized, num_beams=6, max_new_tokens=512
|
46 |
+
)
|
47 |
+
batch_detokenized = tokenizer.batch_decode(
|
48 |
+
model_output,
|
49 |
+
skip_special_tokens=True,
|
50 |
+
)
|
51 |
+
|
52 |
+
print(batch_detokenized)
|
53 |
+
```
|
54 |
|
55 |
## Benchmarks
|
56 |
|
57 |
+
When decoded using Marian, the model has the following test scores.
|
58 |
+
|
59 |
+
| Test set | BLEU | chrF++ | COMET22 |
|
60 |
| -------------------------------------- | ---- | ----- | ----- |
|
61 |
+
| FLORES200 | 20.3 | 47.7 | 0.8182 |
|
62 |
+
| NTREX | 18.2 | 44.9 | 0.79 |
|
63 |
|
64 |
### Acknowledgements
|
65 |
|
66 |
This project has received funding from the European Union's Horizon Europe research and innovation programme under grant agreement No 101070350 and from UK Research and Innovation (UKRI) under the UK government's Horizon Europe funding guarantee [grant number 10052546]
|
67 |
|
68 |
+
Brought to you by researchers from the University of Edinburgh and Charles University in Prague with support from the whole HPLT consortium.
|