Shekarss commited on
Commit
33d6cec
·
verified ·
1 Parent(s): 581467b

Training complete

Browse files
README.md ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ base_model: google/mt5-small
4
+ tags:
5
+ - summarization
6
+ - generated_from_trainer
7
+ metrics:
8
+ - rouge
9
+ model-index:
10
+ - name: mt5-small-finetuned-amazon-en-es
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # mt5-small-finetuned-amazon-en-es
18
+
19
+ This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on the None dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 3.0056
22
+ - Rouge1: 16.1556
23
+ - Rouge2: 8.8085
24
+ - Rougel: 15.8019
25
+ - Rougelsum: 15.7787
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 5.6e-05
45
+ - train_batch_size: 8
46
+ - eval_batch_size: 8
47
+ - seed: 42
48
+ - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
49
+ - lr_scheduler_type: linear
50
+ - num_epochs: 8
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
55
+ |:-------------:|:-----:|:-----:|:---------------:|:-------:|:------:|:-------:|:---------:|
56
+ | 6.9474 | 1.0 | 1399 | 3.3187 | 11.4251 | 4.57 | 10.9308 | 10.7649 |
57
+ | 3.892 | 2.0 | 2798 | 3.1469 | 13.4445 | 6.2409 | 13.013 | 12.9006 |
58
+ | 3.5863 | 3.0 | 4197 | 3.0734 | 15.1812 | 8.3203 | 14.7922 | 14.7632 |
59
+ | 3.4229 | 4.0 | 5596 | 3.0587 | 16.1014 | 8.4524 | 15.8047 | 15.7134 |
60
+ | 3.326 | 5.0 | 6995 | 3.0234 | 17.153 | 9.3495 | 16.6487 | 16.564 |
61
+ | 3.2622 | 6.0 | 8394 | 3.0144 | 15.5588 | 8.1691 | 15.275 | 15.2065 |
62
+ | 3.2163 | 7.0 | 9793 | 3.0139 | 15.8487 | 8.6641 | 15.548 | 15.511 |
63
+ | 3.1948 | 8.0 | 11192 | 3.0056 | 16.1556 | 8.8085 | 15.8019 | 15.7787 |
64
+
65
+
66
+ ### Framework versions
67
+
68
+ - Transformers 4.55.4
69
+ - Pytorch 2.8.0+cu128
70
+ - Datasets 4.0.0
71
+ - Tokenizers 0.21.4
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "decoder_start_token_id": 0,
3
+ "eos_token_id": 1,
4
+ "pad_token_id": 0,
5
+ "transformers_version": "4.55.4"
6
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3fe3819e5f06e15f902e0244b435b64df08c01d773474a58f2bb1ca20b3efeb8
3
  size 1200729512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29c906ea0b2d2012bfe07a85879a2b6bc684782dbc1a999076f2233a1e307ed3
3
  size 1200729512
runs/Aug25_23-13-19_Predator/events.out.tfevents.1756143803.Predator.4938.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0cefaf00cc4877c228ed3d2f3a9a6e77f09e5b8acc937b7b42a04be1737a2651
3
- size 10059
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e3935e85afe3b97bce9d8d928eed2d33588735bbab677ef2ae26e2575e85ba6
3
+ size 11098
runs/Aug25_23-13-19_Predator/events.out.tfevents.1756147083.Predator.4938.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b96fe84efffcfd3f0d53b29214ee5ee494ad27f6fc895228c13f13c67283be85
3
+ size 562