tr-aravindan commited on
Commit
490e98c
·
verified ·
1 Parent(s): 69548e0

t5-large_PREFIX_TUNING_SEQ2SEQ_EMOTION

Browse files
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - emotion
8
+ base_model: t5-large
9
+ model-index:
10
+ - name: output_emotion
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # output_emotion
18
+
19
+ This model is a fine-tuned version of [t5-large](https://huggingface.co/t5-large) on the emotion dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 6.9008
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 1e-05
41
+ - train_batch_size: 16
42
+ - eval_batch_size: 16
43
+ - seed: 42
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: linear
46
+ - num_epochs: 4
47
+
48
+ ### Training results
49
+
50
+ | Training Loss | Epoch | Step | Validation Loss |
51
+ |:-------------:|:-----:|:----:|:---------------:|
52
+ | 7.3148 | 2.0 | 1800 | 6.9641 |
53
+ | 7.243 | 4.0 | 3600 | 6.9008 |
54
+
55
+
56
+ ### Framework versions
57
+
58
+ - PEFT 0.7.1
59
+ - Transformers 4.36.2
60
+ - Pytorch 2.0.0
61
+ - Datasets 2.15.0
62
+ - Tokenizers 0.15.0
adapter_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "t5-large",
4
+ "encoder_hidden_size": 1024,
5
+ "inference_mode": true,
6
+ "num_attention_heads": 16,
7
+ "num_layers": 24,
8
+ "num_transformer_submodules": 2,
9
+ "num_virtual_tokens": 20,
10
+ "peft_type": "PREFIX_TUNING",
11
+ "prefix_projection": false,
12
+ "revision": null,
13
+ "task_type": "SEQ_2_SEQ_LM",
14
+ "token_dim": 1024
15
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bebba8b22ac27f2315d4f4a3ed95623b6b40dd4f234b54ab71a7e032b0d4a187
3
+ size 3932288
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:258163b6c372cb200e4a20cd72507fbac270a717e115e0d2c7d6ac2dc5ea4564
3
+ size 4347