3728km commited on
Commit
fa5f2e1
·
verified ·
1 Parent(s): 36f39a4

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,5 @@
1
  ---
2
- license: apache-2.0
3
- base_model: distilbert/distilbert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,14 +12,14 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # fined-tune-thai-sentiment
15
 
16
- This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - eval_loss: 0.8230
19
- - eval_accuracy: {'accuracy': 0.6904761904761905}
20
- - eval_f1score: {'f1': 0.5717269740966423}
21
- - eval_runtime: 2.733
22
- - eval_samples_per_second: 46.104
23
- - eval_steps_per_second: 5.854
24
  - step: 0
25
 
26
  ## Model description
@@ -40,12 +39,13 @@ More information needed
40
  ### Training hyperparameters
41
 
42
  The following hyperparameters were used during training:
43
- - learning_rate: 2e-05
44
  - train_batch_size: 8
45
  - eval_batch_size: 8
46
  - seed: 42
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: linear
 
49
  - num_epochs: 5
50
 
51
  ### Framework versions
 
1
  ---
2
+ base_model: airesearch/wangchanberta-base-att-spm-uncased
 
3
  tags:
4
  - generated_from_trainer
5
  model-index:
 
12
 
13
  # fined-tune-thai-sentiment
14
 
15
+ This model is a fine-tuned version of [airesearch/wangchanberta-base-att-spm-uncased](https://huggingface.co/airesearch/wangchanberta-base-att-spm-uncased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - eval_loss: 0.7531
18
+ - eval_accuracy: {'accuracy': 0.7142857142857143}
19
+ - eval_f1score: {'f1': 0.6451576462237443}
20
+ - eval_runtime: 1.1661
21
+ - eval_samples_per_second: 108.049
22
+ - eval_steps_per_second: 13.721
23
  - step: 0
24
 
25
  ## Model description
 
39
  ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
+ - learning_rate: 5e-05
43
  - train_batch_size: 8
44
  - eval_batch_size: 8
45
  - seed: 42
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
+ - lr_scheduler_warmup_steps: 31
49
  - num_epochs: 5
50
 
51
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:078eee3098b41dc0facaa509fd2b689c5709f36ee97901ceef062b918ab76b63
3
  size 421011004
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e8e80ea9f3ed8c1c4be5ed128b50b712c8cd6626b1992ace7ffd1b81da4b951
3
  size 421011004
runs/Jun06_07-20-51_27de48a3f45e/events.out.tfevents.1717658799.27de48a3f45e.445.27 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41cc394c4d88a58014310eb1e9e829619147b2fd77fa0621a6fe320b7e0475b2
3
+ size 297
runs/Jun06_07-27-52_27de48a3f45e/events.out.tfevents.1717658880.27de48a3f45e.445.28 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbd809bdd41ad19882e8a4727e9d7a623418dca38db8ca3b409130742c61317b
3
+ size 7608
runs/Jun06_07-27-52_27de48a3f45e/events.out.tfevents.1717659000.27de48a3f45e.445.29 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d7bd8797c2250f41c0685074ce0339496ad19f82b8221bdad871f534dd83def
3
+ size 297
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da65a6b20ba5ff1d20082e6ab909d7de5c93bda9415975c85ac1ecd3922f9dfe
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:323e9edb43846d53450dce7624ef64c1c7a0e5f0098f3f157b153fa93993cdf7
3
  size 5112