alphahg commited on
Commit
abf2aa4
1 Parent(s): 75da275

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -4
README.md CHANGED
@@ -2,6 +2,8 @@
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
 
 
5
  model-index:
6
  - name: koelectra-base-86371428
7
  results: []
@@ -12,7 +14,9 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # koelectra-base-86371428
14
 
15
- This model is a fine-tuned version of [monologg/koelectra-base-v3-discriminator](https://huggingface.co/monologg/koelectra-base-v3-discriminator) on the None dataset.
 
 
16
 
17
  ## Model description
18
 
@@ -31,7 +35,7 @@ More information needed
31
  ### Training hyperparameters
32
 
33
  The following hyperparameters were used during training:
34
- - learning_rate: 0.0005
35
  - train_batch_size: 128
36
  - eval_batch_size: 128
37
  - seed: 30
@@ -39,14 +43,15 @@ The following hyperparameters were used during training:
39
  - total_train_batch_size: 1024
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: cosine
42
- - num_epochs: 1
43
  - mixed_precision_training: Native AMP
44
 
45
  ### Training results
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
- | No log | 0.97 | 22 | 1.8936 |
 
50
 
51
 
52
  ### Framework versions
 
2
  license: apache-2.0
3
  tags:
4
  - generated_from_trainer
5
+ datasets:
6
+ - custom_squad_v2
7
  model-index:
8
  - name: koelectra-base-86371428
9
  results: []
 
14
 
15
  # koelectra-base-86371428
16
 
17
+ This model is a fine-tuned version of [monologg/koelectra-base-v3-discriminator](https://huggingface.co/monologg/koelectra-base-v3-discriminator) on the custom_squad_v2 dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 1.6169
20
 
21
  ## Model description
22
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
+ - learning_rate: 0.0004
39
  - train_batch_size: 128
40
  - eval_batch_size: 128
41
  - seed: 30
 
43
  - total_train_batch_size: 1024
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
+ - num_epochs: 2
47
  - mixed_precision_training: Native AMP
48
 
49
  ### Training results
50
 
51
  | Training Loss | Epoch | Step | Validation Loss |
52
  |:-------------:|:-----:|:----:|:---------------:|
53
+ | No log | 0.94 | 10 | 1.8078 |
54
+ | No log | 1.94 | 20 | 1.6169 |
55
 
56
 
57
  ### Framework versions