Commit
·
e435e5b
1
Parent(s):
71ef33e
Add model and its dependencies
Browse files- config.json +26 -0
- experiments.log +33 -0
- main_config.cfg +27 -0
- merges.txt +0 -0
- pytorch_model.bin +3 -0
- training_args.bin +3 -0
- vocab.json +0 -0
config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "roberta-base",
|
3 |
+
"architectures": [
|
4 |
+
"RobertaForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"eos_token_id": 2,
|
9 |
+
"gradient_checkpointing": false,
|
10 |
+
"hidden_act": "gelu",
|
11 |
+
"hidden_dropout_prob": 0.1,
|
12 |
+
"hidden_size": 768,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 3072,
|
15 |
+
"layer_norm_eps": 1e-05,
|
16 |
+
"max_position_embeddings": 514,
|
17 |
+
"model_type": "roberta",
|
18 |
+
"num_attention_heads": 12,
|
19 |
+
"num_hidden_layers": 12,
|
20 |
+
"pad_token_id": 1,
|
21 |
+
"position_embedding_type": "absolute",
|
22 |
+
"transformers_version": "4.2.2",
|
23 |
+
"type_vocab_size": 4,
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 50265
|
26 |
+
}
|
experiments.log
ADDED
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[56]2021-08-07 10:03:55,527: device: cuda n_gpu: 1
|
2 |
+
[56]2021-08-07 10:06:49,449: ***** Running training *****
|
3 |
+
[56]2021-08-07 10:06:49,449: Batch size = 32
|
4 |
+
[56]2021-08-07 10:06:49,449: Num steps = 15015
|
5 |
+
[56]2021-08-07 11:24:12,835: [epoch 1] ,lr: 1.5e-05 ,tr_loss: 2485.2335910126567
|
6 |
+
[56]2021-08-07 11:24:34,745: ***** Running evaluation *****
|
7 |
+
[56]2021-08-07 11:28:35,164: acc = 0.8122634708956569
|
8 |
+
[56]2021-08-07 11:28:35,164: f1 = 0.3996542284973347
|
9 |
+
[56]2021-08-07 11:28:35,164: precision = 0.46873943899966203
|
10 |
+
[56]2021-08-07 11:28:35,164: recall = 0.3483174284279257
|
11 |
+
[56]2021-08-07 12:44:45,042: [epoch 2] ,lr: 3e-05 ,tr_loss: 1232.1717670541257
|
12 |
+
[56]2021-08-07 12:45:07,200: ***** Running evaluation *****
|
13 |
+
[56]2021-08-07 12:49:07,609: acc = 0.8737159848621373
|
14 |
+
[56]2021-08-07 12:49:07,609: f1 = 0.6822355741979367
|
15 |
+
[56]2021-08-07 12:49:07,609: precision = 0.6218226906385617
|
16 |
+
[56]2021-08-07 12:49:07,609: recall = 0.7556504269211451
|
17 |
+
[56]2021-08-07 14:05:15,094: [epoch 3] ,lr: 0.0 ,tr_loss: 675.0315425428562
|
18 |
+
[56]2021-08-07 14:05:37,318: ***** Running evaluation *****
|
19 |
+
[56]2021-08-07 14:09:37,825: acc = 0.9046224544963056
|
20 |
+
[56]2021-08-07 14:09:37,825: f1 = 0.7179965365658719
|
21 |
+
[56]2021-08-07 14:09:37,825: precision = 0.7645390070921986
|
22 |
+
[56]2021-08-07 14:09:37,825: recall = 0.6767955801104972
|
23 |
+
[56]2021-08-07 14:09:39,472: -----Best Result-----
|
24 |
+
[56]2021-08-07 14:09:39,472: acc = 0.9046224544963056
|
25 |
+
[56]2021-08-07 14:09:39,472: f1 = 0.7179965365658719
|
26 |
+
[56]2021-08-07 14:09:39,472: precision = 0.7645390070921986
|
27 |
+
[56]2021-08-07 14:09:39,472: recall = 0.6767955801104972
|
28 |
+
[56]2021-08-07 14:10:01,830: ***** Running evaluation *****
|
29 |
+
[56]2021-08-07 14:14:02,303: acc = 0.9046224544963056
|
30 |
+
[56]2021-08-07 14:14:02,304: f1 = 0.7179965365658719
|
31 |
+
[56]2021-08-07 14:14:02,304: precision = 0.7645390070921986
|
32 |
+
[56]2021-08-07 14:14:02,304: recall = 0.6767955801104972
|
33 |
+
[56]2021-08-07 14:14:02,304: Saved to saves/roberta-base/2_20210807-1003
|
main_config.cfg
ADDED
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[args]
|
2 |
+
bert_model=roberta-base
|
3 |
+
data_dir=VUA20
|
4 |
+
task_name=vua
|
5 |
+
model_type=MELBERT
|
6 |
+
classifier_hidden=768
|
7 |
+
lr_schedule=warmup_linear
|
8 |
+
warmup_epoch=2
|
9 |
+
drop_ratio=0.2
|
10 |
+
kfold=10
|
11 |
+
num_bagging=0
|
12 |
+
bagging_index=0
|
13 |
+
use_pos=True
|
14 |
+
use_local_context=True
|
15 |
+
max_seq_length=150
|
16 |
+
do_train=True
|
17 |
+
do_test=True
|
18 |
+
do_eval=True
|
19 |
+
do_lower_case=False
|
20 |
+
class_weight=3
|
21 |
+
train_batch_size=32
|
22 |
+
eval_batch_size=8
|
23 |
+
learning_rate=3e-05
|
24 |
+
num_train_epoch=3
|
25 |
+
no_cuda=False
|
26 |
+
seed=42
|
27 |
+
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ceef59e211e065a7c50e7dbb7cb2e3f2ea8ffa1ecc19c5e0646406b880963de5
|
3 |
+
size 508139373
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c4782e64d86a5a2a6eaad688e9cf07aef3df6805304c80e2315c8f8d1258dd57
|
3 |
+
size 1327
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|