lllaplace
Browse filesThis view is limited to 50 files because it contains too many changes. Β
See raw diff
- outputs/args.json +0 -47
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/args.json +0 -0
- outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/logfile.log +312 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/adapter_config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/all_results.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/all_results_la_kron_all_homo_mc_corr_1000.json +0 -0
- outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/all_results_la_kron_last_layer_homo_mc_corr_1000.json +1 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/eval_res.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/eval_res_la_kron_all_homo_mc_corr_1000.json +0 -0
- outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/eval_res_la_kron_last_layer_homo_mc_corr_1000.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/generation_config.json +0 -0
- outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/gpu_stats.json +7 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/head_config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/pytorch_adapter.bin +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/pytorch_model_head.bin +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/special_tokens_map.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/tokenizer.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/tokenizer_config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/vocab.txt +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/adapter_config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/all_results.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/all_results_la_kron_all_homo_mc_corr_1000.json +0 -0
- outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/all_results_la_kron_last_layer_homo_mc_corr_1000.json +1 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/eval_res.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/eval_res_la_kron_all_homo_mc_corr_1000.json +0 -0
- outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/eval_res_la_kron_last_layer_homo_mc_corr_1000.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/generation_config.json +0 -0
- outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/gpu_stats.json +7 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/head_config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/pytorch_adapter.bin +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/pytorch_model_head.bin +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/special_tokens_map.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/tokenizer.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/tokenizer_config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/vocab.txt +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/adapter_config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/all_results.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/all_results_la_kron_all_homo_mc_corr_1000.json +0 -0
- outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/all_results_la_kron_last_layer_homo_mc_corr_1000.json +1 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/eval_res.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/eval_res_la_kron_all_homo_mc_corr_1000.json +0 -0
- outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/eval_res_la_kron_last_layer_homo_mc_corr_1000.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/generation_config.json +0 -0
- outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/gpu_stats.json +7 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/head_config.json +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/pytorch_adapter.bin +0 -0
- outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/pytorch_model_head.bin +0 -0
outputs/args.json
DELETED
@@ -1,47 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"task_name": "cola",
|
3 |
-
"train_file": null,
|
4 |
-
"validation_file": null,
|
5 |
-
"max_length": 256,
|
6 |
-
"pad_to_max_length": false,
|
7 |
-
"model_name_or_path": "google-bert/bert-base-uncased",
|
8 |
-
"use_slow_tokenizer": true,
|
9 |
-
"per_device_train_batch_size": 8,
|
10 |
-
"per_device_eval_batch_size": 8,
|
11 |
-
"learning_rate": 5e-05,
|
12 |
-
"weight_decay": 0.0,
|
13 |
-
"num_train_epochs": 5,
|
14 |
-
"max_train_steps": null,
|
15 |
-
"gradient_accumulation_steps": 1,
|
16 |
-
"lr_scheduler_type": "linear",
|
17 |
-
"num_warmup_steps": 0,
|
18 |
-
"output_dir": "./outputs",
|
19 |
-
"peft_method": null,
|
20 |
-
"seed": 42,
|
21 |
-
"push_to_hub": false,
|
22 |
-
"hub_model_id": null,
|
23 |
-
"hub_token": null,
|
24 |
-
"checkpointing_steps": "1000",
|
25 |
-
"resume_from_checkpoint": null,
|
26 |
-
"with_tracking": false,
|
27 |
-
"report_to": "all",
|
28 |
-
"ignore_mismatched_sizes": true,
|
29 |
-
"save": false,
|
30 |
-
"load_step": 999,
|
31 |
-
"laplace_hessian": "kron",
|
32 |
-
"laplace_sub": "all",
|
33 |
-
"laplace_prior": "homo",
|
34 |
-
"laplace_optim_step": 1000,
|
35 |
-
"testing_set": "val",
|
36 |
-
"laplace_predict": "mc_corr",
|
37 |
-
"lm_head": false,
|
38 |
-
"cache_dir": "/content/cache/huggingface/metrics/glue",
|
39 |
-
"step_list": [
|
40 |
-
0,
|
41 |
-
1068,
|
42 |
-
2137,
|
43 |
-
3206,
|
44 |
-
4275,
|
45 |
-
5344
|
46 |
-
]
|
47 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/args.json
RENAMED
File without changes
|
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/logfile.log
ADDED
@@ -0,0 +1,312 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
04/30/2024 19:12:31 - INFO - __main__ - Distributed environment: NO
|
2 |
+
Num processes: 1
|
3 |
+
Process index: 0
|
4 |
+
Local process index: 0
|
5 |
+
Device: cpu
|
6 |
+
|
7 |
+
Mixed precision type: no
|
8 |
+
|
9 |
+
04/30/2024 19:12:35 - INFO - __main__ - Number of labels detected = 2
|
10 |
+
04/30/2024 19:12:36 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
|
11 |
+
04/30/2024 19:12:37 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/adapter_config.json
|
12 |
+
04/30/2024 19:12:37 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
|
13 |
+
04/30/2024 19:12:37 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/pytorch_adapter.bin
|
14 |
+
04/30/2024 19:12:37 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/head_config.json
|
15 |
+
04/30/2024 19:12:37 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
|
16 |
+
04/30/2024 19:12:37 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/pytorch_model_head.bin
|
17 |
+
04/30/2024 19:12:37 - INFO - __main__ - Adapter Name = cola
|
18 |
+
04/30/2024 19:12:37 - INFO - __main__ - heads.cola.1.weight
|
19 |
+
04/30/2024 19:12:37 - INFO - __main__ - heads.cola.1.bias
|
20 |
+
04/30/2024 19:12:37 - INFO - __main__ - heads.cola.4.weight
|
21 |
+
04/30/2024 19:12:37 - INFO - __main__ - heads.cola.4.bias
|
22 |
+
04/30/2024 19:12:42 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
|
23 |
+
04/30/2024 19:12:42 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
24 |
+
04/30/2024 19:12:42 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
25 |
+
04/30/2024 19:26:31 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
|
26 |
+
04/30/2024 19:26:31 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
|
27 |
+
04/30/2024 19:26:31 - INFO - __main__ - tensor([[ 0.0687, 0.1107],
|
28 |
+
[ 0.1065, 0.1546],
|
29 |
+
[ 0.1252, 0.1252],
|
30 |
+
...,
|
31 |
+
[ 0.0203, 0.1087],
|
32 |
+
[ 0.0962, 0.1235],
|
33 |
+
[ 0.0511, -0.0174]])
|
34 |
+
04/30/2024 19:26:31 - INFO - __main__ - tensor([[[0.2029, 0.1885],
|
35 |
+
[0.1885, 0.2029]],
|
36 |
+
|
37 |
+
[[0.1999, 0.1717],
|
38 |
+
[0.1717, 0.1999]],
|
39 |
+
|
40 |
+
[[0.1960, 0.1759],
|
41 |
+
[0.1759, 0.1960]],
|
42 |
+
|
43 |
+
...,
|
44 |
+
|
45 |
+
[[0.2176, 0.1939],
|
46 |
+
[0.1939, 0.2176]],
|
47 |
+
|
48 |
+
[[0.1863, 0.1736],
|
49 |
+
[0.1736, 0.1863]],
|
50 |
+
|
51 |
+
[[0.1892, 0.1718],
|
52 |
+
[0.1718, 0.1892]]])
|
53 |
+
04/30/2024 19:26:31 - INFO - __main__ - Distributed environment: NO
|
54 |
+
Num processes: 1
|
55 |
+
Process index: 0
|
56 |
+
Local process index: 0
|
57 |
+
Device: cpu
|
58 |
+
|
59 |
+
Mixed precision type: no
|
60 |
+
|
61 |
+
04/30/2024 19:26:35 - INFO - __main__ - Number of labels detected = 2
|
62 |
+
04/30/2024 19:26:35 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
|
63 |
+
04/30/2024 19:26:37 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/adapter_config.json
|
64 |
+
04/30/2024 19:26:37 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
|
65 |
+
04/30/2024 19:26:37 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/pytorch_adapter.bin
|
66 |
+
04/30/2024 19:26:37 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/head_config.json
|
67 |
+
04/30/2024 19:26:37 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
|
68 |
+
04/30/2024 19:26:37 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/pytorch_model_head.bin
|
69 |
+
04/30/2024 19:26:37 - INFO - __main__ - Adapter Name = cola
|
70 |
+
04/30/2024 19:26:37 - INFO - __main__ - heads.cola.1.weight
|
71 |
+
04/30/2024 19:26:37 - INFO - __main__ - heads.cola.1.bias
|
72 |
+
04/30/2024 19:26:37 - INFO - __main__ - heads.cola.4.weight
|
73 |
+
04/30/2024 19:26:37 - INFO - __main__ - heads.cola.4.bias
|
74 |
+
04/30/2024 19:26:41 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
|
75 |
+
04/30/2024 19:26:41 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
76 |
+
04/30/2024 19:26:41 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
77 |
+
04/30/2024 19:40:51 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
|
78 |
+
04/30/2024 19:40:51 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
|
79 |
+
04/30/2024 19:40:51 - INFO - __main__ - tensor([[-1.4529, 1.7744],
|
80 |
+
[-1.1889, 1.5487],
|
81 |
+
[-1.4239, 1.8257],
|
82 |
+
...,
|
83 |
+
[-2.0956, 2.2813],
|
84 |
+
[-0.5643, 0.8205],
|
85 |
+
[-0.9262, 0.8999]])
|
86 |
+
04/30/2024 19:40:51 - INFO - __main__ - tensor([[[0.3071, 0.2560],
|
87 |
+
[0.2560, 0.3071]],
|
88 |
+
|
89 |
+
[[0.2708, 0.2068],
|
90 |
+
[0.2068, 0.2708]],
|
91 |
+
|
92 |
+
[[0.3089, 0.2527],
|
93 |
+
[0.2527, 0.3089]],
|
94 |
+
|
95 |
+
...,
|
96 |
+
|
97 |
+
[[0.4054, 0.3471],
|
98 |
+
[0.3471, 0.4054]],
|
99 |
+
|
100 |
+
[[0.1919, 0.1553],
|
101 |
+
[0.1553, 0.1919]],
|
102 |
+
|
103 |
+
[[0.2052, 0.1670],
|
104 |
+
[0.1670, 0.2052]]])
|
105 |
+
04/30/2024 19:40:51 - INFO - __main__ - Distributed environment: NO
|
106 |
+
Num processes: 1
|
107 |
+
Process index: 0
|
108 |
+
Local process index: 0
|
109 |
+
Device: cpu
|
110 |
+
|
111 |
+
Mixed precision type: no
|
112 |
+
|
113 |
+
04/30/2024 19:40:53 - INFO - __main__ - Number of labels detected = 2
|
114 |
+
04/30/2024 19:40:54 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
|
115 |
+
04/30/2024 19:40:55 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/adapter_config.json
|
116 |
+
04/30/2024 19:40:55 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
|
117 |
+
04/30/2024 19:40:55 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/pytorch_adapter.bin
|
118 |
+
04/30/2024 19:40:55 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/head_config.json
|
119 |
+
04/30/2024 19:40:55 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
|
120 |
+
04/30/2024 19:40:55 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/pytorch_model_head.bin
|
121 |
+
04/30/2024 19:40:55 - INFO - __main__ - Adapter Name = cola
|
122 |
+
04/30/2024 19:40:55 - INFO - __main__ - heads.cola.1.weight
|
123 |
+
04/30/2024 19:40:55 - INFO - __main__ - heads.cola.1.bias
|
124 |
+
04/30/2024 19:40:55 - INFO - __main__ - heads.cola.4.weight
|
125 |
+
04/30/2024 19:40:55 - INFO - __main__ - heads.cola.4.bias
|
126 |
+
04/30/2024 19:41:00 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
|
127 |
+
04/30/2024 19:41:00 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
128 |
+
04/30/2024 19:41:00 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
129 |
+
04/30/2024 19:55:00 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
|
130 |
+
04/30/2024 19:55:00 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
|
131 |
+
04/30/2024 19:55:00 - INFO - __main__ - tensor([[-1.2545, 1.5103],
|
132 |
+
[-1.1312, 1.4272],
|
133 |
+
[-1.0637, 1.4316],
|
134 |
+
...,
|
135 |
+
[-1.9212, 2.1010],
|
136 |
+
[-0.2051, 0.3802],
|
137 |
+
[-0.7883, 0.7410]])
|
138 |
+
04/30/2024 19:55:00 - INFO - __main__ - tensor([[[0.2652, 0.2093],
|
139 |
+
[0.2093, 0.2652]],
|
140 |
+
|
141 |
+
[[0.2405, 0.1822],
|
142 |
+
[0.1822, 0.2405]],
|
143 |
+
|
144 |
+
[[0.2576, 0.2011],
|
145 |
+
[0.2011, 0.2576]],
|
146 |
+
|
147 |
+
...,
|
148 |
+
|
149 |
+
[[0.3661, 0.3102],
|
150 |
+
[0.3102, 0.3661]],
|
151 |
+
|
152 |
+
[[0.1545, 0.1238],
|
153 |
+
[0.1238, 0.1545]],
|
154 |
+
|
155 |
+
[[0.1725, 0.1392],
|
156 |
+
[0.1392, 0.1725]]])
|
157 |
+
04/30/2024 19:55:00 - INFO - __main__ - Distributed environment: NO
|
158 |
+
Num processes: 1
|
159 |
+
Process index: 0
|
160 |
+
Local process index: 0
|
161 |
+
Device: cpu
|
162 |
+
|
163 |
+
Mixed precision type: no
|
164 |
+
|
165 |
+
04/30/2024 19:55:02 - INFO - __main__ - Number of labels detected = 2
|
166 |
+
04/30/2024 19:55:03 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
|
167 |
+
04/30/2024 19:55:04 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_3206/adapter_config.json
|
168 |
+
04/30/2024 19:55:04 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
|
169 |
+
04/30/2024 19:55:04 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_3206/pytorch_adapter.bin
|
170 |
+
04/30/2024 19:55:04 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_3206/head_config.json
|
171 |
+
04/30/2024 19:55:04 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
|
172 |
+
04/30/2024 19:55:04 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_3206/pytorch_model_head.bin
|
173 |
+
04/30/2024 19:55:04 - INFO - __main__ - Adapter Name = cola
|
174 |
+
04/30/2024 19:55:04 - INFO - __main__ - heads.cola.1.weight
|
175 |
+
04/30/2024 19:55:04 - INFO - __main__ - heads.cola.1.bias
|
176 |
+
04/30/2024 19:55:04 - INFO - __main__ - heads.cola.4.weight
|
177 |
+
04/30/2024 19:55:04 - INFO - __main__ - heads.cola.4.bias
|
178 |
+
04/30/2024 19:55:09 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
|
179 |
+
04/30/2024 19:55:09 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
180 |
+
04/30/2024 19:55:09 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
181 |
+
04/30/2024 20:09:20 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
|
182 |
+
04/30/2024 20:09:20 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
|
183 |
+
04/30/2024 20:09:20 - INFO - __main__ - tensor([[-1.5903, 1.8046],
|
184 |
+
[-1.4262, 1.7163],
|
185 |
+
[-1.2728, 1.6234],
|
186 |
+
...,
|
187 |
+
[-2.3189, 2.5137],
|
188 |
+
[-0.0952, 0.2233],
|
189 |
+
[-0.7922, 0.7461]])
|
190 |
+
04/30/2024 20:09:20 - INFO - __main__ - tensor([[[0.3187, 0.2535],
|
191 |
+
[0.2535, 0.3187]],
|
192 |
+
|
193 |
+
[[0.2909, 0.2243],
|
194 |
+
[0.2243, 0.2909]],
|
195 |
+
|
196 |
+
[[0.2970, 0.2309],
|
197 |
+
[0.2309, 0.2970]],
|
198 |
+
|
199 |
+
...,
|
200 |
+
|
201 |
+
[[0.4427, 0.3839],
|
202 |
+
[0.3839, 0.4427]],
|
203 |
+
|
204 |
+
[[0.1634, 0.1292],
|
205 |
+
[0.1292, 0.1634]],
|
206 |
+
|
207 |
+
[[0.1837, 0.1443],
|
208 |
+
[0.1443, 0.1837]]])
|
209 |
+
04/30/2024 20:09:20 - INFO - __main__ - Distributed environment: NO
|
210 |
+
Num processes: 1
|
211 |
+
Process index: 0
|
212 |
+
Local process index: 0
|
213 |
+
Device: cpu
|
214 |
+
|
215 |
+
Mixed precision type: no
|
216 |
+
|
217 |
+
04/30/2024 20:09:22 - INFO - __main__ - Number of labels detected = 2
|
218 |
+
04/30/2024 20:09:23 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
|
219 |
+
04/30/2024 20:09:23 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_4275/adapter_config.json
|
220 |
+
04/30/2024 20:09:23 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
|
221 |
+
04/30/2024 20:09:23 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_4275/pytorch_adapter.bin
|
222 |
+
04/30/2024 20:09:23 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_4275/head_config.json
|
223 |
+
04/30/2024 20:09:23 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
|
224 |
+
04/30/2024 20:09:24 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_4275/pytorch_model_head.bin
|
225 |
+
04/30/2024 20:09:24 - INFO - __main__ - Adapter Name = cola
|
226 |
+
04/30/2024 20:09:24 - INFO - __main__ - heads.cola.1.weight
|
227 |
+
04/30/2024 20:09:24 - INFO - __main__ - heads.cola.1.bias
|
228 |
+
04/30/2024 20:09:24 - INFO - __main__ - heads.cola.4.weight
|
229 |
+
04/30/2024 20:09:24 - INFO - __main__ - heads.cola.4.bias
|
230 |
+
04/30/2024 20:09:28 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
|
231 |
+
04/30/2024 20:09:28 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
232 |
+
04/30/2024 20:09:28 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
233 |
+
04/30/2024 20:23:30 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
|
234 |
+
04/30/2024 20:23:30 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
|
235 |
+
04/30/2024 20:23:30 - INFO - __main__ - tensor([[-1.8506, 2.0807],
|
236 |
+
[-1.5632, 1.8705],
|
237 |
+
[-1.4465, 1.8179],
|
238 |
+
...,
|
239 |
+
[-2.5331, 2.7452],
|
240 |
+
[-0.1701, 0.3093],
|
241 |
+
[-0.9411, 0.9126]])
|
242 |
+
04/30/2024 20:23:30 - INFO - __main__ - tensor([[[0.3666, 0.2971],
|
243 |
+
[0.2971, 0.3666]],
|
244 |
+
|
245 |
+
[[0.3195, 0.2491],
|
246 |
+
[0.2491, 0.3195]],
|
247 |
+
|
248 |
+
[[0.3326, 0.2601],
|
249 |
+
[0.2601, 0.3326]],
|
250 |
+
|
251 |
+
...,
|
252 |
+
|
253 |
+
[[0.4860, 0.4264],
|
254 |
+
[0.4264, 0.4860]],
|
255 |
+
|
256 |
+
[[0.1715, 0.1336],
|
257 |
+
[0.1336, 0.1715]],
|
258 |
+
|
259 |
+
[[0.2016, 0.1572],
|
260 |
+
[0.1572, 0.2016]]])
|
261 |
+
04/30/2024 20:23:30 - INFO - __main__ - Distributed environment: NO
|
262 |
+
Num processes: 1
|
263 |
+
Process index: 0
|
264 |
+
Local process index: 0
|
265 |
+
Device: cpu
|
266 |
+
|
267 |
+
Mixed precision type: no
|
268 |
+
|
269 |
+
04/30/2024 20:23:32 - INFO - __main__ - Number of labels detected = 2
|
270 |
+
04/30/2024 20:23:33 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
|
271 |
+
04/30/2024 20:23:34 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_5344/adapter_config.json
|
272 |
+
04/30/2024 20:23:34 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
|
273 |
+
04/30/2024 20:23:34 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_5344/pytorch_adapter.bin
|
274 |
+
04/30/2024 20:23:34 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_5344/head_config.json
|
275 |
+
04/30/2024 20:23:34 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
|
276 |
+
04/30/2024 20:23:34 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_5344/pytorch_model_head.bin
|
277 |
+
04/30/2024 20:23:34 - INFO - __main__ - Adapter Name = cola
|
278 |
+
04/30/2024 20:23:34 - INFO - __main__ - heads.cola.1.weight
|
279 |
+
04/30/2024 20:23:34 - INFO - __main__ - heads.cola.1.bias
|
280 |
+
04/30/2024 20:23:34 - INFO - __main__ - heads.cola.4.weight
|
281 |
+
04/30/2024 20:23:34 - INFO - __main__ - heads.cola.4.bias
|
282 |
+
04/30/2024 20:23:39 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
|
283 |
+
04/30/2024 20:23:39 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
284 |
+
04/30/2024 20:23:39 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
285 |
+
04/30/2024 20:37:44 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
|
286 |
+
04/30/2024 20:37:44 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
|
287 |
+
04/30/2024 20:37:44 - INFO - __main__ - tensor([[-1.8640, 2.1037],
|
288 |
+
[-1.5643, 1.8815],
|
289 |
+
[-1.4149, 1.7945],
|
290 |
+
...,
|
291 |
+
[-2.4889, 2.7067],
|
292 |
+
[-0.1663, 0.3097],
|
293 |
+
[-0.8635, 0.8440]])
|
294 |
+
04/30/2024 20:37:44 - INFO - __main__ - tensor([[[0.3703, 0.3002],
|
295 |
+
[0.3002, 0.3703]],
|
296 |
+
|
297 |
+
[[0.3209, 0.2497],
|
298 |
+
[0.2497, 0.3209]],
|
299 |
+
|
300 |
+
[[0.3293, 0.2559],
|
301 |
+
[0.2559, 0.3293]],
|
302 |
+
|
303 |
+
...,
|
304 |
+
|
305 |
+
[[0.4776, 0.4180],
|
306 |
+
[0.4180, 0.4776]],
|
307 |
+
|
308 |
+
[[0.1715, 0.1335],
|
309 |
+
[0.1335, 0.1715]],
|
310 |
+
|
311 |
+
[[0.1975, 0.1524],
|
312 |
+
[0.1524, 0.1975]]])
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/adapter_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/all_results.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/all_results_la_kron_all_homo_mc_corr_1000.json
RENAMED
File without changes
|
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/all_results_la_kron_last_layer_homo_mc_corr_1000.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_matthews_correlation": 0.01029747168884776}
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/eval_res.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/eval_res_la_kron_all_homo_mc_corr_1000.json
RENAMED
File without changes
|
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/eval_res_la_kron_last_layer_homo_mc_corr_1000.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/generation_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/gpu_stats.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"memory_allocated": 0,
|
3 |
+
"max_memory_allocated": 0,
|
4 |
+
"memory_reserved": 0,
|
5 |
+
"max_memory_reserved": 0,
|
6 |
+
"memory_stats": {}
|
7 |
+
}
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/head_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/pytorch_adapter.bin
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/pytorch_model_head.bin
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/special_tokens_map.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/tokenizer.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/tokenizer_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/vocab.txt
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/adapter_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/all_results.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/all_results_la_kron_all_homo_mc_corr_1000.json
RENAMED
File without changes
|
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/all_results_la_kron_last_layer_homo_mc_corr_1000.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_matthews_correlation": 0.46484295300587036}
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/eval_res.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/eval_res_la_kron_all_homo_mc_corr_1000.json
RENAMED
File without changes
|
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/eval_res_la_kron_last_layer_homo_mc_corr_1000.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/generation_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/gpu_stats.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"memory_allocated": 0,
|
3 |
+
"max_memory_allocated": 0,
|
4 |
+
"memory_reserved": 0,
|
5 |
+
"max_memory_reserved": 0,
|
6 |
+
"memory_stats": {}
|
7 |
+
}
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/head_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/pytorch_adapter.bin
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/pytorch_model_head.bin
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/special_tokens_map.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/tokenizer.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/tokenizer_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/vocab.txt
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/adapter_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/all_results.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/all_results_la_kron_all_homo_mc_corr_1000.json
RENAMED
File without changes
|
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/all_results_la_kron_last_layer_homo_mc_corr_1000.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_matthews_correlation": 0.5072753983504138}
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/eval_res.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/eval_res_la_kron_all_homo_mc_corr_1000.json
RENAMED
File without changes
|
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/eval_res_la_kron_last_layer_homo_mc_corr_1000.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/generation_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/gpu_stats.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"memory_allocated": 0,
|
3 |
+
"max_memory_allocated": 0,
|
4 |
+
"memory_reserved": 0,
|
5 |
+
"max_memory_reserved": 0,
|
6 |
+
"memory_stats": {}
|
7 |
+
}
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/head_config.json
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/pytorch_adapter.bin
RENAMED
File without changes
|
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/pytorch_model_head.bin
RENAMED
File without changes
|