jdorairaj commited on
Commit
d61c3e4
Β·
1 Parent(s): a479626
This view is limited to 50 files because it contains too many changes. Β  See raw diff
Files changed (50) hide show
  1. outputs/args.json +0 -47
  2. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/args.json +0 -0
  3. outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/logfile.log +312 -0
  4. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/adapter_config.json +0 -0
  5. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/all_results.json +0 -0
  6. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/all_results_la_kron_all_homo_mc_corr_1000.json +0 -0
  7. outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/all_results_la_kron_last_layer_homo_mc_corr_1000.json +1 -0
  8. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/config.json +0 -0
  9. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/eval_res.json +0 -0
  10. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/eval_res_la_kron_all_homo_mc_corr_1000.json +0 -0
  11. outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/eval_res_la_kron_last_layer_homo_mc_corr_1000.json +0 -0
  12. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/generation_config.json +0 -0
  13. outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/gpu_stats.json +7 -0
  14. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/head_config.json +0 -0
  15. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/pytorch_adapter.bin +0 -0
  16. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/pytorch_model_head.bin +0 -0
  17. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/special_tokens_map.json +0 -0
  18. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/tokenizer.json +0 -0
  19. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/tokenizer_config.json +0 -0
  20. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/vocab.txt +0 -0
  21. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/adapter_config.json +0 -0
  22. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/all_results.json +0 -0
  23. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/all_results_la_kron_all_homo_mc_corr_1000.json +0 -0
  24. outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/all_results_la_kron_last_layer_homo_mc_corr_1000.json +1 -0
  25. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/config.json +0 -0
  26. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/eval_res.json +0 -0
  27. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/eval_res_la_kron_all_homo_mc_corr_1000.json +0 -0
  28. outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/eval_res_la_kron_last_layer_homo_mc_corr_1000.json +0 -0
  29. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/generation_config.json +0 -0
  30. outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/gpu_stats.json +7 -0
  31. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/head_config.json +0 -0
  32. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/pytorch_adapter.bin +0 -0
  33. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/pytorch_model_head.bin +0 -0
  34. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/special_tokens_map.json +0 -0
  35. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/tokenizer.json +0 -0
  36. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/tokenizer_config.json +0 -0
  37. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/vocab.txt +0 -0
  38. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/adapter_config.json +0 -0
  39. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/all_results.json +0 -0
  40. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/all_results_la_kron_all_homo_mc_corr_1000.json +0 -0
  41. outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/all_results_la_kron_last_layer_homo_mc_corr_1000.json +1 -0
  42. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/config.json +0 -0
  43. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/eval_res.json +0 -0
  44. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/eval_res_la_kron_all_homo_mc_corr_1000.json +0 -0
  45. outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/eval_res_la_kron_last_layer_homo_mc_corr_1000.json +0 -0
  46. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/generation_config.json +0 -0
  47. outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/gpu_stats.json +7 -0
  48. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/head_config.json +0 -0
  49. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/pytorch_adapter.bin +0 -0
  50. outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/pytorch_model_head.bin +0 -0
outputs/args.json DELETED
@@ -1,47 +0,0 @@
1
- {
2
- "task_name": "cola",
3
- "train_file": null,
4
- "validation_file": null,
5
- "max_length": 256,
6
- "pad_to_max_length": false,
7
- "model_name_or_path": "google-bert/bert-base-uncased",
8
- "use_slow_tokenizer": true,
9
- "per_device_train_batch_size": 8,
10
- "per_device_eval_batch_size": 8,
11
- "learning_rate": 5e-05,
12
- "weight_decay": 0.0,
13
- "num_train_epochs": 5,
14
- "max_train_steps": null,
15
- "gradient_accumulation_steps": 1,
16
- "lr_scheduler_type": "linear",
17
- "num_warmup_steps": 0,
18
- "output_dir": "./outputs",
19
- "peft_method": null,
20
- "seed": 42,
21
- "push_to_hub": false,
22
- "hub_model_id": null,
23
- "hub_token": null,
24
- "checkpointing_steps": "1000",
25
- "resume_from_checkpoint": null,
26
- "with_tracking": false,
27
- "report_to": "all",
28
- "ignore_mismatched_sizes": true,
29
- "save": false,
30
- "load_step": 999,
31
- "laplace_hessian": "kron",
32
- "laplace_sub": "all",
33
- "laplace_prior": "homo",
34
- "laplace_optim_step": 1000,
35
- "testing_set": "val",
36
- "laplace_predict": "mc_corr",
37
- "lm_head": false,
38
- "cache_dir": "/content/cache/huggingface/metrics/glue",
39
- "step_list": [
40
- 0,
41
- 1068,
42
- 2137,
43
- 3206,
44
- 4275,
45
- 5344
46
- ]
47
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/args.json RENAMED
File without changes
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/logfile.log ADDED
@@ -0,0 +1,312 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 04/30/2024 19:12:31 - INFO - __main__ - Distributed environment: NO
2
+ Num processes: 1
3
+ Process index: 0
4
+ Local process index: 0
5
+ Device: cpu
6
+
7
+ Mixed precision type: no
8
+
9
+ 04/30/2024 19:12:35 - INFO - __main__ - Number of labels detected = 2
10
+ 04/30/2024 19:12:36 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
11
+ 04/30/2024 19:12:37 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/adapter_config.json
12
+ 04/30/2024 19:12:37 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
13
+ 04/30/2024 19:12:37 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/pytorch_adapter.bin
14
+ 04/30/2024 19:12:37 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/head_config.json
15
+ 04/30/2024 19:12:37 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
16
+ 04/30/2024 19:12:37 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/pytorch_model_head.bin
17
+ 04/30/2024 19:12:37 - INFO - __main__ - Adapter Name = cola
18
+ 04/30/2024 19:12:37 - INFO - __main__ - heads.cola.1.weight
19
+ 04/30/2024 19:12:37 - INFO - __main__ - heads.cola.1.bias
20
+ 04/30/2024 19:12:37 - INFO - __main__ - heads.cola.4.weight
21
+ 04/30/2024 19:12:37 - INFO - __main__ - heads.cola.4.bias
22
+ 04/30/2024 19:12:42 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
23
+ 04/30/2024 19:12:42 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
24
+ 04/30/2024 19:12:42 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
25
+ 04/30/2024 19:26:31 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
26
+ 04/30/2024 19:26:31 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
27
+ 04/30/2024 19:26:31 - INFO - __main__ - tensor([[ 0.0687, 0.1107],
28
+ [ 0.1065, 0.1546],
29
+ [ 0.1252, 0.1252],
30
+ ...,
31
+ [ 0.0203, 0.1087],
32
+ [ 0.0962, 0.1235],
33
+ [ 0.0511, -0.0174]])
34
+ 04/30/2024 19:26:31 - INFO - __main__ - tensor([[[0.2029, 0.1885],
35
+ [0.1885, 0.2029]],
36
+
37
+ [[0.1999, 0.1717],
38
+ [0.1717, 0.1999]],
39
+
40
+ [[0.1960, 0.1759],
41
+ [0.1759, 0.1960]],
42
+
43
+ ...,
44
+
45
+ [[0.2176, 0.1939],
46
+ [0.1939, 0.2176]],
47
+
48
+ [[0.1863, 0.1736],
49
+ [0.1736, 0.1863]],
50
+
51
+ [[0.1892, 0.1718],
52
+ [0.1718, 0.1892]]])
53
+ 04/30/2024 19:26:31 - INFO - __main__ - Distributed environment: NO
54
+ Num processes: 1
55
+ Process index: 0
56
+ Local process index: 0
57
+ Device: cpu
58
+
59
+ Mixed precision type: no
60
+
61
+ 04/30/2024 19:26:35 - INFO - __main__ - Number of labels detected = 2
62
+ 04/30/2024 19:26:35 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
63
+ 04/30/2024 19:26:37 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/adapter_config.json
64
+ 04/30/2024 19:26:37 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
65
+ 04/30/2024 19:26:37 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/pytorch_adapter.bin
66
+ 04/30/2024 19:26:37 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/head_config.json
67
+ 04/30/2024 19:26:37 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
68
+ 04/30/2024 19:26:37 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/pytorch_model_head.bin
69
+ 04/30/2024 19:26:37 - INFO - __main__ - Adapter Name = cola
70
+ 04/30/2024 19:26:37 - INFO - __main__ - heads.cola.1.weight
71
+ 04/30/2024 19:26:37 - INFO - __main__ - heads.cola.1.bias
72
+ 04/30/2024 19:26:37 - INFO - __main__ - heads.cola.4.weight
73
+ 04/30/2024 19:26:37 - INFO - __main__ - heads.cola.4.bias
74
+ 04/30/2024 19:26:41 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
75
+ 04/30/2024 19:26:41 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
76
+ 04/30/2024 19:26:41 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
77
+ 04/30/2024 19:40:51 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
78
+ 04/30/2024 19:40:51 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
79
+ 04/30/2024 19:40:51 - INFO - __main__ - tensor([[-1.4529, 1.7744],
80
+ [-1.1889, 1.5487],
81
+ [-1.4239, 1.8257],
82
+ ...,
83
+ [-2.0956, 2.2813],
84
+ [-0.5643, 0.8205],
85
+ [-0.9262, 0.8999]])
86
+ 04/30/2024 19:40:51 - INFO - __main__ - tensor([[[0.3071, 0.2560],
87
+ [0.2560, 0.3071]],
88
+
89
+ [[0.2708, 0.2068],
90
+ [0.2068, 0.2708]],
91
+
92
+ [[0.3089, 0.2527],
93
+ [0.2527, 0.3089]],
94
+
95
+ ...,
96
+
97
+ [[0.4054, 0.3471],
98
+ [0.3471, 0.4054]],
99
+
100
+ [[0.1919, 0.1553],
101
+ [0.1553, 0.1919]],
102
+
103
+ [[0.2052, 0.1670],
104
+ [0.1670, 0.2052]]])
105
+ 04/30/2024 19:40:51 - INFO - __main__ - Distributed environment: NO
106
+ Num processes: 1
107
+ Process index: 0
108
+ Local process index: 0
109
+ Device: cpu
110
+
111
+ Mixed precision type: no
112
+
113
+ 04/30/2024 19:40:53 - INFO - __main__ - Number of labels detected = 2
114
+ 04/30/2024 19:40:54 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
115
+ 04/30/2024 19:40:55 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/adapter_config.json
116
+ 04/30/2024 19:40:55 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
117
+ 04/30/2024 19:40:55 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/pytorch_adapter.bin
118
+ 04/30/2024 19:40:55 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/head_config.json
119
+ 04/30/2024 19:40:55 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
120
+ 04/30/2024 19:40:55 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/pytorch_model_head.bin
121
+ 04/30/2024 19:40:55 - INFO - __main__ - Adapter Name = cola
122
+ 04/30/2024 19:40:55 - INFO - __main__ - heads.cola.1.weight
123
+ 04/30/2024 19:40:55 - INFO - __main__ - heads.cola.1.bias
124
+ 04/30/2024 19:40:55 - INFO - __main__ - heads.cola.4.weight
125
+ 04/30/2024 19:40:55 - INFO - __main__ - heads.cola.4.bias
126
+ 04/30/2024 19:41:00 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
127
+ 04/30/2024 19:41:00 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
128
+ 04/30/2024 19:41:00 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
129
+ 04/30/2024 19:55:00 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
130
+ 04/30/2024 19:55:00 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
131
+ 04/30/2024 19:55:00 - INFO - __main__ - tensor([[-1.2545, 1.5103],
132
+ [-1.1312, 1.4272],
133
+ [-1.0637, 1.4316],
134
+ ...,
135
+ [-1.9212, 2.1010],
136
+ [-0.2051, 0.3802],
137
+ [-0.7883, 0.7410]])
138
+ 04/30/2024 19:55:00 - INFO - __main__ - tensor([[[0.2652, 0.2093],
139
+ [0.2093, 0.2652]],
140
+
141
+ [[0.2405, 0.1822],
142
+ [0.1822, 0.2405]],
143
+
144
+ [[0.2576, 0.2011],
145
+ [0.2011, 0.2576]],
146
+
147
+ ...,
148
+
149
+ [[0.3661, 0.3102],
150
+ [0.3102, 0.3661]],
151
+
152
+ [[0.1545, 0.1238],
153
+ [0.1238, 0.1545]],
154
+
155
+ [[0.1725, 0.1392],
156
+ [0.1392, 0.1725]]])
157
+ 04/30/2024 19:55:00 - INFO - __main__ - Distributed environment: NO
158
+ Num processes: 1
159
+ Process index: 0
160
+ Local process index: 0
161
+ Device: cpu
162
+
163
+ Mixed precision type: no
164
+
165
+ 04/30/2024 19:55:02 - INFO - __main__ - Number of labels detected = 2
166
+ 04/30/2024 19:55:03 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
167
+ 04/30/2024 19:55:04 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_3206/adapter_config.json
168
+ 04/30/2024 19:55:04 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
169
+ 04/30/2024 19:55:04 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_3206/pytorch_adapter.bin
170
+ 04/30/2024 19:55:04 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_3206/head_config.json
171
+ 04/30/2024 19:55:04 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
172
+ 04/30/2024 19:55:04 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_3206/pytorch_model_head.bin
173
+ 04/30/2024 19:55:04 - INFO - __main__ - Adapter Name = cola
174
+ 04/30/2024 19:55:04 - INFO - __main__ - heads.cola.1.weight
175
+ 04/30/2024 19:55:04 - INFO - __main__ - heads.cola.1.bias
176
+ 04/30/2024 19:55:04 - INFO - __main__ - heads.cola.4.weight
177
+ 04/30/2024 19:55:04 - INFO - __main__ - heads.cola.4.bias
178
+ 04/30/2024 19:55:09 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
179
+ 04/30/2024 19:55:09 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
180
+ 04/30/2024 19:55:09 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
181
+ 04/30/2024 20:09:20 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
182
+ 04/30/2024 20:09:20 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
183
+ 04/30/2024 20:09:20 - INFO - __main__ - tensor([[-1.5903, 1.8046],
184
+ [-1.4262, 1.7163],
185
+ [-1.2728, 1.6234],
186
+ ...,
187
+ [-2.3189, 2.5137],
188
+ [-0.0952, 0.2233],
189
+ [-0.7922, 0.7461]])
190
+ 04/30/2024 20:09:20 - INFO - __main__ - tensor([[[0.3187, 0.2535],
191
+ [0.2535, 0.3187]],
192
+
193
+ [[0.2909, 0.2243],
194
+ [0.2243, 0.2909]],
195
+
196
+ [[0.2970, 0.2309],
197
+ [0.2309, 0.2970]],
198
+
199
+ ...,
200
+
201
+ [[0.4427, 0.3839],
202
+ [0.3839, 0.4427]],
203
+
204
+ [[0.1634, 0.1292],
205
+ [0.1292, 0.1634]],
206
+
207
+ [[0.1837, 0.1443],
208
+ [0.1443, 0.1837]]])
209
+ 04/30/2024 20:09:20 - INFO - __main__ - Distributed environment: NO
210
+ Num processes: 1
211
+ Process index: 0
212
+ Local process index: 0
213
+ Device: cpu
214
+
215
+ Mixed precision type: no
216
+
217
+ 04/30/2024 20:09:22 - INFO - __main__ - Number of labels detected = 2
218
+ 04/30/2024 20:09:23 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
219
+ 04/30/2024 20:09:23 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_4275/adapter_config.json
220
+ 04/30/2024 20:09:23 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
221
+ 04/30/2024 20:09:23 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_4275/pytorch_adapter.bin
222
+ 04/30/2024 20:09:23 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_4275/head_config.json
223
+ 04/30/2024 20:09:23 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
224
+ 04/30/2024 20:09:24 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_4275/pytorch_model_head.bin
225
+ 04/30/2024 20:09:24 - INFO - __main__ - Adapter Name = cola
226
+ 04/30/2024 20:09:24 - INFO - __main__ - heads.cola.1.weight
227
+ 04/30/2024 20:09:24 - INFO - __main__ - heads.cola.1.bias
228
+ 04/30/2024 20:09:24 - INFO - __main__ - heads.cola.4.weight
229
+ 04/30/2024 20:09:24 - INFO - __main__ - heads.cola.4.bias
230
+ 04/30/2024 20:09:28 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
231
+ 04/30/2024 20:09:28 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
232
+ 04/30/2024 20:09:28 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
233
+ 04/30/2024 20:23:30 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
234
+ 04/30/2024 20:23:30 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
235
+ 04/30/2024 20:23:30 - INFO - __main__ - tensor([[-1.8506, 2.0807],
236
+ [-1.5632, 1.8705],
237
+ [-1.4465, 1.8179],
238
+ ...,
239
+ [-2.5331, 2.7452],
240
+ [-0.1701, 0.3093],
241
+ [-0.9411, 0.9126]])
242
+ 04/30/2024 20:23:30 - INFO - __main__ - tensor([[[0.3666, 0.2971],
243
+ [0.2971, 0.3666]],
244
+
245
+ [[0.3195, 0.2491],
246
+ [0.2491, 0.3195]],
247
+
248
+ [[0.3326, 0.2601],
249
+ [0.2601, 0.3326]],
250
+
251
+ ...,
252
+
253
+ [[0.4860, 0.4264],
254
+ [0.4264, 0.4860]],
255
+
256
+ [[0.1715, 0.1336],
257
+ [0.1336, 0.1715]],
258
+
259
+ [[0.2016, 0.1572],
260
+ [0.1572, 0.2016]]])
261
+ 04/30/2024 20:23:30 - INFO - __main__ - Distributed environment: NO
262
+ Num processes: 1
263
+ Process index: 0
264
+ Local process index: 0
265
+ Device: cpu
266
+
267
+ Mixed precision type: no
268
+
269
+ 04/30/2024 20:23:32 - INFO - __main__ - Number of labels detected = 2
270
+ 04/30/2024 20:23:33 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
271
+ 04/30/2024 20:23:34 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_5344/adapter_config.json
272
+ 04/30/2024 20:23:34 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
273
+ 04/30/2024 20:23:34 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_5344/pytorch_adapter.bin
274
+ 04/30/2024 20:23:34 - INFO - adapters.loading - Loading module configuration from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_5344/head_config.json
275
+ 04/30/2024 20:23:34 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
276
+ 04/30/2024 20:23:34 - INFO - adapters.loading - Loading module weights from ./outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_5344/pytorch_model_head.bin
277
+ 04/30/2024 20:23:34 - INFO - __main__ - Adapter Name = cola
278
+ 04/30/2024 20:23:34 - INFO - __main__ - heads.cola.1.weight
279
+ 04/30/2024 20:23:34 - INFO - __main__ - heads.cola.1.bias
280
+ 04/30/2024 20:23:34 - INFO - __main__ - heads.cola.4.weight
281
+ 04/30/2024 20:23:34 - INFO - __main__ - heads.cola.4.bias
282
+ 04/30/2024 20:23:39 - INFO - __main__ - Sample 5238 of the training set: {'input_ids': [101, 2009, 1005, 1055, 2986, 2008, 2002, 3825, 1998, 17806, 1010, 2021, 1045, 2123, 1005, 1056, 2428, 2729, 2055, 2010, 15531, 1010, 2030, 1996, 2769, 1010, 2030, 2505, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
283
+ 04/30/2024 20:23:39 - INFO - __main__ - Sample 912 of the training set: {'input_ids': [101, 1045, 2113, 2029, 2338, 23848, 3191, 1010, 1998, 2029, 2338, 3960, 2356, 2339, 2017, 2910, 1005, 1056, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
284
+ 04/30/2024 20:23:39 - INFO - __main__ - Sample 204 of the training set: {'input_ids': [101, 1996, 26108, 2002, 4152, 1010, 1996, 2062, 2198, 6010, 11067, 2229, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
285
+ 04/30/2024 20:37:44 - INFO - __main__ - f_mu shape : torch.Size([1043, 2])
286
+ 04/30/2024 20:37:44 - INFO - __main__ - f_var shape : torch.Size([1043, 2, 2])
287
+ 04/30/2024 20:37:44 - INFO - __main__ - tensor([[-1.8640, 2.1037],
288
+ [-1.5643, 1.8815],
289
+ [-1.4149, 1.7945],
290
+ ...,
291
+ [-2.4889, 2.7067],
292
+ [-0.1663, 0.3097],
293
+ [-0.8635, 0.8440]])
294
+ 04/30/2024 20:37:44 - INFO - __main__ - tensor([[[0.3703, 0.3002],
295
+ [0.3002, 0.3703]],
296
+
297
+ [[0.3209, 0.2497],
298
+ [0.2497, 0.3209]],
299
+
300
+ [[0.3293, 0.2559],
301
+ [0.2559, 0.3293]],
302
+
303
+ ...,
304
+
305
+ [[0.4776, 0.4180],
306
+ [0.4180, 0.4776]],
307
+
308
+ [[0.1715, 0.1335],
309
+ [0.1335, 0.1715]],
310
+
311
+ [[0.1975, 0.1524],
312
+ [0.1524, 0.1975]]])
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/adapter_config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/all_results.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/all_results_la_kron_all_homo_mc_corr_1000.json RENAMED
File without changes
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/all_results_la_kron_last_layer_homo_mc_corr_1000.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eval_matthews_correlation": 0.01029747168884776}
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/eval_res.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/eval_res_la_kron_all_homo_mc_corr_1000.json RENAMED
File without changes
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/eval_res_la_kron_last_layer_homo_mc_corr_1000.json ADDED
The diff for this file is too large to render. See raw diff
 
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/generation_config.json RENAMED
File without changes
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_0/gpu_stats.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "memory_allocated": 0,
3
+ "max_memory_allocated": 0,
4
+ "memory_reserved": 0,
5
+ "max_memory_reserved": 0,
6
+ "memory_stats": {}
7
+ }
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/head_config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/pytorch_adapter.bin RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/pytorch_model_head.bin RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/special_tokens_map.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/tokenizer.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/tokenizer_config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_0/vocab.txt RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/adapter_config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/all_results.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/all_results_la_kron_all_homo_mc_corr_1000.json RENAMED
File without changes
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/all_results_la_kron_last_layer_homo_mc_corr_1000.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eval_matthews_correlation": 0.46484295300587036}
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/eval_res.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/eval_res_la_kron_all_homo_mc_corr_1000.json RENAMED
File without changes
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/eval_res_la_kron_last_layer_homo_mc_corr_1000.json ADDED
The diff for this file is too large to render. See raw diff
 
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/generation_config.json RENAMED
File without changes
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_1068/gpu_stats.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "memory_allocated": 0,
3
+ "max_memory_allocated": 0,
4
+ "memory_reserved": 0,
5
+ "max_memory_reserved": 0,
6
+ "memory_stats": {}
7
+ }
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/head_config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/pytorch_adapter.bin RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/pytorch_model_head.bin RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/special_tokens_map.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/tokenizer.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/tokenizer_config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_1068/vocab.txt RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/adapter_config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/all_results.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/all_results_la_kron_all_homo_mc_corr_1000.json RENAMED
File without changes
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/all_results_la_kron_last_layer_homo_mc_corr_1000.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eval_matthews_correlation": 0.5072753983504138}
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/eval_res.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/eval_res_la_kron_all_homo_mc_corr_1000.json RENAMED
File without changes
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/eval_res_la_kron_last_layer_homo_mc_corr_1000.json ADDED
The diff for this file is too large to render. See raw diff
 
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/generation_config.json RENAMED
File without changes
outputs/cola/google-bert/bert-base-uncased_adapterstrain_val_5e-05_42/step_2137/gpu_stats.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "memory_allocated": 0,
3
+ "max_memory_allocated": 0,
4
+ "memory_reserved": 0,
5
+ "max_memory_reserved": 0,
6
+ "memory_stats": {}
7
+ }
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/head_config.json RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/pytorch_adapter.bin RENAMED
File without changes
outputs/cola/google-bert/{bert-base-uncased_adapters_5e-05_42 β†’ bert-base-uncased_adapterstrain_val_5e-05_42}/step_2137/pytorch_model_head.bin RENAMED
File without changes