Shawon16 commited on
Commit
8e137d2
·
verified ·
1 Parent(s): 6cae7f6

End of training

Browse files
README.md CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [facebook/timesformer-base-finetuned-k400](https://huggingface.co/facebook/timesformer-base-finetuned-k400) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 1.1312
22
- - Accuracy: 0.7304
23
 
24
  ## Model description
25
 
@@ -45,13 +45,13 @@ The following hyperparameters were used during training:
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_ratio: 0.1
48
- - training_steps: 669
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
53
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
54
- | 0.1824 | 1.0 | 669 | 0.0990 | 0.9933 |
55
 
56
 
57
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [facebook/timesformer-base-finetuned-k400](https://huggingface.co/facebook/timesformer-base-finetuned-k400) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.2304
22
+ - Accuracy: 0.7288
23
 
24
  ## Model description
25
 
 
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_ratio: 0.1
48
+ - training_steps: 502
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
53
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
54
+ | 0.3006 | 1.0 | 502 | 0.2494 | 0.9806 |
55
 
56
 
57
  ### Framework versions
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_accuracy": 0.7304075235109718,
4
- "eval_loss": 1.1312494277954102,
5
- "eval_runtime": 377.402,
6
- "eval_samples_per_second": 3.381,
7
- "eval_steps_per_second": 0.424
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_accuracy": 0.7288401253918495,
4
+ "eval_loss": 1.230407953262329,
5
+ "eval_runtime": 286.6871,
6
+ "eval_samples_per_second": 4.451,
7
+ "eval_steps_per_second": 0.558
8
  }
confusion_matrix_test_fold_2.png ADDED
confusion_matrix_train_fold_2.png ADDED
confusion_matrix_valid_fold_2.png ADDED
test_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_accuracy": 0.7304075235109718,
4
- "eval_loss": 1.1312494277954102,
5
- "eval_runtime": 377.402,
6
- "eval_samples_per_second": 3.381,
7
- "eval_steps_per_second": 0.424
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_accuracy": 0.7288401253918495,
4
+ "eval_loss": 1.230407953262329,
5
+ "eval_runtime": 286.6871,
6
+ "eval_samples_per_second": 4.451,
7
+ "eval_steps_per_second": 0.558
8
  }
trainer_state.json CHANGED
@@ -1,85 +1,78 @@
1
  {
2
- "best_metric": 0.9932768924302788,
3
- "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/10 fold timesformer/Timesformer_default_fold_10_10_epoch_noAug_batch8_codecheck/checkpoint-669",
4
  "epoch": 1.0,
5
  "eval_steps": 500,
6
- "global_step": 669,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.14947683109118087,
13
- "grad_norm": 11.951628684997559,
14
- "learning_rate": 4.725913621262459e-05,
15
- "loss": 3.7222,
16
  "step": 100
17
  },
18
  {
19
- "epoch": 0.29895366218236175,
20
- "grad_norm": 10.303458213806152,
21
- "learning_rate": 3.895348837209303e-05,
22
- "loss": 1.9162,
23
  "step": 200
24
  },
25
  {
26
- "epoch": 0.4484304932735426,
27
- "grad_norm": 7.89475154876709,
28
- "learning_rate": 3.064784053156146e-05,
29
- "loss": 0.971,
30
  "step": 300
31
  },
32
  {
33
- "epoch": 0.5979073243647235,
34
- "grad_norm": 3.1657724380493164,
35
- "learning_rate": 2.23421926910299e-05,
36
- "loss": 0.478,
37
  "step": 400
38
  },
39
  {
40
- "epoch": 0.7473841554559043,
41
- "grad_norm": 2.475733757019043,
42
- "learning_rate": 1.403654485049834e-05,
43
- "loss": 0.2504,
44
  "step": 500
45
  },
46
- {
47
- "epoch": 0.8968609865470852,
48
- "grad_norm": 5.244026184082031,
49
- "learning_rate": 5.730897009966778e-06,
50
- "loss": 0.1824,
51
- "step": 600
52
- },
53
  {
54
  "epoch": 1.0,
55
- "eval_accuracy": 0.9932768924302788,
56
- "eval_loss": 0.09897388517856598,
57
- "eval_runtime": 887.6237,
58
- "eval_samples_per_second": 4.524,
59
- "eval_steps_per_second": 0.566,
60
- "step": 669
61
  },
62
  {
63
  "epoch": 1.0,
64
- "step": 669,
65
- "total_flos": 4.690865157815402e+18,
66
- "train_loss": 1.136438287070692,
67
- "train_runtime": 2516.9818,
68
- "train_samples_per_second": 2.126,
69
- "train_steps_per_second": 0.266
70
  },
71
  {
72
  "epoch": 1.0,
73
- "eval_accuracy": 0.7304075235109718,
74
- "eval_loss": 1.1312494277954102,
75
- "eval_runtime": 377.402,
76
- "eval_samples_per_second": 3.381,
77
- "eval_steps_per_second": 0.424,
78
- "step": 669
79
  }
80
  ],
81
  "logging_steps": 100,
82
- "max_steps": 669,
83
  "num_input_tokens_seen": 0,
84
  "num_train_epochs": 9223372036854775807,
85
  "save_steps": 500,
@@ -95,7 +88,7 @@
95
  "attributes": {}
96
  }
97
  },
98
- "total_flos": 4.690865157815402e+18,
99
  "train_batch_size": 8,
100
  "trial_name": null,
101
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.9805728518057285,
3
+ "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/10 fold timesformer/Timesformer_default_fold_10_10_epoch_noAug_batch8_codecheck/checkpoint-502",
4
  "epoch": 1.0,
5
  "eval_steps": 500,
6
+ "global_step": 502,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.199203187250996,
13
+ "grad_norm": 11.452275276184082,
14
+ "learning_rate": 4.456762749445677e-05,
15
+ "loss": 3.5774,
16
  "step": 100
17
  },
18
  {
19
+ "epoch": 0.398406374501992,
20
+ "grad_norm": 10.499092102050781,
21
+ "learning_rate": 3.348115299334812e-05,
22
+ "loss": 1.8278,
23
  "step": 200
24
  },
25
  {
26
+ "epoch": 0.5976095617529881,
27
+ "grad_norm": 7.440881729125977,
28
+ "learning_rate": 2.239467849223947e-05,
29
+ "loss": 0.9037,
30
  "step": 300
31
  },
32
  {
33
+ "epoch": 0.796812749003984,
34
+ "grad_norm": 2.855297088623047,
35
+ "learning_rate": 1.130820399113082e-05,
36
+ "loss": 0.4349,
37
  "step": 400
38
  },
39
  {
40
+ "epoch": 0.9960159362549801,
41
+ "grad_norm": 3.7258729934692383,
42
+ "learning_rate": 2.2172949002217296e-07,
43
+ "loss": 0.3006,
44
  "step": 500
45
  },
 
 
 
 
 
 
 
46
  {
47
  "epoch": 1.0,
48
+ "eval_accuracy": 0.9805728518057285,
49
+ "eval_loss": 0.24935545027256012,
50
+ "eval_runtime": 1419.7674,
51
+ "eval_samples_per_second": 2.828,
52
+ "eval_steps_per_second": 0.354,
53
+ "step": 502
54
  },
55
  {
56
  "epoch": 1.0,
57
+ "step": 502,
58
+ "total_flos": 3.5199018075087176e+18,
59
+ "train_loss": 1.403769117427537,
60
+ "train_runtime": 3097.7448,
61
+ "train_samples_per_second": 1.296,
62
+ "train_steps_per_second": 0.162
63
  },
64
  {
65
  "epoch": 1.0,
66
+ "eval_accuracy": 0.7288401253918495,
67
+ "eval_loss": 1.230407953262329,
68
+ "eval_runtime": 286.6871,
69
+ "eval_samples_per_second": 4.451,
70
+ "eval_steps_per_second": 0.558,
71
+ "step": 502
72
  }
73
  ],
74
  "logging_steps": 100,
75
+ "max_steps": 502,
76
  "num_input_tokens_seen": 0,
77
  "num_train_epochs": 9223372036854775807,
78
  "save_steps": 500,
 
88
  "attributes": {}
89
  }
90
  },
91
+ "total_flos": 3.5199018075087176e+18,
92
  "train_batch_size": 8,
93
  "trial_name": null,
94
  "trial_params": null