Shawon16 commited on
Commit
df375a2
·
verified ·
1 Parent(s): 69760b3

End of training

Browse files
README.md CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [facebook/timesformer-base-finetuned-k400](https://huggingface.co/facebook/timesformer-base-finetuned-k400) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.7724
22
- - Accuracy: 0.8111
23
 
24
  ## Model description
25
 
@@ -45,14 +45,14 @@ The following hyperparameters were used during training:
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_ratio: 0.1
48
- - training_steps: 1002
49
 
50
  ### Training results
51
 
52
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
53
- |:-------------:|:------:|:----:|:---------------:|:--------:|
54
- | 0.3111 | 0.5010 | 502 | 0.2995 | 0.9457 |
55
- | 0.0629 | 1.4990 | 1002 | 0.0642 | 0.9918 |
56
 
57
 
58
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [facebook/timesformer-base-finetuned-k400](https://huggingface.co/facebook/timesformer-base-finetuned-k400) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.8394
22
+ - Accuracy: 0.8017
23
 
24
  ## Model description
25
 
 
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_ratio: 0.1
48
+ - training_steps: 1004
49
 
50
  ### Training results
51
 
52
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
53
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
54
+ | 0.3785 | 0.5 | 502 | 0.3069 | 0.9479 |
55
+ | 0.0487 | 1.5 | 1004 | 0.0754 | 0.9883 |
56
 
57
 
58
  ### Framework versions
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 1.499001996007984,
3
- "eval_accuracy": 0.8111285266457681,
4
- "eval_loss": 0.7723951935768127,
5
- "eval_runtime": 509.1163,
6
- "eval_samples_per_second": 2.506,
7
- "eval_steps_per_second": 0.314
8
  }
 
1
  {
2
+ "epoch": 1.5,
3
+ "eval_accuracy": 0.8017241379310345,
4
+ "eval_loss": 0.8394390940666199,
5
+ "eval_runtime": 424.8162,
6
+ "eval_samples_per_second": 3.004,
7
+ "eval_steps_per_second": 0.377
8
  }
confusion_matrix_kfold_fold_2.png ADDED
confusion_matrix_test_fold_2.png CHANGED
confusion_matrix_trainfold_fold_2.png ADDED
test_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 1.499001996007984,
3
- "eval_accuracy": 0.8111285266457681,
4
- "eval_loss": 0.7723951935768127,
5
- "eval_runtime": 509.1163,
6
- "eval_samples_per_second": 2.506,
7
- "eval_steps_per_second": 0.314
8
  }
 
1
  {
2
+ "epoch": 1.5,
3
+ "eval_accuracy": 0.8017241379310345,
4
+ "eval_loss": 0.8394390940666199,
5
+ "eval_runtime": 424.8162,
6
+ "eval_samples_per_second": 3.004,
7
+ "eval_steps_per_second": 0.377
8
  }
trainer_state.json CHANGED
@@ -1,122 +1,122 @@
1
  {
2
- "best_metric": 0.9917828685258964,
3
- "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/10 fold timesformer/Timesformer_default_fold_10_10_epoch_noAug_batch8_codecheck/checkpoint-1002",
4
- "epoch": 1.499001996007984,
5
  "eval_steps": 500,
6
- "global_step": 1002,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.0998003992015968,
13
- "grad_norm": 12.41899585723877,
14
  "learning_rate": 4.950495049504951e-05,
15
- "loss": 3.8639,
16
  "step": 100
17
  },
18
  {
19
- "epoch": 0.1996007984031936,
20
- "grad_norm": 11.086575508117676,
21
- "learning_rate": 4.4506104328523865e-05,
22
- "loss": 2.3042,
23
  "step": 200
24
  },
25
  {
26
- "epoch": 0.2994011976047904,
27
- "grad_norm": 12.340323448181152,
28
- "learning_rate": 3.895671476137625e-05,
29
- "loss": 1.027,
30
  "step": 300
31
  },
32
  {
33
- "epoch": 0.3992015968063872,
34
- "grad_norm": 6.579195499420166,
35
- "learning_rate": 3.340732519422864e-05,
36
- "loss": 0.6417,
37
  "step": 400
38
  },
39
  {
40
- "epoch": 0.499001996007984,
41
- "grad_norm": 8.828213691711426,
42
- "learning_rate": 2.785793562708102e-05,
43
- "loss": 0.3111,
44
  "step": 500
45
  },
46
  {
47
- "epoch": 0.500998003992016,
48
- "eval_accuracy": 0.9457171314741036,
49
- "eval_loss": 0.2995070815086365,
50
- "eval_runtime": 1755.2973,
51
- "eval_samples_per_second": 2.288,
52
- "eval_steps_per_second": 0.286,
53
  "step": 502
54
  },
55
  {
56
- "epoch": 1.0978043912175648,
57
- "grad_norm": 7.735666275024414,
58
- "learning_rate": 2.230854605993341e-05,
59
- "loss": 0.1745,
60
  "step": 600
61
  },
62
  {
63
- "epoch": 1.1976047904191618,
64
- "grad_norm": 1.334542155265808,
65
- "learning_rate": 1.6759156492785795e-05,
66
- "loss": 0.1142,
67
  "step": 700
68
  },
69
  {
70
- "epoch": 1.2974051896207586,
71
- "grad_norm": 6.124719619750977,
72
- "learning_rate": 1.120976692563818e-05,
73
- "loss": 0.0919,
74
  "step": 800
75
  },
76
  {
77
- "epoch": 1.3972055888223553,
78
- "grad_norm": 7.422993183135986,
79
- "learning_rate": 5.660377358490566e-06,
80
- "loss": 0.0588,
81
  "step": 900
82
  },
83
  {
84
- "epoch": 1.4970059880239521,
85
- "grad_norm": 1.5655475854873657,
86
- "learning_rate": 1.1098779134295228e-07,
87
- "loss": 0.0629,
88
  "step": 1000
89
  },
90
  {
91
- "epoch": 1.499001996007984,
92
- "eval_accuracy": 0.9917828685258964,
93
- "eval_loss": 0.06420521438121796,
94
- "eval_runtime": 1671.9392,
95
- "eval_samples_per_second": 2.402,
96
- "eval_steps_per_second": 0.3,
97
- "step": 1002
98
  },
99
  {
100
- "epoch": 1.499001996007984,
101
- "step": 1002,
102
- "total_flos": 7.024903632266527e+18,
103
- "train_loss": 0.863516083377564,
104
- "train_runtime": 7438.8722,
105
- "train_samples_per_second": 1.078,
106
- "train_steps_per_second": 0.135
107
  },
108
  {
109
- "epoch": 1.499001996007984,
110
- "eval_accuracy": 0.8111285266457681,
111
- "eval_loss": 0.7723951935768127,
112
- "eval_runtime": 509.1163,
113
- "eval_samples_per_second": 2.506,
114
- "eval_steps_per_second": 0.314,
115
- "step": 1002
116
  }
117
  ],
118
  "logging_steps": 100,
119
- "max_steps": 1002,
120
  "num_input_tokens_seen": 0,
121
  "num_train_epochs": 9223372036854775807,
122
  "save_steps": 500,
@@ -132,7 +132,7 @@
132
  "attributes": {}
133
  }
134
  },
135
- "total_flos": 7.024903632266527e+18,
136
  "train_batch_size": 8,
137
  "trial_name": null,
138
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.988293897882939,
3
+ "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/10 fold timesformer/Timesformer_default_fold_10_10_epoch_noAug_batch8_codecheck/checkpoint-1004",
4
+ "epoch": 1.5,
5
  "eval_steps": 500,
6
+ "global_step": 1004,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.099601593625498,
13
+ "grad_norm": 12.111989974975586,
14
  "learning_rate": 4.950495049504951e-05,
15
+ "loss": 3.7788,
16
  "step": 100
17
  },
18
  {
19
+ "epoch": 0.199203187250996,
20
+ "grad_norm": 10.870170593261719,
21
+ "learning_rate": 4.451827242524917e-05,
22
+ "loss": 2.1601,
23
  "step": 200
24
  },
25
  {
26
+ "epoch": 0.29880478087649404,
27
+ "grad_norm": 9.179058074951172,
28
+ "learning_rate": 3.89811738648948e-05,
29
+ "loss": 1.1244,
30
  "step": 300
31
  },
32
  {
33
+ "epoch": 0.398406374501992,
34
+ "grad_norm": 2.3942956924438477,
35
+ "learning_rate": 3.344407530454042e-05,
36
+ "loss": 0.5512,
37
  "step": 400
38
  },
39
  {
40
+ "epoch": 0.49800796812749004,
41
+ "grad_norm": 3.746593952178955,
42
+ "learning_rate": 2.7906976744186048e-05,
43
+ "loss": 0.3785,
44
  "step": 500
45
  },
46
  {
47
+ "epoch": 0.5,
48
+ "eval_accuracy": 0.947945205479452,
49
+ "eval_loss": 0.30693259835243225,
50
+ "eval_runtime": 1802.7473,
51
+ "eval_samples_per_second": 2.227,
52
+ "eval_steps_per_second": 0.278,
53
  "step": 502
54
  },
55
  {
56
+ "epoch": 1.097609561752988,
57
+ "grad_norm": 3.173877716064453,
58
+ "learning_rate": 2.2369878183831673e-05,
59
+ "loss": 0.1585,
60
  "step": 600
61
  },
62
  {
63
+ "epoch": 1.197211155378486,
64
+ "grad_norm": 3.913566827774048,
65
+ "learning_rate": 1.6832779623477298e-05,
66
+ "loss": 0.1062,
67
  "step": 700
68
  },
69
  {
70
+ "epoch": 1.296812749003984,
71
+ "grad_norm": 0.6361425518989563,
72
+ "learning_rate": 1.1295681063122925e-05,
73
+ "loss": 0.1037,
74
  "step": 800
75
  },
76
  {
77
+ "epoch": 1.3964143426294822,
78
+ "grad_norm": 0.8113210797309875,
79
+ "learning_rate": 5.75858250276855e-06,
80
+ "loss": 0.0565,
81
  "step": 900
82
  },
83
  {
84
+ "epoch": 1.49601593625498,
85
+ "grad_norm": 0.3181760609149933,
86
+ "learning_rate": 2.2148394241417497e-07,
87
+ "loss": 0.0487,
88
  "step": 1000
89
  },
90
  {
91
+ "epoch": 1.5,
92
+ "eval_accuracy": 0.988293897882939,
93
+ "eval_loss": 0.07536407560110092,
94
+ "eval_runtime": 1562.8471,
95
+ "eval_samples_per_second": 2.569,
96
+ "eval_steps_per_second": 0.321,
97
+ "step": 1004
98
  },
99
  {
100
+ "epoch": 1.5,
101
+ "step": 1004,
102
+ "total_flos": 7.039803615017435e+18,
103
+ "train_loss": 0.8433415542145174,
104
+ "train_runtime": 7467.2826,
105
+ "train_samples_per_second": 1.076,
106
+ "train_steps_per_second": 0.134
107
  },
108
  {
109
+ "epoch": 1.5,
110
+ "eval_accuracy": 0.8017241379310345,
111
+ "eval_loss": 0.8394390940666199,
112
+ "eval_runtime": 424.8162,
113
+ "eval_samples_per_second": 3.004,
114
+ "eval_steps_per_second": 0.377,
115
+ "step": 1004
116
  }
117
  ],
118
  "logging_steps": 100,
119
+ "max_steps": 1004,
120
  "num_input_tokens_seen": 0,
121
  "num_train_epochs": 9223372036854775807,
122
  "save_steps": 500,
 
132
  "attributes": {}
133
  }
134
  },
135
+ "total_flos": 7.039803615017435e+18,
136
  "train_batch_size": 8,
137
  "trial_name": null,
138
  "trial_params": null