fats-fme commited on
Commit
79d29e4
·
verified ·
1 Parent(s): 434aeb2

Training in progress, step 82, checkpoint

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8487ce7bdbc23b357aa541d0b6f6552ab3cd137d9189f6805444f0a71649243
3
  size 25986148
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39b94c672a97c9015e8295fd7a26b83b30f58fc04dbbdb99d68bcad055f93947
3
  size 25986148
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3595634126481ea254b652ccb287819db6abe86cb9d82adc40da7f64a80c4f6e
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41f026569d2524bee91b5ef6c45e9ccac7bd6699e35f824786c4e9bc2cd3b716
3
  size 14512
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9485cd7ca685639d72a742a35da6508ee70e4def8667e09a65fece70b6b717f
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1453247989fe4f40cb58011506f6735ba25507deafeaf2cc886ff35d4634f2b0
3
  size 14512
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:725ee499aaa0bc04b490ac3af0c734c514c976dd8cd2f204b00fdb43d2a90bf8
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d20550e97739760e033bfa2dbeb5dbd23fb1ea480b478b5de311b00b3043e221
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.838709677419355,
5
  "eval_steps": 7,
6
- "global_step": 77,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -642,6 +642,41 @@
642
  "eval_samples_per_second": 13.773,
643
  "eval_steps_per_second": 1.797,
644
  "step": 77
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
645
  }
646
  ],
647
  "logging_steps": 1,
@@ -656,12 +691,12 @@
656
  "should_evaluate": false,
657
  "should_log": false,
658
  "should_save": true,
659
- "should_training_stop": false
660
  },
661
  "attributes": {}
662
  }
663
  },
664
- "total_flos": 6.34091163573289e+16,
665
  "train_batch_size": 4,
666
  "trial_name": null,
667
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 3.0230414746543777,
5
  "eval_steps": 7,
6
+ "global_step": 82,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
642
  "eval_samples_per_second": 13.773,
643
  "eval_steps_per_second": 1.797,
644
  "step": 77
645
+ },
646
+ {
647
+ "epoch": 2.8755760368663594,
648
+ "grad_norm": Infinity,
649
+ "learning_rate": 0.00015600000000000002,
650
+ "loss": 1.4261,
651
+ "step": 78
652
+ },
653
+ {
654
+ "epoch": 2.912442396313364,
655
+ "grad_norm": Infinity,
656
+ "learning_rate": 0.00015800000000000002,
657
+ "loss": 1.4717,
658
+ "step": 79
659
+ },
660
+ {
661
+ "epoch": 2.9493087557603688,
662
+ "grad_norm": Infinity,
663
+ "learning_rate": 0.00016,
664
+ "loss": 1.4148,
665
+ "step": 80
666
+ },
667
+ {
668
+ "epoch": 2.986175115207373,
669
+ "grad_norm": Infinity,
670
+ "learning_rate": 0.000162,
671
+ "loss": 1.4234,
672
+ "step": 81
673
+ },
674
+ {
675
+ "epoch": 3.0230414746543777,
676
+ "grad_norm": Infinity,
677
+ "learning_rate": 0.000164,
678
+ "loss": 2.5737,
679
+ "step": 82
680
  }
681
  ],
682
  "logging_steps": 1,
 
691
  "should_evaluate": false,
692
  "should_log": false,
693
  "should_save": true,
694
+ "should_training_stop": true
695
  },
696
  "attributes": {}
697
  }
698
  },
699
+ "total_flos": 6.752659144546714e+16,
700
  "train_batch_size": 4,
701
  "trial_name": null,
702
  "trial_params": null