Training in progress, epoch 1
Browse files- model.safetensors +1 -1
- run-2/checkpoint-58/config.json +27 -0
- run-2/checkpoint-58/model.safetensors +3 -0
- run-2/checkpoint-58/optimizer.pt +3 -0
- run-2/checkpoint-58/rng_state.pth +3 -0
- run-2/checkpoint-58/scheduler.pt +3 -0
- run-2/checkpoint-58/special_tokens_map.json +37 -0
- run-2/checkpoint-58/tokenizer.json +0 -0
- run-2/checkpoint-58/tokenizer_config.json +59 -0
- run-2/checkpoint-58/trainer_state.json +56 -0
- run-2/checkpoint-58/training_args.bin +3 -0
- run-2/checkpoint-58/vocab.txt +0 -0
- run-2/checkpoint-87/config.json +27 -0
- run-2/checkpoint-87/model.safetensors +3 -0
- run-2/checkpoint-87/optimizer.pt +3 -0
- run-2/checkpoint-87/rng_state.pth +3 -0
- run-2/checkpoint-87/scheduler.pt +3 -0
- run-2/checkpoint-87/special_tokens_map.json +37 -0
- run-2/checkpoint-87/tokenizer.json +0 -0
- run-2/checkpoint-87/tokenizer_config.json +59 -0
- run-2/checkpoint-87/trainer_state.json +65 -0
- run-2/checkpoint-87/training_args.bin +3 -0
- run-2/checkpoint-87/vocab.txt +0 -0
- run-3/checkpoint-29/config.json +27 -0
- run-3/checkpoint-29/model.safetensors +3 -0
- run-3/checkpoint-29/optimizer.pt +3 -0
- run-3/checkpoint-29/rng_state.pth +3 -0
- run-3/checkpoint-29/scheduler.pt +3 -0
- run-3/checkpoint-29/special_tokens_map.json +37 -0
- run-3/checkpoint-29/tokenizer.json +0 -0
- run-3/checkpoint-29/tokenizer_config.json +59 -0
- run-3/checkpoint-29/trainer_state.json +47 -0
- run-3/checkpoint-29/training_args.bin +3 -0
- run-3/checkpoint-29/vocab.txt +0 -0
- runs/Aug01_13-47-28_1d376fbba8fb/events.out.tfevents.1722520428.1d376fbba8fb.194.7 +3 -0
- training_args.bin +1 -1
    	
        model.safetensors
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
             
            size 433270768
         | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:4982e2b41506ddf13232beab76be2fa3415c215ada31514b02a193021a1e5f15
         | 
| 3 | 
             
            size 433270768
         | 
    	
        run-2/checkpoint-58/config.json
    ADDED
    
    | @@ -0,0 +1,27 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_name_or_path": "AIMH/mental-bert-base-cased",
         | 
| 3 | 
            +
              "architectures": [
         | 
| 4 | 
            +
                "BertForSequenceClassification"
         | 
| 5 | 
            +
              ],
         | 
| 6 | 
            +
              "attention_probs_dropout_prob": 0.1,
         | 
| 7 | 
            +
              "classifier_dropout": null,
         | 
| 8 | 
            +
              "gradient_checkpointing": false,
         | 
| 9 | 
            +
              "hidden_act": "gelu",
         | 
| 10 | 
            +
              "hidden_dropout_prob": 0.1,
         | 
| 11 | 
            +
              "hidden_size": 768,
         | 
| 12 | 
            +
              "initializer_range": 0.02,
         | 
| 13 | 
            +
              "intermediate_size": 3072,
         | 
| 14 | 
            +
              "layer_norm_eps": 1e-12,
         | 
| 15 | 
            +
              "max_position_embeddings": 512,
         | 
| 16 | 
            +
              "model_type": "bert",
         | 
| 17 | 
            +
              "num_attention_heads": 12,
         | 
| 18 | 
            +
              "num_hidden_layers": 12,
         | 
| 19 | 
            +
              "pad_token_id": 0,
         | 
| 20 | 
            +
              "position_embedding_type": "absolute",
         | 
| 21 | 
            +
              "problem_type": "single_label_classification",
         | 
| 22 | 
            +
              "torch_dtype": "float32",
         | 
| 23 | 
            +
              "transformers_version": "4.42.4",
         | 
| 24 | 
            +
              "type_vocab_size": 2,
         | 
| 25 | 
            +
              "use_cache": true,
         | 
| 26 | 
            +
              "vocab_size": 28996
         | 
| 27 | 
            +
            }
         | 
    	
        run-2/checkpoint-58/model.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:852e74afdea9b1c2539ec7c1ede00b20a12cfd9b000d1ed07be1acde42f6af69
         | 
| 3 | 
            +
            size 433270768
         | 
    	
        run-2/checkpoint-58/optimizer.pt
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:d8d6db0efd65543274e5f3051f234c479f850b113e5ed89fb1f71d4f89d17edf
         | 
| 3 | 
            +
            size 866662650
         | 
    	
        run-2/checkpoint-58/rng_state.pth
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:97b48f623e38a363d501c36fe8c4bc4a983ea51f88de73556367877c745a5e4c
         | 
| 3 | 
            +
            size 14244
         | 
    	
        run-2/checkpoint-58/scheduler.pt
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:56d8a52dcdbc99eb9ac1df472b4dd3923147a3ae998f5802cd5ec30e94b252d3
         | 
| 3 | 
            +
            size 1064
         | 
    	
        run-2/checkpoint-58/special_tokens_map.json
    ADDED
    
    | @@ -0,0 +1,37 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "cls_token": {
         | 
| 3 | 
            +
                "content": "[CLS]",
         | 
| 4 | 
            +
                "lstrip": false,
         | 
| 5 | 
            +
                "normalized": false,
         | 
| 6 | 
            +
                "rstrip": false,
         | 
| 7 | 
            +
                "single_word": false
         | 
| 8 | 
            +
              },
         | 
| 9 | 
            +
              "mask_token": {
         | 
| 10 | 
            +
                "content": "[MASK]",
         | 
| 11 | 
            +
                "lstrip": false,
         | 
| 12 | 
            +
                "normalized": false,
         | 
| 13 | 
            +
                "rstrip": false,
         | 
| 14 | 
            +
                "single_word": false
         | 
| 15 | 
            +
              },
         | 
| 16 | 
            +
              "pad_token": {
         | 
| 17 | 
            +
                "content": "[PAD]",
         | 
| 18 | 
            +
                "lstrip": false,
         | 
| 19 | 
            +
                "normalized": false,
         | 
| 20 | 
            +
                "rstrip": false,
         | 
| 21 | 
            +
                "single_word": false
         | 
| 22 | 
            +
              },
         | 
| 23 | 
            +
              "sep_token": {
         | 
| 24 | 
            +
                "content": "[SEP]",
         | 
| 25 | 
            +
                "lstrip": false,
         | 
| 26 | 
            +
                "normalized": false,
         | 
| 27 | 
            +
                "rstrip": false,
         | 
| 28 | 
            +
                "single_word": false
         | 
| 29 | 
            +
              },
         | 
| 30 | 
            +
              "unk_token": {
         | 
| 31 | 
            +
                "content": "[UNK]",
         | 
| 32 | 
            +
                "lstrip": false,
         | 
| 33 | 
            +
                "normalized": false,
         | 
| 34 | 
            +
                "rstrip": false,
         | 
| 35 | 
            +
                "single_word": false
         | 
| 36 | 
            +
              }
         | 
| 37 | 
            +
            }
         | 
    	
        run-2/checkpoint-58/tokenizer.json
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        run-2/checkpoint-58/tokenizer_config.json
    ADDED
    
    | @@ -0,0 +1,59 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "added_tokens_decoder": {
         | 
| 3 | 
            +
                "0": {
         | 
| 4 | 
            +
                  "content": "[PAD]",
         | 
| 5 | 
            +
                  "lstrip": false,
         | 
| 6 | 
            +
                  "normalized": false,
         | 
| 7 | 
            +
                  "rstrip": false,
         | 
| 8 | 
            +
                  "single_word": false,
         | 
| 9 | 
            +
                  "special": true
         | 
| 10 | 
            +
                },
         | 
| 11 | 
            +
                "100": {
         | 
| 12 | 
            +
                  "content": "[UNK]",
         | 
| 13 | 
            +
                  "lstrip": false,
         | 
| 14 | 
            +
                  "normalized": false,
         | 
| 15 | 
            +
                  "rstrip": false,
         | 
| 16 | 
            +
                  "single_word": false,
         | 
| 17 | 
            +
                  "special": true
         | 
| 18 | 
            +
                },
         | 
| 19 | 
            +
                "101": {
         | 
| 20 | 
            +
                  "content": "[CLS]",
         | 
| 21 | 
            +
                  "lstrip": false,
         | 
| 22 | 
            +
                  "normalized": false,
         | 
| 23 | 
            +
                  "rstrip": false,
         | 
| 24 | 
            +
                  "single_word": false,
         | 
| 25 | 
            +
                  "special": true
         | 
| 26 | 
            +
                },
         | 
| 27 | 
            +
                "102": {
         | 
| 28 | 
            +
                  "content": "[SEP]",
         | 
| 29 | 
            +
                  "lstrip": false,
         | 
| 30 | 
            +
                  "normalized": false,
         | 
| 31 | 
            +
                  "rstrip": false,
         | 
| 32 | 
            +
                  "single_word": false,
         | 
| 33 | 
            +
                  "special": true
         | 
| 34 | 
            +
                },
         | 
| 35 | 
            +
                "103": {
         | 
| 36 | 
            +
                  "content": "[MASK]",
         | 
| 37 | 
            +
                  "lstrip": false,
         | 
| 38 | 
            +
                  "normalized": false,
         | 
| 39 | 
            +
                  "rstrip": false,
         | 
| 40 | 
            +
                  "single_word": false,
         | 
| 41 | 
            +
                  "special": true
         | 
| 42 | 
            +
                }
         | 
| 43 | 
            +
              },
         | 
| 44 | 
            +
              "clean_up_tokenization_spaces": true,
         | 
| 45 | 
            +
              "cls_token": "[CLS]",
         | 
| 46 | 
            +
              "do_lower_case": false,
         | 
| 47 | 
            +
              "mask_token": "[MASK]",
         | 
| 48 | 
            +
              "max_length": 512,
         | 
| 49 | 
            +
              "model_max_length": 512,
         | 
| 50 | 
            +
              "pad_token": "[PAD]",
         | 
| 51 | 
            +
              "sep_token": "[SEP]",
         | 
| 52 | 
            +
              "stride": 0,
         | 
| 53 | 
            +
              "strip_accents": null,
         | 
| 54 | 
            +
              "tokenize_chinese_chars": true,
         | 
| 55 | 
            +
              "tokenizer_class": "BertTokenizer",
         | 
| 56 | 
            +
              "truncation_side": "right",
         | 
| 57 | 
            +
              "truncation_strategy": "longest_first",
         | 
| 58 | 
            +
              "unk_token": "[UNK]"
         | 
| 59 | 
            +
            }
         | 
    	
        run-2/checkpoint-58/trainer_state.json
    ADDED
    
    | @@ -0,0 +1,56 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "best_metric": 0.604,
         | 
| 3 | 
            +
              "best_model_checkpoint": "BERTForDetectingDepression-Twitter2020/run-2/checkpoint-58",
         | 
| 4 | 
            +
              "epoch": 2.0,
         | 
| 5 | 
            +
              "eval_steps": 500,
         | 
| 6 | 
            +
              "global_step": 58,
         | 
| 7 | 
            +
              "is_hyper_param_search": true,
         | 
| 8 | 
            +
              "is_local_process_zero": true,
         | 
| 9 | 
            +
              "is_world_process_zero": true,
         | 
| 10 | 
            +
              "log_history": [
         | 
| 11 | 
            +
                {
         | 
| 12 | 
            +
                  "epoch": 1.0,
         | 
| 13 | 
            +
                  "eval_accuracy": 0.591,
         | 
| 14 | 
            +
                  "eval_loss": 0.6620234251022339,
         | 
| 15 | 
            +
                  "eval_runtime": 3.0479,
         | 
| 16 | 
            +
                  "eval_samples_per_second": 656.189,
         | 
| 17 | 
            +
                  "eval_steps_per_second": 41.012,
         | 
| 18 | 
            +
                  "step": 29
         | 
| 19 | 
            +
                },
         | 
| 20 | 
            +
                {
         | 
| 21 | 
            +
                  "epoch": 2.0,
         | 
| 22 | 
            +
                  "eval_accuracy": 0.604,
         | 
| 23 | 
            +
                  "eval_loss": 0.6554160118103027,
         | 
| 24 | 
            +
                  "eval_runtime": 3.0697,
         | 
| 25 | 
            +
                  "eval_samples_per_second": 651.538,
         | 
| 26 | 
            +
                  "eval_steps_per_second": 40.721,
         | 
| 27 | 
            +
                  "step": 58
         | 
| 28 | 
            +
                }
         | 
| 29 | 
            +
              ],
         | 
| 30 | 
            +
              "logging_steps": 500,
         | 
| 31 | 
            +
              "max_steps": 87,
         | 
| 32 | 
            +
              "num_input_tokens_seen": 0,
         | 
| 33 | 
            +
              "num_train_epochs": 3,
         | 
| 34 | 
            +
              "save_steps": 500,
         | 
| 35 | 
            +
              "stateful_callbacks": {
         | 
| 36 | 
            +
                "TrainerControl": {
         | 
| 37 | 
            +
                  "args": {
         | 
| 38 | 
            +
                    "should_epoch_stop": false,
         | 
| 39 | 
            +
                    "should_evaluate": false,
         | 
| 40 | 
            +
                    "should_log": false,
         | 
| 41 | 
            +
                    "should_save": true,
         | 
| 42 | 
            +
                    "should_training_stop": false
         | 
| 43 | 
            +
                  },
         | 
| 44 | 
            +
                  "attributes": {}
         | 
| 45 | 
            +
                }
         | 
| 46 | 
            +
              },
         | 
| 47 | 
            +
              "total_flos": 0,
         | 
| 48 | 
            +
              "train_batch_size": 64,
         | 
| 49 | 
            +
              "trial_name": null,
         | 
| 50 | 
            +
              "trial_params": {
         | 
| 51 | 
            +
                "learning_rate": 2.0937702506046973e-05,
         | 
| 52 | 
            +
                "num_train_epochs": 3,
         | 
| 53 | 
            +
                "per_device_train_batch_size": 64,
         | 
| 54 | 
            +
                "seed": 12
         | 
| 55 | 
            +
              }
         | 
| 56 | 
            +
            }
         | 
    	
        run-2/checkpoint-58/training_args.bin
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:f09db55f5a58c50b427033c39fc8017a6d9bf92a954a6b68eaa7be80ef4a39a6
         | 
| 3 | 
            +
            size 5176
         | 
    	
        run-2/checkpoint-58/vocab.txt
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        run-2/checkpoint-87/config.json
    ADDED
    
    | @@ -0,0 +1,27 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_name_or_path": "AIMH/mental-bert-base-cased",
         | 
| 3 | 
            +
              "architectures": [
         | 
| 4 | 
            +
                "BertForSequenceClassification"
         | 
| 5 | 
            +
              ],
         | 
| 6 | 
            +
              "attention_probs_dropout_prob": 0.1,
         | 
| 7 | 
            +
              "classifier_dropout": null,
         | 
| 8 | 
            +
              "gradient_checkpointing": false,
         | 
| 9 | 
            +
              "hidden_act": "gelu",
         | 
| 10 | 
            +
              "hidden_dropout_prob": 0.1,
         | 
| 11 | 
            +
              "hidden_size": 768,
         | 
| 12 | 
            +
              "initializer_range": 0.02,
         | 
| 13 | 
            +
              "intermediate_size": 3072,
         | 
| 14 | 
            +
              "layer_norm_eps": 1e-12,
         | 
| 15 | 
            +
              "max_position_embeddings": 512,
         | 
| 16 | 
            +
              "model_type": "bert",
         | 
| 17 | 
            +
              "num_attention_heads": 12,
         | 
| 18 | 
            +
              "num_hidden_layers": 12,
         | 
| 19 | 
            +
              "pad_token_id": 0,
         | 
| 20 | 
            +
              "position_embedding_type": "absolute",
         | 
| 21 | 
            +
              "problem_type": "single_label_classification",
         | 
| 22 | 
            +
              "torch_dtype": "float32",
         | 
| 23 | 
            +
              "transformers_version": "4.42.4",
         | 
| 24 | 
            +
              "type_vocab_size": 2,
         | 
| 25 | 
            +
              "use_cache": true,
         | 
| 26 | 
            +
              "vocab_size": 28996
         | 
| 27 | 
            +
            }
         | 
    	
        run-2/checkpoint-87/model.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:02fdb7e3f08881ceb5e97f78a746d3e989e4e85e59201882824940befbf97aa0
         | 
| 3 | 
            +
            size 433270768
         | 
    	
        run-2/checkpoint-87/optimizer.pt
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:d0893a19b2c783f7b8407fb3195f198bd007a395feed2b6c880afb13d26fccca
         | 
| 3 | 
            +
            size 866662650
         | 
    	
        run-2/checkpoint-87/rng_state.pth
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:19d622a0e31f1454dfd90e8656011fba87df11138c8317eb6499c2af9956ea69
         | 
| 3 | 
            +
            size 14244
         | 
    	
        run-2/checkpoint-87/scheduler.pt
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:9055753750e928908df3f843f3d3e08044bfe994fe9b1e290250f9c52c233510
         | 
| 3 | 
            +
            size 1064
         | 
    	
        run-2/checkpoint-87/special_tokens_map.json
    ADDED
    
    | @@ -0,0 +1,37 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "cls_token": {
         | 
| 3 | 
            +
                "content": "[CLS]",
         | 
| 4 | 
            +
                "lstrip": false,
         | 
| 5 | 
            +
                "normalized": false,
         | 
| 6 | 
            +
                "rstrip": false,
         | 
| 7 | 
            +
                "single_word": false
         | 
| 8 | 
            +
              },
         | 
| 9 | 
            +
              "mask_token": {
         | 
| 10 | 
            +
                "content": "[MASK]",
         | 
| 11 | 
            +
                "lstrip": false,
         | 
| 12 | 
            +
                "normalized": false,
         | 
| 13 | 
            +
                "rstrip": false,
         | 
| 14 | 
            +
                "single_word": false
         | 
| 15 | 
            +
              },
         | 
| 16 | 
            +
              "pad_token": {
         | 
| 17 | 
            +
                "content": "[PAD]",
         | 
| 18 | 
            +
                "lstrip": false,
         | 
| 19 | 
            +
                "normalized": false,
         | 
| 20 | 
            +
                "rstrip": false,
         | 
| 21 | 
            +
                "single_word": false
         | 
| 22 | 
            +
              },
         | 
| 23 | 
            +
              "sep_token": {
         | 
| 24 | 
            +
                "content": "[SEP]",
         | 
| 25 | 
            +
                "lstrip": false,
         | 
| 26 | 
            +
                "normalized": false,
         | 
| 27 | 
            +
                "rstrip": false,
         | 
| 28 | 
            +
                "single_word": false
         | 
| 29 | 
            +
              },
         | 
| 30 | 
            +
              "unk_token": {
         | 
| 31 | 
            +
                "content": "[UNK]",
         | 
| 32 | 
            +
                "lstrip": false,
         | 
| 33 | 
            +
                "normalized": false,
         | 
| 34 | 
            +
                "rstrip": false,
         | 
| 35 | 
            +
                "single_word": false
         | 
| 36 | 
            +
              }
         | 
| 37 | 
            +
            }
         | 
    	
        run-2/checkpoint-87/tokenizer.json
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        run-2/checkpoint-87/tokenizer_config.json
    ADDED
    
    | @@ -0,0 +1,59 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "added_tokens_decoder": {
         | 
| 3 | 
            +
                "0": {
         | 
| 4 | 
            +
                  "content": "[PAD]",
         | 
| 5 | 
            +
                  "lstrip": false,
         | 
| 6 | 
            +
                  "normalized": false,
         | 
| 7 | 
            +
                  "rstrip": false,
         | 
| 8 | 
            +
                  "single_word": false,
         | 
| 9 | 
            +
                  "special": true
         | 
| 10 | 
            +
                },
         | 
| 11 | 
            +
                "100": {
         | 
| 12 | 
            +
                  "content": "[UNK]",
         | 
| 13 | 
            +
                  "lstrip": false,
         | 
| 14 | 
            +
                  "normalized": false,
         | 
| 15 | 
            +
                  "rstrip": false,
         | 
| 16 | 
            +
                  "single_word": false,
         | 
| 17 | 
            +
                  "special": true
         | 
| 18 | 
            +
                },
         | 
| 19 | 
            +
                "101": {
         | 
| 20 | 
            +
                  "content": "[CLS]",
         | 
| 21 | 
            +
                  "lstrip": false,
         | 
| 22 | 
            +
                  "normalized": false,
         | 
| 23 | 
            +
                  "rstrip": false,
         | 
| 24 | 
            +
                  "single_word": false,
         | 
| 25 | 
            +
                  "special": true
         | 
| 26 | 
            +
                },
         | 
| 27 | 
            +
                "102": {
         | 
| 28 | 
            +
                  "content": "[SEP]",
         | 
| 29 | 
            +
                  "lstrip": false,
         | 
| 30 | 
            +
                  "normalized": false,
         | 
| 31 | 
            +
                  "rstrip": false,
         | 
| 32 | 
            +
                  "single_word": false,
         | 
| 33 | 
            +
                  "special": true
         | 
| 34 | 
            +
                },
         | 
| 35 | 
            +
                "103": {
         | 
| 36 | 
            +
                  "content": "[MASK]",
         | 
| 37 | 
            +
                  "lstrip": false,
         | 
| 38 | 
            +
                  "normalized": false,
         | 
| 39 | 
            +
                  "rstrip": false,
         | 
| 40 | 
            +
                  "single_word": false,
         | 
| 41 | 
            +
                  "special": true
         | 
| 42 | 
            +
                }
         | 
| 43 | 
            +
              },
         | 
| 44 | 
            +
              "clean_up_tokenization_spaces": true,
         | 
| 45 | 
            +
              "cls_token": "[CLS]",
         | 
| 46 | 
            +
              "do_lower_case": false,
         | 
| 47 | 
            +
              "mask_token": "[MASK]",
         | 
| 48 | 
            +
              "max_length": 512,
         | 
| 49 | 
            +
              "model_max_length": 512,
         | 
| 50 | 
            +
              "pad_token": "[PAD]",
         | 
| 51 | 
            +
              "sep_token": "[SEP]",
         | 
| 52 | 
            +
              "stride": 0,
         | 
| 53 | 
            +
              "strip_accents": null,
         | 
| 54 | 
            +
              "tokenize_chinese_chars": true,
         | 
| 55 | 
            +
              "tokenizer_class": "BertTokenizer",
         | 
| 56 | 
            +
              "truncation_side": "right",
         | 
| 57 | 
            +
              "truncation_strategy": "longest_first",
         | 
| 58 | 
            +
              "unk_token": "[UNK]"
         | 
| 59 | 
            +
            }
         | 
    	
        run-2/checkpoint-87/trainer_state.json
    ADDED
    
    | @@ -0,0 +1,65 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "best_metric": 0.6085,
         | 
| 3 | 
            +
              "best_model_checkpoint": "BERTForDetectingDepression-Twitter2020/run-2/checkpoint-87",
         | 
| 4 | 
            +
              "epoch": 3.0,
         | 
| 5 | 
            +
              "eval_steps": 500,
         | 
| 6 | 
            +
              "global_step": 87,
         | 
| 7 | 
            +
              "is_hyper_param_search": true,
         | 
| 8 | 
            +
              "is_local_process_zero": true,
         | 
| 9 | 
            +
              "is_world_process_zero": true,
         | 
| 10 | 
            +
              "log_history": [
         | 
| 11 | 
            +
                {
         | 
| 12 | 
            +
                  "epoch": 1.0,
         | 
| 13 | 
            +
                  "eval_accuracy": 0.591,
         | 
| 14 | 
            +
                  "eval_loss": 0.6620234251022339,
         | 
| 15 | 
            +
                  "eval_runtime": 3.0479,
         | 
| 16 | 
            +
                  "eval_samples_per_second": 656.189,
         | 
| 17 | 
            +
                  "eval_steps_per_second": 41.012,
         | 
| 18 | 
            +
                  "step": 29
         | 
| 19 | 
            +
                },
         | 
| 20 | 
            +
                {
         | 
| 21 | 
            +
                  "epoch": 2.0,
         | 
| 22 | 
            +
                  "eval_accuracy": 0.604,
         | 
| 23 | 
            +
                  "eval_loss": 0.6554160118103027,
         | 
| 24 | 
            +
                  "eval_runtime": 3.0697,
         | 
| 25 | 
            +
                  "eval_samples_per_second": 651.538,
         | 
| 26 | 
            +
                  "eval_steps_per_second": 40.721,
         | 
| 27 | 
            +
                  "step": 58
         | 
| 28 | 
            +
                },
         | 
| 29 | 
            +
                {
         | 
| 30 | 
            +
                  "epoch": 3.0,
         | 
| 31 | 
            +
                  "eval_accuracy": 0.6085,
         | 
| 32 | 
            +
                  "eval_loss": 0.6581469178199768,
         | 
| 33 | 
            +
                  "eval_runtime": 3.1368,
         | 
| 34 | 
            +
                  "eval_samples_per_second": 637.583,
         | 
| 35 | 
            +
                  "eval_steps_per_second": 39.849,
         | 
| 36 | 
            +
                  "step": 87
         | 
| 37 | 
            +
                }
         | 
| 38 | 
            +
              ],
         | 
| 39 | 
            +
              "logging_steps": 500,
         | 
| 40 | 
            +
              "max_steps": 87,
         | 
| 41 | 
            +
              "num_input_tokens_seen": 0,
         | 
| 42 | 
            +
              "num_train_epochs": 3,
         | 
| 43 | 
            +
              "save_steps": 500,
         | 
| 44 | 
            +
              "stateful_callbacks": {
         | 
| 45 | 
            +
                "TrainerControl": {
         | 
| 46 | 
            +
                  "args": {
         | 
| 47 | 
            +
                    "should_epoch_stop": false,
         | 
| 48 | 
            +
                    "should_evaluate": false,
         | 
| 49 | 
            +
                    "should_log": false,
         | 
| 50 | 
            +
                    "should_save": true,
         | 
| 51 | 
            +
                    "should_training_stop": true
         | 
| 52 | 
            +
                  },
         | 
| 53 | 
            +
                  "attributes": {}
         | 
| 54 | 
            +
                }
         | 
| 55 | 
            +
              },
         | 
| 56 | 
            +
              "total_flos": 0,
         | 
| 57 | 
            +
              "train_batch_size": 64,
         | 
| 58 | 
            +
              "trial_name": null,
         | 
| 59 | 
            +
              "trial_params": {
         | 
| 60 | 
            +
                "learning_rate": 2.0937702506046973e-05,
         | 
| 61 | 
            +
                "num_train_epochs": 3,
         | 
| 62 | 
            +
                "per_device_train_batch_size": 64,
         | 
| 63 | 
            +
                "seed": 12
         | 
| 64 | 
            +
              }
         | 
| 65 | 
            +
            }
         | 
    	
        run-2/checkpoint-87/training_args.bin
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:f09db55f5a58c50b427033c39fc8017a6d9bf92a954a6b68eaa7be80ef4a39a6
         | 
| 3 | 
            +
            size 5176
         | 
    	
        run-2/checkpoint-87/vocab.txt
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        run-3/checkpoint-29/config.json
    ADDED
    
    | @@ -0,0 +1,27 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_name_or_path": "AIMH/mental-bert-base-cased",
         | 
| 3 | 
            +
              "architectures": [
         | 
| 4 | 
            +
                "BertForSequenceClassification"
         | 
| 5 | 
            +
              ],
         | 
| 6 | 
            +
              "attention_probs_dropout_prob": 0.1,
         | 
| 7 | 
            +
              "classifier_dropout": null,
         | 
| 8 | 
            +
              "gradient_checkpointing": false,
         | 
| 9 | 
            +
              "hidden_act": "gelu",
         | 
| 10 | 
            +
              "hidden_dropout_prob": 0.1,
         | 
| 11 | 
            +
              "hidden_size": 768,
         | 
| 12 | 
            +
              "initializer_range": 0.02,
         | 
| 13 | 
            +
              "intermediate_size": 3072,
         | 
| 14 | 
            +
              "layer_norm_eps": 1e-12,
         | 
| 15 | 
            +
              "max_position_embeddings": 512,
         | 
| 16 | 
            +
              "model_type": "bert",
         | 
| 17 | 
            +
              "num_attention_heads": 12,
         | 
| 18 | 
            +
              "num_hidden_layers": 12,
         | 
| 19 | 
            +
              "pad_token_id": 0,
         | 
| 20 | 
            +
              "position_embedding_type": "absolute",
         | 
| 21 | 
            +
              "problem_type": "single_label_classification",
         | 
| 22 | 
            +
              "torch_dtype": "float32",
         | 
| 23 | 
            +
              "transformers_version": "4.42.4",
         | 
| 24 | 
            +
              "type_vocab_size": 2,
         | 
| 25 | 
            +
              "use_cache": true,
         | 
| 26 | 
            +
              "vocab_size": 28996
         | 
| 27 | 
            +
            }
         | 
    	
        run-3/checkpoint-29/model.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:4982e2b41506ddf13232beab76be2fa3415c215ada31514b02a193021a1e5f15
         | 
| 3 | 
            +
            size 433270768
         | 
    	
        run-3/checkpoint-29/optimizer.pt
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:49b16e043f475cf92b7184aded02aca3435697022ca088d62d7c1c87a5d3d603
         | 
| 3 | 
            +
            size 866662650
         | 
    	
        run-3/checkpoint-29/rng_state.pth
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:cab59dbc2b7fcdf027f9c5f942fd746d9187d76636c2efb570b6ca0734df8f13
         | 
| 3 | 
            +
            size 14308
         | 
    	
        run-3/checkpoint-29/scheduler.pt
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:136674cb95a8495ea25e25c94b383047cdda3841d03222e65c61e3c0f6d716a5
         | 
| 3 | 
            +
            size 1064
         | 
    	
        run-3/checkpoint-29/special_tokens_map.json
    ADDED
    
    | @@ -0,0 +1,37 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "cls_token": {
         | 
| 3 | 
            +
                "content": "[CLS]",
         | 
| 4 | 
            +
                "lstrip": false,
         | 
| 5 | 
            +
                "normalized": false,
         | 
| 6 | 
            +
                "rstrip": false,
         | 
| 7 | 
            +
                "single_word": false
         | 
| 8 | 
            +
              },
         | 
| 9 | 
            +
              "mask_token": {
         | 
| 10 | 
            +
                "content": "[MASK]",
         | 
| 11 | 
            +
                "lstrip": false,
         | 
| 12 | 
            +
                "normalized": false,
         | 
| 13 | 
            +
                "rstrip": false,
         | 
| 14 | 
            +
                "single_word": false
         | 
| 15 | 
            +
              },
         | 
| 16 | 
            +
              "pad_token": {
         | 
| 17 | 
            +
                "content": "[PAD]",
         | 
| 18 | 
            +
                "lstrip": false,
         | 
| 19 | 
            +
                "normalized": false,
         | 
| 20 | 
            +
                "rstrip": false,
         | 
| 21 | 
            +
                "single_word": false
         | 
| 22 | 
            +
              },
         | 
| 23 | 
            +
              "sep_token": {
         | 
| 24 | 
            +
                "content": "[SEP]",
         | 
| 25 | 
            +
                "lstrip": false,
         | 
| 26 | 
            +
                "normalized": false,
         | 
| 27 | 
            +
                "rstrip": false,
         | 
| 28 | 
            +
                "single_word": false
         | 
| 29 | 
            +
              },
         | 
| 30 | 
            +
              "unk_token": {
         | 
| 31 | 
            +
                "content": "[UNK]",
         | 
| 32 | 
            +
                "lstrip": false,
         | 
| 33 | 
            +
                "normalized": false,
         | 
| 34 | 
            +
                "rstrip": false,
         | 
| 35 | 
            +
                "single_word": false
         | 
| 36 | 
            +
              }
         | 
| 37 | 
            +
            }
         | 
    	
        run-3/checkpoint-29/tokenizer.json
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        run-3/checkpoint-29/tokenizer_config.json
    ADDED
    
    | @@ -0,0 +1,59 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "added_tokens_decoder": {
         | 
| 3 | 
            +
                "0": {
         | 
| 4 | 
            +
                  "content": "[PAD]",
         | 
| 5 | 
            +
                  "lstrip": false,
         | 
| 6 | 
            +
                  "normalized": false,
         | 
| 7 | 
            +
                  "rstrip": false,
         | 
| 8 | 
            +
                  "single_word": false,
         | 
| 9 | 
            +
                  "special": true
         | 
| 10 | 
            +
                },
         | 
| 11 | 
            +
                "100": {
         | 
| 12 | 
            +
                  "content": "[UNK]",
         | 
| 13 | 
            +
                  "lstrip": false,
         | 
| 14 | 
            +
                  "normalized": false,
         | 
| 15 | 
            +
                  "rstrip": false,
         | 
| 16 | 
            +
                  "single_word": false,
         | 
| 17 | 
            +
                  "special": true
         | 
| 18 | 
            +
                },
         | 
| 19 | 
            +
                "101": {
         | 
| 20 | 
            +
                  "content": "[CLS]",
         | 
| 21 | 
            +
                  "lstrip": false,
         | 
| 22 | 
            +
                  "normalized": false,
         | 
| 23 | 
            +
                  "rstrip": false,
         | 
| 24 | 
            +
                  "single_word": false,
         | 
| 25 | 
            +
                  "special": true
         | 
| 26 | 
            +
                },
         | 
| 27 | 
            +
                "102": {
         | 
| 28 | 
            +
                  "content": "[SEP]",
         | 
| 29 | 
            +
                  "lstrip": false,
         | 
| 30 | 
            +
                  "normalized": false,
         | 
| 31 | 
            +
                  "rstrip": false,
         | 
| 32 | 
            +
                  "single_word": false,
         | 
| 33 | 
            +
                  "special": true
         | 
| 34 | 
            +
                },
         | 
| 35 | 
            +
                "103": {
         | 
| 36 | 
            +
                  "content": "[MASK]",
         | 
| 37 | 
            +
                  "lstrip": false,
         | 
| 38 | 
            +
                  "normalized": false,
         | 
| 39 | 
            +
                  "rstrip": false,
         | 
| 40 | 
            +
                  "single_word": false,
         | 
| 41 | 
            +
                  "special": true
         | 
| 42 | 
            +
                }
         | 
| 43 | 
            +
              },
         | 
| 44 | 
            +
              "clean_up_tokenization_spaces": true,
         | 
| 45 | 
            +
              "cls_token": "[CLS]",
         | 
| 46 | 
            +
              "do_lower_case": false,
         | 
| 47 | 
            +
              "mask_token": "[MASK]",
         | 
| 48 | 
            +
              "max_length": 512,
         | 
| 49 | 
            +
              "model_max_length": 512,
         | 
| 50 | 
            +
              "pad_token": "[PAD]",
         | 
| 51 | 
            +
              "sep_token": "[SEP]",
         | 
| 52 | 
            +
              "stride": 0,
         | 
| 53 | 
            +
              "strip_accents": null,
         | 
| 54 | 
            +
              "tokenize_chinese_chars": true,
         | 
| 55 | 
            +
              "tokenizer_class": "BertTokenizer",
         | 
| 56 | 
            +
              "truncation_side": "right",
         | 
| 57 | 
            +
              "truncation_strategy": "longest_first",
         | 
| 58 | 
            +
              "unk_token": "[UNK]"
         | 
| 59 | 
            +
            }
         | 
    	
        run-3/checkpoint-29/trainer_state.json
    ADDED
    
    | @@ -0,0 +1,47 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "best_metric": 0.5795,
         | 
| 3 | 
            +
              "best_model_checkpoint": "BERTForDetectingDepression-Twitter2020/run-3/checkpoint-29",
         | 
| 4 | 
            +
              "epoch": 1.0,
         | 
| 5 | 
            +
              "eval_steps": 500,
         | 
| 6 | 
            +
              "global_step": 29,
         | 
| 7 | 
            +
              "is_hyper_param_search": true,
         | 
| 8 | 
            +
              "is_local_process_zero": true,
         | 
| 9 | 
            +
              "is_world_process_zero": true,
         | 
| 10 | 
            +
              "log_history": [
         | 
| 11 | 
            +
                {
         | 
| 12 | 
            +
                  "epoch": 1.0,
         | 
| 13 | 
            +
                  "eval_accuracy": 0.5795,
         | 
| 14 | 
            +
                  "eval_loss": 0.6693535447120667,
         | 
| 15 | 
            +
                  "eval_runtime": 3.1138,
         | 
| 16 | 
            +
                  "eval_samples_per_second": 642.306,
         | 
| 17 | 
            +
                  "eval_steps_per_second": 40.144,
         | 
| 18 | 
            +
                  "step": 29
         | 
| 19 | 
            +
                }
         | 
| 20 | 
            +
              ],
         | 
| 21 | 
            +
              "logging_steps": 500,
         | 
| 22 | 
            +
              "max_steps": 29,
         | 
| 23 | 
            +
              "num_input_tokens_seen": 0,
         | 
| 24 | 
            +
              "num_train_epochs": 1,
         | 
| 25 | 
            +
              "save_steps": 500,
         | 
| 26 | 
            +
              "stateful_callbacks": {
         | 
| 27 | 
            +
                "TrainerControl": {
         | 
| 28 | 
            +
                  "args": {
         | 
| 29 | 
            +
                    "should_epoch_stop": false,
         | 
| 30 | 
            +
                    "should_evaluate": false,
         | 
| 31 | 
            +
                    "should_log": false,
         | 
| 32 | 
            +
                    "should_save": true,
         | 
| 33 | 
            +
                    "should_training_stop": true
         | 
| 34 | 
            +
                  },
         | 
| 35 | 
            +
                  "attributes": {}
         | 
| 36 | 
            +
                }
         | 
| 37 | 
            +
              },
         | 
| 38 | 
            +
              "total_flos": 0,
         | 
| 39 | 
            +
              "train_batch_size": 64,
         | 
| 40 | 
            +
              "trial_name": null,
         | 
| 41 | 
            +
              "trial_params": {
         | 
| 42 | 
            +
                "learning_rate": 7.916782335882902e-05,
         | 
| 43 | 
            +
                "num_train_epochs": 1,
         | 
| 44 | 
            +
                "per_device_train_batch_size": 64,
         | 
| 45 | 
            +
                "seed": 26
         | 
| 46 | 
            +
              }
         | 
| 47 | 
            +
            }
         | 
    	
        run-3/checkpoint-29/training_args.bin
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:76e74b88a99996021c40214a52388477c30fd978e5cc29af1fb108a76756da46
         | 
| 3 | 
            +
            size 5176
         | 
    	
        run-3/checkpoint-29/vocab.txt
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        runs/Aug01_13-47-28_1d376fbba8fb/events.out.tfevents.1722520428.1d376fbba8fb.194.7
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:22d1232fc3639fa73dfe90cf2d30e0d2418ccb57467ee3ae209c7b54e5ac3d10
         | 
| 3 | 
            +
            size 5613
         | 
    	
        training_args.bin
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
             
            size 5176
         | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:76e74b88a99996021c40214a52388477c30fd978e5cc29af1fb108a76756da46
         | 
| 3 | 
             
            size 5176
         |