New model from https://wandb.ai/wandb/huggingtweets/runs/k0swqzf3
Browse files- README.md +10 -10
 - config.json +1 -1
 - generation_config.json +1 -1
 - pytorch_model.bin +1 -1
 - tokenizer_config.json +1 -1
 - training_args.bin +1 -1
 
    	
        README.md
    CHANGED
    
    | 
         @@ -20,7 +20,7 @@ widget: 
     | 
|
| 20 | 
         
             
                    </div>
         
     | 
| 21 | 
         
             
                </div>
         
     | 
| 22 | 
         
             
                <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div>
         
     | 
| 23 | 
         
            -
                <div style="text-align: center; font-size: 16px; font-weight: 800"> 
     | 
| 24 | 
         
             
                <div style="text-align: center; font-size: 14px;">@dril</div>
         
     | 
| 25 | 
         
             
            </div>
         
     | 
| 26 | 
         | 
| 
         @@ -38,24 +38,24 @@ To understand how the model was developed, check the [W&B report](https://wandb. 
     | 
|
| 38 | 
         | 
| 39 | 
         
             
            ## Training data
         
     | 
| 40 | 
         | 
| 41 | 
         
            -
            The model was trained on tweets from  
     | 
| 42 | 
         | 
| 43 | 
         
            -
            | Data |  
     | 
| 44 | 
         
             
            | --- | --- |
         
     | 
| 45 | 
         
            -
            | Tweets downloaded |  
     | 
| 46 | 
         
            -
            | Retweets |  
     | 
| 47 | 
         
            -
            | Short tweets |  
     | 
| 48 | 
         
            -
            | Tweets kept |  
     | 
| 49 | 
         | 
| 50 | 
         
            -
            [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/ 
     | 
| 51 | 
         | 
| 52 | 
         
             
            ## Training procedure
         
     | 
| 53 | 
         | 
| 54 | 
         
             
            The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @dril's tweets.
         
     | 
| 55 | 
         | 
| 56 | 
         
            -
            Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/ 
     | 
| 57 | 
         | 
| 58 | 
         
            -
            At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/ 
     | 
| 59 | 
         | 
| 60 | 
         
             
            ## How to use
         
     | 
| 61 | 
         | 
| 
         | 
|
| 20 | 
         
             
                    </div>
         
     | 
| 21 | 
         
             
                </div>
         
     | 
| 22 | 
         
             
                <div style="text-align: center; margin-top: 3px; font-size: 16px; font-weight: 800">🤖 AI BOT 🤖</div>
         
     | 
| 23 | 
         
            +
                <div style="text-align: center; font-size: 16px; font-weight: 800">slave to Woke</div>
         
     | 
| 24 | 
         
             
                <div style="text-align: center; font-size: 14px;">@dril</div>
         
     | 
| 25 | 
         
             
            </div>
         
     | 
| 26 | 
         | 
| 
         | 
|
| 38 | 
         | 
| 39 | 
         
             
            ## Training data
         
     | 
| 40 | 
         | 
| 41 | 
         
            +
            The model was trained on tweets from slave to Woke.
         
     | 
| 42 | 
         | 
| 43 | 
         
            +
            | Data | slave to Woke |
         
     | 
| 44 | 
         
             
            | --- | --- |
         
     | 
| 45 | 
         
            +
            | Tweets downloaded | 3191 |
         
     | 
| 46 | 
         
            +
            | Retweets | 512 |
         
     | 
| 47 | 
         
            +
            | Short tweets | 274 |
         
     | 
| 48 | 
         
            +
            | Tweets kept | 2405 |
         
     | 
| 49 | 
         | 
| 50 | 
         
            +
            [Explore the data](https://wandb.ai/wandb/huggingtweets/runs/a5mb4z84/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline.
         
     | 
| 51 | 
         | 
| 52 | 
         
             
            ## Training procedure
         
     | 
| 53 | 
         | 
| 54 | 
         
             
            The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on @dril's tweets.
         
     | 
| 55 | 
         | 
| 56 | 
         
            +
            Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/wandb/huggingtweets/runs/k0swqzf3) for full transparency and reproducibility.
         
     | 
| 57 | 
         | 
| 58 | 
         
            +
            At the end of training, [the final model](https://wandb.ai/wandb/huggingtweets/runs/k0swqzf3/artifacts) is logged and versioned.
         
     | 
| 59 | 
         | 
| 60 | 
         
             
            ## How to use
         
     | 
| 61 | 
         | 
    	
        config.json
    CHANGED
    
    | 
         @@ -37,7 +37,7 @@ 
     | 
|
| 37 | 
         
             
                }
         
     | 
| 38 | 
         
             
              },
         
     | 
| 39 | 
         
             
              "torch_dtype": "float32",
         
     | 
| 40 | 
         
            -
              "transformers_version": "4. 
     | 
| 41 | 
         
             
              "use_cache": true,
         
     | 
| 42 | 
         
             
              "vocab_size": 50257
         
     | 
| 43 | 
         
             
            }
         
     | 
| 
         | 
|
| 37 | 
         
             
                }
         
     | 
| 38 | 
         
             
              },
         
     | 
| 39 | 
         
             
              "torch_dtype": "float32",
         
     | 
| 40 | 
         
            +
              "transformers_version": "4.28.1",
         
     | 
| 41 | 
         
             
              "use_cache": true,
         
     | 
| 42 | 
         
             
              "vocab_size": 50257
         
     | 
| 43 | 
         
             
            }
         
     | 
    	
        generation_config.json
    CHANGED
    
    | 
         @@ -2,5 +2,5 @@ 
     | 
|
| 2 | 
         
             
              "_from_model_config": true,
         
     | 
| 3 | 
         
             
              "bos_token_id": 50256,
         
     | 
| 4 | 
         
             
              "eos_token_id": 50256,
         
     | 
| 5 | 
         
            -
              "transformers_version": "4. 
     | 
| 6 | 
         
             
            }
         
     | 
| 
         | 
|
| 2 | 
         
             
              "_from_model_config": true,
         
     | 
| 3 | 
         
             
              "bos_token_id": 50256,
         
     | 
| 4 | 
         
             
              "eos_token_id": 50256,
         
     | 
| 5 | 
         
            +
              "transformers_version": "4.28.1"
         
     | 
| 6 | 
         
             
            }
         
     | 
    	
        pytorch_model.bin
    CHANGED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
             
            size 510398013
         
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:6444af3859c305d57eea54eec41ead554a9b9b7b7a0bdb869e9eb578b3eb613b
         
     | 
| 3 | 
         
             
            size 510398013
         
     | 
    	
        tokenizer_config.json
    CHANGED
    
    | 
         @@ -1,9 +1,9 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "add_prefix_space": false,
         
     | 
| 3 | 
         
             
              "bos_token": "<|endoftext|>",
         
     | 
| 
         | 
|
| 4 | 
         
             
              "eos_token": "<|endoftext|>",
         
     | 
| 5 | 
         
             
              "model_max_length": 1024,
         
     | 
| 6 | 
         
            -
              "special_tokens_map_file": null,
         
     | 
| 7 | 
         
             
              "tokenizer_class": "GPT2Tokenizer",
         
     | 
| 8 | 
         
             
              "unk_token": "<|endoftext|>"
         
     | 
| 9 | 
         
             
            }
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "add_prefix_space": false,
         
     | 
| 3 | 
         
             
              "bos_token": "<|endoftext|>",
         
     | 
| 4 | 
         
            +
              "clean_up_tokenization_spaces": true,
         
     | 
| 5 | 
         
             
              "eos_token": "<|endoftext|>",
         
     | 
| 6 | 
         
             
              "model_max_length": 1024,
         
     | 
| 
         | 
|
| 7 | 
         
             
              "tokenizer_class": "GPT2Tokenizer",
         
     | 
| 8 | 
         
             
              "unk_token": "<|endoftext|>"
         
     | 
| 9 | 
         
             
            }
         
     | 
    	
        training_args.bin
    CHANGED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
             
            size 3579
         
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:89195ccee0e48abd5b134d19f89ba0a68c3ced43976e93363c7f4a0dad65b112
         
     | 
| 3 | 
         
             
            size 3579
         
     |