End of training
Browse files
    	
        README.md
    CHANGED
    
    | @@ -35,11 +35,11 @@ This model was trained with SFT. | |
| 35 |  | 
| 36 | 
             
            ### Framework versions
         | 
| 37 |  | 
| 38 | 
            -
            - TRL: 0. | 
| 39 | 
            -
            - Transformers: 4. | 
| 40 | 
             
            - Pytorch: 2.7.0
         | 
| 41 | 
             
            - Datasets: 3.6.0
         | 
| 42 | 
            -
            - Tokenizers: 0.21. | 
| 43 |  | 
| 44 | 
             
            ## Citations
         | 
| 45 |  | 
|  | |
| 35 |  | 
| 36 | 
             
            ### Framework versions
         | 
| 37 |  | 
| 38 | 
            +
            - TRL: 0.21.0
         | 
| 39 | 
            +
            - Transformers: 4.55.0
         | 
| 40 | 
             
            - Pytorch: 2.7.0
         | 
| 41 | 
             
            - Datasets: 3.6.0
         | 
| 42 | 
            +
            - Tokenizers: 0.21.4
         | 
| 43 |  | 
| 44 | 
             
            ## Citations
         | 
| 45 |  | 
    	
        adapter_config.json
    CHANGED
    
    | @@ -25,14 +25,15 @@ | |
| 25 | 
             
              "rank_pattern": {},
         | 
| 26 | 
             
              "revision": null,
         | 
| 27 | 
             
              "target_modules": [
         | 
|  | |
|  | |
|  | |
| 28 | 
             
                "v_proj",
         | 
| 29 | 
             
                "o_proj",
         | 
| 30 | 
            -
                "gate_proj",
         | 
| 31 | 
            -
                "up_proj",
         | 
| 32 | 
            -
                "q_proj",
         | 
| 33 | 
             
                "k_proj",
         | 
| 34 | 
            -
                " | 
| 35 | 
             
              ],
         | 
|  | |
| 36 | 
             
              "task_type": "CAUSAL_LM",
         | 
| 37 | 
             
              "trainable_token_indices": null,
         | 
| 38 | 
             
              "use_dora": false,
         | 
|  | |
| 25 | 
             
              "rank_pattern": {},
         | 
| 26 | 
             
              "revision": null,
         | 
| 27 | 
             
              "target_modules": [
         | 
| 28 | 
            +
                "q_proj",
         | 
| 29 | 
            +
                "up_proj",
         | 
| 30 | 
            +
                "down_proj",
         | 
| 31 | 
             
                "v_proj",
         | 
| 32 | 
             
                "o_proj",
         | 
|  | |
|  | |
|  | |
| 33 | 
             
                "k_proj",
         | 
| 34 | 
            +
                "gate_proj"
         | 
| 35 | 
             
              ],
         | 
| 36 | 
            +
              "target_parameters": null,
         | 
| 37 | 
             
              "task_type": "CAUSAL_LM",
         | 
| 38 | 
             
              "trainable_token_indices": null,
         | 
| 39 | 
             
              "use_dora": false,
         | 
    	
        adapter_model.safetensors
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
             
            size 194563400
         | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:b1a78aa4966aed68a01b6e118d9be6af2e46e7b09fbc0f01a3464a62537ba645
         | 
| 3 | 
             
            size 194563400
         | 
    	
        chat_template.jinja
    CHANGED
    
    | @@ -1,6 +1,11 @@ | |
| 1 | 
             
            {{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ '<|start_header_id|>system<|end_header_id|>
         | 
|  | |
| 2 | 
             
            ' + messages[0]['content'] + '<|eot_id|>' }}{% set loop_messages = messages[1:] %}{% else %}{{ '<|start_header_id|>system<|end_header_id|>
         | 
|  | |
| 3 | 
             
            Below are some instructions that describe some tasks. Write responses that appropriately complete each request.<|eot_id|>' }}{% set loop_messages = messages %}{% endif %}{% for message in loop_messages %}{% if message['role'] == 'user' %}{{ '<|start_header_id|>user<|end_header_id|>
         | 
|  | |
| 4 | 
             
            ' + message['content'] + '<|eot_id|>' }}{% elif message['role'] == 'assistant' %}{{ '<|start_header_id|>assistant<|end_header_id|>
         | 
|  | |
| 5 | 
             
            ' + message['content'] + '<|eot_id|>' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>
         | 
|  | |
| 6 | 
             
            ' }}{% endif %}
         | 
|  | |
| 1 | 
             
            {{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ '<|start_header_id|>system<|end_header_id|>
         | 
| 2 | 
            +
             | 
| 3 | 
             
            ' + messages[0]['content'] + '<|eot_id|>' }}{% set loop_messages = messages[1:] %}{% else %}{{ '<|start_header_id|>system<|end_header_id|>
         | 
| 4 | 
            +
             | 
| 5 | 
             
            Below are some instructions that describe some tasks. Write responses that appropriately complete each request.<|eot_id|>' }}{% set loop_messages = messages %}{% endif %}{% for message in loop_messages %}{% if message['role'] == 'user' %}{{ '<|start_header_id|>user<|end_header_id|>
         | 
| 6 | 
            +
             | 
| 7 | 
             
            ' + message['content'] + '<|eot_id|>' }}{% elif message['role'] == 'assistant' %}{{ '<|start_header_id|>assistant<|end_header_id|>
         | 
| 8 | 
            +
             | 
| 9 | 
             
            ' + message['content'] + '<|eot_id|>' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>
         | 
| 10 | 
            +
             | 
| 11 | 
             
            ' }}{% endif %}
         | 
    	
        runs/Aug08_09-43-50_54b39aeabec8/events.out.tfevents.1754646239.54b39aeabec8.16942.0
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:8ce26b7ec4004278672b72d77464263fa6799ca43e2655de3ae4483e4196af86
         | 
| 3 | 
            +
            size 27196
         | 
    	
        tokenizer_config.json
    CHANGED
    
    | @@ -2060,7 +2060,7 @@ | |
| 2060 | 
             
              ],
         | 
| 2061 | 
             
              "model_max_length": 131072,
         | 
| 2062 | 
             
              "pad_token": "<|finetune_right_pad_id|>",
         | 
| 2063 | 
            -
              "padding_side": " | 
| 2064 | 
             
              "tokenizer_class": "PreTrainedTokenizerFast",
         | 
| 2065 | 
             
              "unk_token": null
         | 
| 2066 | 
             
            }
         | 
|  | |
| 2060 | 
             
              ],
         | 
| 2061 | 
             
              "model_max_length": 131072,
         | 
| 2062 | 
             
              "pad_token": "<|finetune_right_pad_id|>",
         | 
| 2063 | 
            +
              "padding_side": "right",
         | 
| 2064 | 
             
              "tokenizer_class": "PreTrainedTokenizerFast",
         | 
| 2065 | 
             
              "unk_token": null
         | 
| 2066 | 
             
            }
         | 
    	
        training_args.bin
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
            -
            size  | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:b7b5ac3eb18e7aa01f0b881a93e3aea65416ae2b625af2c90964540ec6af9c6e
         | 
| 3 | 
            +
            size 6225
         |