Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/README.md +10 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/adapter_config.json +23 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/adapter_model.bin +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/special_tokens_map.json +24 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer.json +0 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer.model +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer_config.json +39 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/README.md +10 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/adapter_config.json +23 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/adapter_model.bin +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/optimizer.pt +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_0.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_1.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_2.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_3.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/scheduler.pt +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/special_tokens_map.json +24 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer.json +0 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer.model +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer_config.json +39 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/trainer_state.json +0 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/training_args.bin +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/README.md +12 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/adapter_config.json +23 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/adapter_model.bin +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/optimizer.pt +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_0.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_1.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_2.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_3.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/scheduler.pt +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/special_tokens_map.json +24 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer.json +0 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer.model +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer_config.json +39 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/trainer_state.json +0 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/training_args.bin +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/README.md +10 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/adapter_config.json +23 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/adapter_model.bin +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/optimizer.pt +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_0.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_1.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_2.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_3.pth +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/scheduler.pt +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/special_tokens_map.json +24 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer.json +0 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer.model +3 -0
- modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer_config.json +39 -0
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/README.md
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: peft
|
3 |
+
---
|
4 |
+
## Training procedure
|
5 |
+
|
6 |
+
### Framework versions
|
7 |
+
|
8 |
+
- PEFT 0.5.0
|
9 |
+
|
10 |
+
- PEFT 0.5.0
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/adapter_config.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"auto_mapping": null,
|
3 |
+
"base_model_name_or_path": "/shared/nas/data/m1/xuejunz2/model_weights/Llama-2-7b-hf",
|
4 |
+
"bias": "none",
|
5 |
+
"fan_in_fan_out": false,
|
6 |
+
"inference_mode": true,
|
7 |
+
"init_lora_weights": true,
|
8 |
+
"layers_pattern": null,
|
9 |
+
"layers_to_transform": null,
|
10 |
+
"lora_alpha": 1,
|
11 |
+
"lora_dropout": 0.0,
|
12 |
+
"modules_to_save": null,
|
13 |
+
"peft_type": "LORA",
|
14 |
+
"r": 64,
|
15 |
+
"revision": null,
|
16 |
+
"target_modules": [
|
17 |
+
"q_proj",
|
18 |
+
"k_proj",
|
19 |
+
"v_proj",
|
20 |
+
"o_proj"
|
21 |
+
],
|
22 |
+
"task_type": "CAUSAL_LM"
|
23 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:efbbbb7ea75ce5b9aaf3b587a4588a7ac6f69b6dddfa062eaa3c05dbef52abb7
|
3 |
+
size 268528394
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "</s>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<unk>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
|
3 |
+
size 499723
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer_config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "<unk>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<s>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "</s>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"bos_token": "<s>",
|
29 |
+
"clean_up_tokenization_spaces": false,
|
30 |
+
"eos_token": "</s>",
|
31 |
+
"legacy": false,
|
32 |
+
"model_max_length": 1000000000000000019884624838656,
|
33 |
+
"pad_token": "</s>",
|
34 |
+
"padding_side": "right",
|
35 |
+
"sp_model_kwargs": {},
|
36 |
+
"tokenizer_class": "LlamaTokenizer",
|
37 |
+
"unk_token": "<unk>",
|
38 |
+
"use_default_system_prompt": true
|
39 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/README.md
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: peft
|
3 |
+
---
|
4 |
+
## Training procedure
|
5 |
+
|
6 |
+
### Framework versions
|
7 |
+
|
8 |
+
- PEFT 0.5.0
|
9 |
+
|
10 |
+
- PEFT 0.5.0
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/adapter_config.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"auto_mapping": null,
|
3 |
+
"base_model_name_or_path": "/shared/nas/data/m1/xuejunz2/model_weights/Llama-2-7b-hf",
|
4 |
+
"bias": "none",
|
5 |
+
"fan_in_fan_out": false,
|
6 |
+
"inference_mode": true,
|
7 |
+
"init_lora_weights": true,
|
8 |
+
"layers_pattern": null,
|
9 |
+
"layers_to_transform": null,
|
10 |
+
"lora_alpha": 1,
|
11 |
+
"lora_dropout": 0.0,
|
12 |
+
"modules_to_save": null,
|
13 |
+
"peft_type": "LORA",
|
14 |
+
"r": 64,
|
15 |
+
"revision": null,
|
16 |
+
"target_modules": [
|
17 |
+
"q_proj",
|
18 |
+
"k_proj",
|
19 |
+
"v_proj",
|
20 |
+
"o_proj"
|
21 |
+
],
|
22 |
+
"task_type": "CAUSAL_LM"
|
23 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:efbbbb7ea75ce5b9aaf3b587a4588a7ac6f69b6dddfa062eaa3c05dbef52abb7
|
3 |
+
size 268528394
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fb1da41ea15c3fc705c67353801979096dbcd946d2bb8a1c3bbb9f300505fda8
|
3 |
+
size 537086714
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e812528c12db18ea461d9e3bd3cc22ea90b0101393312170b1e2fc74c940bd81
|
3 |
+
size 14960
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37194a6d48612e1a46a2d5d317ead97c70d9fc4569b0118fcd5f84c3dc9daa5a
|
3 |
+
size 15024
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:17c179483659a784aa1ace2427daff48c556a6bcc3c330e6f3274e4dc95e4b49
|
3 |
+
size 15024
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b56857c9b117629f35af2c3d64f522d33a9d8aa94faa81ec6956380a895118c4
|
3 |
+
size 15024
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3db0714409ba3b0176819041101c2b4b3a70bd1514b4d0f04431adbcb070ab2
|
3 |
+
size 1064
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "</s>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<unk>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
|
3 |
+
size 499723
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer_config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "<unk>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<s>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "</s>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"bos_token": "<s>",
|
29 |
+
"clean_up_tokenization_spaces": false,
|
30 |
+
"eos_token": "</s>",
|
31 |
+
"legacy": false,
|
32 |
+
"model_max_length": 1000000000000000019884624838656,
|
33 |
+
"pad_token": "</s>",
|
34 |
+
"padding_side": "right",
|
35 |
+
"sp_model_kwargs": {},
|
36 |
+
"tokenizer_class": "LlamaTokenizer",
|
37 |
+
"unk_token": "<unk>",
|
38 |
+
"use_default_system_prompt": true
|
39 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae8acf45b101a66520f11d811c9b4c5c8fb3d83195649a5dae75ae3803a963a9
|
3 |
+
size 4600
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/README.md
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: peft
|
3 |
+
---
|
4 |
+
## Training procedure
|
5 |
+
|
6 |
+
### Framework versions
|
7 |
+
|
8 |
+
- PEFT 0.5.0
|
9 |
+
- PEFT 0.5.0
|
10 |
+
- PEFT 0.5.0
|
11 |
+
|
12 |
+
- PEFT 0.5.0
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/adapter_config.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"auto_mapping": null,
|
3 |
+
"base_model_name_or_path": "/shared/nas/data/m1/xuejunz2/model_weights/Llama-2-7b-hf",
|
4 |
+
"bias": "none",
|
5 |
+
"fan_in_fan_out": false,
|
6 |
+
"inference_mode": true,
|
7 |
+
"init_lora_weights": true,
|
8 |
+
"layers_pattern": null,
|
9 |
+
"layers_to_transform": null,
|
10 |
+
"lora_alpha": 1,
|
11 |
+
"lora_dropout": 0.0,
|
12 |
+
"modules_to_save": null,
|
13 |
+
"peft_type": "LORA",
|
14 |
+
"r": 64,
|
15 |
+
"revision": null,
|
16 |
+
"target_modules": [
|
17 |
+
"q_proj",
|
18 |
+
"k_proj",
|
19 |
+
"v_proj",
|
20 |
+
"o_proj"
|
21 |
+
],
|
22 |
+
"task_type": "CAUSAL_LM"
|
23 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2d7e162b663d08768968547f75192156e0a95e127c23f3940afab27fe75ccfa0
|
3 |
+
size 268528394
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a662864d65425ab1679cd1d884e22d0a597eb512537a6b516ae11e651e2268b3
|
3 |
+
size 537086714
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b0f69b5414e85c17b526e8c65193f06ef8d6195a8c8383efe1d2f9d044824583
|
3 |
+
size 14960
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:49d1438e98cc9c53a6852464635ce62e9788e61eb3646b73e33813f487c4b6ae
|
3 |
+
size 15024
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4388add9cec90932f8ff0100d27a0574d98e1bad52ff89d44e31967d2b4fbfde
|
3 |
+
size 15024
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a705d6dfaae4f2c1b4b2be6b25a6eb521ffae6fcba21cc1531e97b60037ed079
|
3 |
+
size 15024
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c57b62113ea6696a3b704766d3edd1245c2a899e4f0575dbf1544286d733557
|
3 |
+
size 1064
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "</s>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<unk>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
|
3 |
+
size 499723
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer_config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "<unk>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<s>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "</s>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"bos_token": "<s>",
|
29 |
+
"clean_up_tokenization_spaces": false,
|
30 |
+
"eos_token": "</s>",
|
31 |
+
"legacy": false,
|
32 |
+
"model_max_length": 1000000000000000019884624838656,
|
33 |
+
"pad_token": "</s>",
|
34 |
+
"padding_side": "right",
|
35 |
+
"sp_model_kwargs": {},
|
36 |
+
"tokenizer_class": "LlamaTokenizer",
|
37 |
+
"unk_token": "<unk>",
|
38 |
+
"use_default_system_prompt": true
|
39 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae8acf45b101a66520f11d811c9b4c5c8fb3d83195649a5dae75ae3803a963a9
|
3 |
+
size 4600
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/README.md
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: peft
|
3 |
+
---
|
4 |
+
## Training procedure
|
5 |
+
|
6 |
+
### Framework versions
|
7 |
+
|
8 |
+
- PEFT 0.5.0
|
9 |
+
|
10 |
+
- PEFT 0.5.0
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/adapter_config.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"auto_mapping": null,
|
3 |
+
"base_model_name_or_path": "/shared/nas/data/m1/xuejunz2/model_weights/Llama-2-7b-hf",
|
4 |
+
"bias": "none",
|
5 |
+
"fan_in_fan_out": false,
|
6 |
+
"inference_mode": true,
|
7 |
+
"init_lora_weights": true,
|
8 |
+
"layers_pattern": null,
|
9 |
+
"layers_to_transform": null,
|
10 |
+
"lora_alpha": 1,
|
11 |
+
"lora_dropout": 0.0,
|
12 |
+
"modules_to_save": null,
|
13 |
+
"peft_type": "LORA",
|
14 |
+
"r": 64,
|
15 |
+
"revision": null,
|
16 |
+
"target_modules": [
|
17 |
+
"q_proj",
|
18 |
+
"k_proj",
|
19 |
+
"v_proj",
|
20 |
+
"o_proj"
|
21 |
+
],
|
22 |
+
"task_type": "CAUSAL_LM"
|
23 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:602e6dc995a5cd30d73313ae99d682334b0e1be28dfa910e2652688004b500e9
|
3 |
+
size 268528394
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:54d41c858553cea0ff8abc84acacd331c3b767369779a0370d31ed8f11e0f437
|
3 |
+
size 537086714
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:331df8a1fce70deea17ec1af86cf0dc22ecc098b0c4860cf6b22d89e76c71a14
|
3 |
+
size 14960
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6e07ace389d24bc1307b74f42a1e7b8f0117b0db853e2df64ff3f15cb92916a2
|
3 |
+
size 15024
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:da6a990f346d7014dffb28fa2bc7d3b890bd3c53712503fce3656da48d3d6e50
|
3 |
+
size 15024
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e95f356ca38179b05993f55daece0223e96fa10b9a1b9ea2102a739211333f63
|
3 |
+
size 15024
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f6a64bfb00f5fc6b3e5400798323479c248f49e6235678500b3ac4eb6986d3f
|
3 |
+
size 1064
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "</s>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<unk>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
|
3 |
+
size 499723
|
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer_config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "<unk>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<s>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "</s>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"bos_token": "<s>",
|
29 |
+
"clean_up_tokenization_spaces": false,
|
30 |
+
"eos_token": "</s>",
|
31 |
+
"legacy": false,
|
32 |
+
"model_max_length": 1000000000000000019884624838656,
|
33 |
+
"pad_token": "</s>",
|
34 |
+
"padding_side": "right",
|
35 |
+
"sp_model_kwargs": {},
|
36 |
+
"tokenizer_class": "LlamaTokenizer",
|
37 |
+
"unk_token": "<unk>",
|
38 |
+
"use_default_system_prompt": true
|
39 |
+
}
|