yifan-Eva commited on
Commit
51b8a6f
·
verified ·
1 Parent(s): 8b73c7e

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/README.md +10 -0
  2. modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/adapter_config.json +23 -0
  3. modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/adapter_model.bin +3 -0
  4. modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/special_tokens_map.json +24 -0
  5. modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer.json +0 -0
  6. modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer.model +3 -0
  7. modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer_config.json +39 -0
  8. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/README.md +10 -0
  9. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/adapter_config.json +23 -0
  10. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/adapter_model.bin +3 -0
  11. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/optimizer.pt +3 -0
  12. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_0.pth +3 -0
  13. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_1.pth +3 -0
  14. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_2.pth +3 -0
  15. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_3.pth +3 -0
  16. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/scheduler.pt +3 -0
  17. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/special_tokens_map.json +24 -0
  18. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer.json +0 -0
  19. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer.model +3 -0
  20. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer_config.json +39 -0
  21. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/trainer_state.json +0 -0
  22. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/training_args.bin +3 -0
  23. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/README.md +12 -0
  24. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/adapter_config.json +23 -0
  25. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/adapter_model.bin +3 -0
  26. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/optimizer.pt +3 -0
  27. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_0.pth +3 -0
  28. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_1.pth +3 -0
  29. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_2.pth +3 -0
  30. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_3.pth +3 -0
  31. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/scheduler.pt +3 -0
  32. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/special_tokens_map.json +24 -0
  33. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer.json +0 -0
  34. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer.model +3 -0
  35. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer_config.json +39 -0
  36. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/trainer_state.json +0 -0
  37. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/training_args.bin +3 -0
  38. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/README.md +10 -0
  39. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/adapter_config.json +23 -0
  40. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/adapter_model.bin +3 -0
  41. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/optimizer.pt +3 -0
  42. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_0.pth +3 -0
  43. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_1.pth +3 -0
  44. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_2.pth +3 -0
  45. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_3.pth +3 -0
  46. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/scheduler.pt +3 -0
  47. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/special_tokens_map.json +24 -0
  48. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer.json +0 -0
  49. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer.model +3 -0
  50. modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer_config.json +39 -0
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/README.md ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+ - PEFT 0.5.0
9
+
10
+ - PEFT 0.5.0
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/adapter_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "/shared/nas/data/m1/xuejunz2/model_weights/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 1,
11
+ "lora_dropout": 0.0,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 64,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "q_proj",
18
+ "k_proj",
19
+ "v_proj",
20
+ "o_proj"
21
+ ],
22
+ "task_type": "CAUSAL_LM"
23
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efbbbb7ea75ce5b9aaf3b587a4588a7ac6f69b6dddfa062eaa3c05dbef52abb7
3
+ size 268528394
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
modpo/lm/(0.1)helpful+(1-0.1)harmless/best_checkpoint/tokenizer_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "bos_token": "<s>",
29
+ "clean_up_tokenization_spaces": false,
30
+ "eos_token": "</s>",
31
+ "legacy": false,
32
+ "model_max_length": 1000000000000000019884624838656,
33
+ "pad_token": "</s>",
34
+ "padding_side": "right",
35
+ "sp_model_kwargs": {},
36
+ "tokenizer_class": "LlamaTokenizer",
37
+ "unk_token": "<unk>",
38
+ "use_default_system_prompt": true
39
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/README.md ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+ - PEFT 0.5.0
9
+
10
+ - PEFT 0.5.0
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/adapter_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "/shared/nas/data/m1/xuejunz2/model_weights/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 1,
11
+ "lora_dropout": 0.0,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 64,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "q_proj",
18
+ "k_proj",
19
+ "v_proj",
20
+ "o_proj"
21
+ ],
22
+ "task_type": "CAUSAL_LM"
23
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efbbbb7ea75ce5b9aaf3b587a4588a7ac6f69b6dddfa062eaa3c05dbef52abb7
3
+ size 268528394
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb1da41ea15c3fc705c67353801979096dbcd946d2bb8a1c3bbb9f300505fda8
3
+ size 537086714
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e812528c12db18ea461d9e3bd3cc22ea90b0101393312170b1e2fc74c940bd81
3
+ size 14960
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37194a6d48612e1a46a2d5d317ead97c70d9fc4569b0118fcd5f84c3dc9daa5a
3
+ size 15024
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17c179483659a784aa1ace2427daff48c556a6bcc3c330e6f3274e4dc95e4b49
3
+ size 15024
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b56857c9b117629f35af2c3d64f522d33a9d8aa94faa81ec6956380a895118c4
3
+ size 15024
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3db0714409ba3b0176819041101c2b4b3a70bd1514b4d0f04431adbcb070ab2
3
+ size 1064
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/tokenizer_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "bos_token": "<s>",
29
+ "clean_up_tokenization_spaces": false,
30
+ "eos_token": "</s>",
31
+ "legacy": false,
32
+ "model_max_length": 1000000000000000019884624838656,
33
+ "pad_token": "</s>",
34
+ "padding_side": "right",
35
+ "sp_model_kwargs": {},
36
+ "tokenizer_class": "LlamaTokenizer",
37
+ "unk_token": "<unk>",
38
+ "use_default_system_prompt": true
39
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-12000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae8acf45b101a66520f11d811c9b4c5c8fb3d83195649a5dae75ae3803a963a9
3
+ size 4600
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/README.md ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+ - PEFT 0.5.0
9
+ - PEFT 0.5.0
10
+ - PEFT 0.5.0
11
+
12
+ - PEFT 0.5.0
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/adapter_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "/shared/nas/data/m1/xuejunz2/model_weights/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 1,
11
+ "lora_dropout": 0.0,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 64,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "q_proj",
18
+ "k_proj",
19
+ "v_proj",
20
+ "o_proj"
21
+ ],
22
+ "task_type": "CAUSAL_LM"
23
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d7e162b663d08768968547f75192156e0a95e127c23f3940afab27fe75ccfa0
3
+ size 268528394
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a662864d65425ab1679cd1d884e22d0a597eb512537a6b516ae11e651e2268b3
3
+ size 537086714
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0f69b5414e85c17b526e8c65193f06ef8d6195a8c8383efe1d2f9d044824583
3
+ size 14960
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49d1438e98cc9c53a6852464635ce62e9788e61eb3646b73e33813f487c4b6ae
3
+ size 15024
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4388add9cec90932f8ff0100d27a0574d98e1bad52ff89d44e31967d2b4fbfde
3
+ size 15024
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a705d6dfaae4f2c1b4b2be6b25a6eb521ffae6fcba21cc1531e97b60037ed079
3
+ size 15024
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c57b62113ea6696a3b704766d3edd1245c2a899e4f0575dbf1544286d733557
3
+ size 1064
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/tokenizer_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "bos_token": "<s>",
29
+ "clean_up_tokenization_spaces": false,
30
+ "eos_token": "</s>",
31
+ "legacy": false,
32
+ "model_max_length": 1000000000000000019884624838656,
33
+ "pad_token": "</s>",
34
+ "padding_side": "right",
35
+ "sp_model_kwargs": {},
36
+ "tokenizer_class": "LlamaTokenizer",
37
+ "unk_token": "<unk>",
38
+ "use_default_system_prompt": true
39
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-6000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae8acf45b101a66520f11d811c9b4c5c8fb3d83195649a5dae75ae3803a963a9
3
+ size 4600
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/README.md ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+ - PEFT 0.5.0
9
+
10
+ - PEFT 0.5.0
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/adapter_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "/shared/nas/data/m1/xuejunz2/model_weights/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 1,
11
+ "lora_dropout": 0.0,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 64,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "q_proj",
18
+ "k_proj",
19
+ "v_proj",
20
+ "o_proj"
21
+ ],
22
+ "task_type": "CAUSAL_LM"
23
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:602e6dc995a5cd30d73313ae99d682334b0e1be28dfa910e2652688004b500e9
3
+ size 268528394
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54d41c858553cea0ff8abc84acacd331c3b767369779a0370d31ed8f11e0f437
3
+ size 537086714
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:331df8a1fce70deea17ec1af86cf0dc22ecc098b0c4860cf6b22d89e76c71a14
3
+ size 14960
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e07ace389d24bc1307b74f42a1e7b8f0117b0db853e2df64ff3f15cb92916a2
3
+ size 15024
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da6a990f346d7014dffb28fa2bc7d3b890bd3c53712503fce3656da48d3d6e50
3
+ size 15024
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e95f356ca38179b05993f55daece0223e96fa10b9a1b9ea2102a739211333f63
3
+ size 15024
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f6a64bfb00f5fc6b3e5400798323479c248f49e6235678500b3ac4eb6986d3f
3
+ size 1064
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
modpo/lm/(0.1)helpful+(1-0.1)harmless/checkpoint-9000/tokenizer_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "bos_token": "<s>",
29
+ "clean_up_tokenization_spaces": false,
30
+ "eos_token": "</s>",
31
+ "legacy": false,
32
+ "model_max_length": 1000000000000000019884624838656,
33
+ "pad_token": "</s>",
34
+ "padding_side": "right",
35
+ "sp_model_kwargs": {},
36
+ "tokenizer_class": "LlamaTokenizer",
37
+ "unk_token": "<unk>",
38
+ "use_default_system_prompt": true
39
+ }