cipherpy commited on
Commit
f3a402e
·
verified ·
1 Parent(s): 33d58c8

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -26,17 +26,17 @@
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
- "v_proj",
30
- "q_proj",
31
- "lm_head",
32
- "fc1",
33
- "o_proj",
34
  "fc2",
35
- "k_proj",
36
  "out_proj",
37
- "up_proj",
38
  "gate_proj",
39
- "down_proj"
 
 
 
 
 
 
40
  ],
41
  "task_type": "CAUSAL_LM",
42
  "use_dora": false,
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
 
 
 
 
 
29
  "fc2",
 
30
  "out_proj",
31
+ "v_proj",
32
  "gate_proj",
33
+ "up_proj",
34
+ "fc1",
35
+ "k_proj",
36
+ "o_proj",
37
+ "down_proj",
38
+ "lm_head",
39
+ "q_proj"
40
  ],
41
  "task_type": "CAUSAL_LM",
42
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2053c303ffc8a167adda423e7487364c00a4bb7b979f0c552ca2d69a0df28c87
3
  size 2839124552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b206d65f700e6507904f0e722b50ca644ce1ca45868f4d58628e8a59121577e
3
  size 2839124552
runs/Apr01_09-21-29_gpu-dom-cmlre/events.out.tfevents.1743479499.gpu-dom-cmlre ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf0c9ee8a61b7ba3c2f4cad194d654a4cb6868ac62830308afe8c88d9a2ce2ed
3
+ size 97327
runs/Apr01_10-28-10_gpu-dom-cmlre/events.out.tfevents.1743483496.gpu-dom-cmlre ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f4e941ffc49c663d00b54d6c2cc1bfe3b9c9ae286e55354dc8ca47f061a00a5
3
+ size 43618
runs/Mar23_17-15-34_gpu-dom-cmlre/events.out.tfevents.1742730336.gpu-dom-cmlre CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe6d7cb68544c10cfb4afcc5396de1cebda29fa4285f6d0042742dadc0ed34ca
3
- size 3611953
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e750157ac2704c690fdb8980e8450fc4bf6a8ee08b489e6dc07fa39468c810ac
3
+ size 3830853
tokenizer_config.json CHANGED
@@ -51335,8 +51335,12 @@
51335
  "image_token": "<image_soft_token>"
51336
  },
51337
  "image_token": "<image_soft_token>",
 
51338
  "model_max_length": 1000000000000000019884624838656,
 
51339
  "pad_token": "<pad>",
 
 
51340
  "processor_class": "Gemma3Processor",
51341
  "sp_model_kwargs": null,
51342
  "spaces_between_special_tokens": false,
 
51335
  "image_token": "<image_soft_token>"
51336
  },
51337
  "image_token": "<image_soft_token>",
51338
+ "max_length": null,
51339
  "model_max_length": 1000000000000000019884624838656,
51340
+ "pad_to_multiple_of": null,
51341
  "pad_token": "<pad>",
51342
+ "pad_token_type_id": 0,
51343
+ "padding_side": "left",
51344
  "processor_class": "Gemma3Processor",
51345
  "sp_model_kwargs": null,
51346
  "spaces_between_special_tokens": false,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e190478124f4869598fce383b4dc80723b7eb1e11b0acc89b86b58f7c20c7520
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cead071e7dc4914a2fab5ad07bf61bdf81c2b5683141c087ac6ac23f8e29f1b
3
  size 5624