Esha89 commited on
Commit
b5283f1
·
verified ·
1 Parent(s): 891b6a0

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -25,8 +25,8 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
- "q_proj",
29
- "v_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "trainable_token_indices": null,
 
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
+ "v_proj",
29
+ "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f12874d2be69cf3f99b098a8157dbda4ef6aad4cff54a1bfad75de72a6987c1d
3
  size 4517152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d630aea7f7424e985a0862bfd9bcfe085a149fcbc77cd681dbf730741a0dd66
3
  size 4517152
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 512,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 512
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 2,
16
+ "pad_type_id": 0,
17
+ "pad_token": "</s>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0a6e312f9eeec5fac451e54a979932c3926cfcab1071471a5bf9477a9d7d7a3
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e17160c409174eebeb953806735f5288fae127978f6e9531dd11ddf7ab787131
3
  size 5304