praeclarumjj3 commited on
Commit
fa7c997
1 Parent(s): 2580e32

Upload folder using huggingface_hub

Browse files
.ipynb_checkpoints/config-checkpoint.json CHANGED
@@ -45,7 +45,7 @@
45
  "tune_mm_mlp_adapter": false,
46
  "tune_mm_vision_resampler": false,
47
  "unfreeze_mm_vision_tower": false,
48
- "use_cache": false,
49
  "use_mm2_proj": true,
50
  "use_mm_proj": true,
51
  "vocab_size": 32000
 
45
  "tune_mm_mlp_adapter": false,
46
  "tune_mm_vision_resampler": false,
47
  "unfreeze_mm_vision_tower": false,
48
+ "use_cache": true,
49
  "use_mm2_proj": true,
50
  "use_mm_proj": true,
51
  "vocab_size": 32000
.ipynb_checkpoints/generation_config-checkpoint.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "max_length": 4096,
6
+ "pad_token_id": 0,
7
+ "transformers_version": "4.31.0"
8
+ }
config.json CHANGED
@@ -45,7 +45,7 @@
45
  "tune_mm_mlp_adapter": false,
46
  "tune_mm_vision_resampler": false,
47
  "unfreeze_mm_vision_tower": false,
48
- "use_cache": false,
49
  "use_mm2_proj": true,
50
  "use_mm_proj": true,
51
  "vocab_size": 32000
 
45
  "tune_mm_mlp_adapter": false,
46
  "tune_mm_vision_resampler": false,
47
  "unfreeze_mm_vision_tower": false,
48
+ "use_cache": true,
49
  "use_mm2_proj": true,
50
  "use_mm_proj": true,
51
  "vocab_size": 32000