panoyo9829 commited on
Commit
3d8a16c
·
verified ·
1 Parent(s): 400a660

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
3
  "architectures": [
4
- "LlavaLlamaForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
@@ -49,10 +49,23 @@
49
  "mm_vision_select_layer": -2,
50
  "mm_vision_tower": "openai/clip-vit-large-patch14-336",
51
  "mm_vision_tower_lr": 2e-06,
52
- "model_type": "llava",
53
  "num_attention_heads": 32,
54
  "num_hidden_layers": 32,
55
  "num_key_value_heads": 8,
 
 
 
 
 
 
 
 
 
 
 
 
 
56
  "rms_norm_eps": 1e-05,
57
  "rope_theta": 1000000.0,
58
  "sliding_window": null,
@@ -60,22 +73,11 @@
60
  "tokenizer_model_max_length": 4096,
61
  "tokenizer_padding_side": "left",
62
  "torch_dtype": "bfloat16",
63
- "transformers_version": "4.36.2",
64
  "tune_mm_mlp_adapter": false,
65
  "tune_mm_vision_resampler": false,
66
  "unfreeze_mm_vision_tower": true,
67
  "use_cache": true,
68
  "use_mm_proj": true,
69
- "vocab_size": 32000,
70
- "quantization_config": {
71
- "bits": 4,
72
- "group_size": 128,
73
- "damp_percent": 0.1,
74
- "desc_act": true,
75
- "sym": true,
76
- "true_sequential": true,
77
- "model_name_or_path": null,
78
- "model_file_base_name": "model",
79
- "quant_method": "gptq"
80
- }
81
  }
 
1
  {
2
+ "_name_or_path": "../../submission/models/llava-v1.6-mistral-7b",
3
  "architectures": [
4
+ "LlavaMistralForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
 
49
  "mm_vision_select_layer": -2,
50
  "mm_vision_tower": "openai/clip-vit-large-patch14-336",
51
  "mm_vision_tower_lr": 2e-06,
52
+ "model_type": "llava_mistral",
53
  "num_attention_heads": 32,
54
  "num_hidden_layers": 32,
55
  "num_key_value_heads": 8,
56
+ "quantization_config": {
57
+ "bits": 4,
58
+ "damp_percent": 0.01,
59
+ "desc_act": false,
60
+ "group_size": 128,
61
+ "is_marlin_format": false,
62
+ "model_file_base_name": null,
63
+ "model_name_or_path": null,
64
+ "quant_method": "gptq",
65
+ "static_groups": false,
66
+ "sym": true,
67
+ "true_sequential": true
68
+ },
69
  "rms_norm_eps": 1e-05,
70
  "rope_theta": 1000000.0,
71
  "sliding_window": null,
 
73
  "tokenizer_model_max_length": 4096,
74
  "tokenizer_padding_side": "left",
75
  "torch_dtype": "bfloat16",
76
+ "transformers_version": "4.38.1",
77
  "tune_mm_mlp_adapter": false,
78
  "tune_mm_vision_resampler": false,
79
  "unfreeze_mm_vision_tower": true,
80
  "use_cache": true,
81
  "use_mm_proj": true,
82
+ "vocab_size": 32000
 
 
 
 
 
 
 
 
 
 
 
83
  }
gptq_model-4bit-128g.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:116d3c4ba4388fbdd027f13f83d060d39cc6e008930caf9c3037d5a5071c3d14
3
+ size 5981509688
quantize_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "damp_percent": 0.01,
5
+ "desc_act": false,
6
+ "static_groups": false,
7
+ "sym": true,
8
+ "true_sequential": true,
9
+ "model_name_or_path": null,
10
+ "model_file_base_name": null,
11
+ "is_marlin_format": false,
12
+ "quant_method": "gptq"
13
+ }