Update models
Browse files- .gitattributes +12 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q2_K.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q3_K_L.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q3_K_M.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q3_K_S.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q4_0.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q4_K_M.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q4_K_S.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q5_0.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q5_K_S.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q6_K.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-Q8_0.gguf +3 -0
- Mistral-Small-3.1-24B-Instruct-2503-f16.gguf +3 -0
- config.json +46 -0
.gitattributes
CHANGED
@@ -34,3 +34,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
Mistral-Small-3.1-24B-Instruct-2503-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
Mistral-Small-3.1-24B-Instruct-2503-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
45 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
+
Mistral-Small-3.1-24B-Instruct-2503-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
48 |
+
Mistral-Small-3.1-24B-Instruct-2503-f16.gguf filter=lfs diff=lfs merge=lfs -text
|
Mistral-Small-3.1-24B-Instruct-2503-Q2_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4567f26678ea6766779158275b43d9d75dc2f0c0ab768b82bb1e846a1a129446
|
3 |
+
size 8890326304
|
Mistral-Small-3.1-24B-Instruct-2503-Q3_K_L.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2db99d72fb9c3883cbd9b6add572d84e89fd09c66a0ae6c42006c88308fb8407
|
3 |
+
size 12400762144
|
Mistral-Small-3.1-24B-Instruct-2503-Q3_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:644bb5b0b2df89910b99e84e0d9f4e0510f4e7ee8e8cecb000b439b17767098f
|
3 |
+
size 11474083104
|
Mistral-Small-3.1-24B-Instruct-2503-Q3_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5a612758d9de996d98e01c2577c93545340ca7bdad68619355d6edef83181957
|
3 |
+
size 10400275744
|
Mistral-Small-3.1-24B-Instruct-2503-Q4_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a847ea05e66d2e46bbf02542fa086d7f7d056ad3cebca66ffdcad64dcad92ca7
|
3 |
+
size 13441801504
|
Mistral-Small-3.1-24B-Instruct-2503-Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b47a312714cc10b3d2790c863b707b6a900ccfe6bd0a3a6112b3090a43ab0f65
|
3 |
+
size 14333910304
|
Mistral-Small-3.1-24B-Instruct-2503-Q4_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:515bf6d66d540552c2701a90f4b0000ae06b9562fa013f6f7a987729774bd861
|
3 |
+
size 13549280544
|
Mistral-Small-3.1-24B-Instruct-2503-Q5_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d555ea054661290d53a380274f72b2a9315895b9a0276a0270a1f0f6769a879
|
3 |
+
size 16304413984
|
Mistral-Small-3.1-24B-Instruct-2503-Q5_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28ddf94c4bdc7d239877611f18e8207feb186f80741ef24f2ec7fa8bfb476f36
|
3 |
+
size 16304413984
|
Mistral-Small-3.1-24B-Instruct-2503-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:377f455025b9d8bb535c66621471334e9fd6e187049a93ccdc44b6754b56171e
|
3 |
+
size 19345939744
|
Mistral-Small-3.1-24B-Instruct-2503-Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9b1019a6ba3f81472f915f08c1ed90053715a14f21f7ff4fd77306c5becfe465
|
3 |
+
size 25054780704
|
Mistral-Small-3.1-24B-Instruct-2503-f16.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c08ee83640e5239b85be9a9e993cf7c32f0945051e84e957ef2b54079f7dce39
|
3 |
+
size 47153519904
|
config.json
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"Mistral3ForConditionalGeneration"
|
4 |
+
],
|
5 |
+
"image_token_index": 10,
|
6 |
+
"model_type": "mistral3",
|
7 |
+
"multimodal_projector_bias": false,
|
8 |
+
"projector_hidden_act": "gelu",
|
9 |
+
"spatial_merge_size": 2,
|
10 |
+
"text_config": {
|
11 |
+
"attention_dropout": 0.0,
|
12 |
+
"head_dim": 128,
|
13 |
+
"hidden_act": "silu",
|
14 |
+
"hidden_size": 5120,
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"intermediate_size": 32768,
|
17 |
+
"max_position_embeddings": 131072,
|
18 |
+
"model_type": "mistral",
|
19 |
+
"num_attention_heads": 32,
|
20 |
+
"num_hidden_layers": 40,
|
21 |
+
"num_key_value_heads": 8,
|
22 |
+
"rms_norm_eps": 1e-05,
|
23 |
+
"rope_theta": 1000000000.0,
|
24 |
+
"sliding_window": null,
|
25 |
+
"use_cache": true,
|
26 |
+
"vocab_size": 131072
|
27 |
+
},
|
28 |
+
"torch_dtype": "bfloat16",
|
29 |
+
"transformers_version": "4.50.0.dev0",
|
30 |
+
"vision_config": {
|
31 |
+
"attention_dropout": 0.0,
|
32 |
+
"head_dim": 64,
|
33 |
+
"hidden_act": "gelu",
|
34 |
+
"hidden_size": 1024,
|
35 |
+
"image_size": 1540,
|
36 |
+
"initializer_range": 0.02,
|
37 |
+
"intermediate_size": 4096,
|
38 |
+
"model_type": "pixtral",
|
39 |
+
"num_attention_heads": 16,
|
40 |
+
"num_channels": 3,
|
41 |
+
"num_hidden_layers": 24,
|
42 |
+
"patch_size": 14,
|
43 |
+
"rope_theta": 10000.0
|
44 |
+
},
|
45 |
+
"vision_feature_layer": -1
|
46 |
+
}
|