danielhanchen commited on
Commit
48c2d46
·
verified ·
1 Parent(s): e34d02f

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -80,3 +80,5 @@ Q4_0/Llama-4-Scout-17B-16E-Instruct-Q4_0-00002-of-00002.gguf filter=lfs diff=lfs
80
  Q4_1/Llama-4-Scout-17B-16E-Instruct-Q4_1-00002-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
81
  Q4_1/Llama-4-Scout-17B-16E-Instruct-Q4_1-00001-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
82
  Q4_0/Llama-4-Scout-17B-16E-Instruct-Q4_0-00001-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
80
  Q4_1/Llama-4-Scout-17B-16E-Instruct-Q4_1-00002-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
81
  Q4_1/Llama-4-Scout-17B-16E-Instruct-Q4_1-00001-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
82
  Q4_0/Llama-4-Scout-17B-16E-Instruct-Q4_0-00001-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
83
+ UD-Q5_K_XL/Llama-4-Scout-17B-16E-Instruct-UD-Q5_K_XL-00002-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
84
+ UD-Q5_K_XL/Llama-4-Scout-17B-16E-Instruct-UD-Q5_K_XL-00001-of-00002.gguf filter=lfs diff=lfs merge=lfs -text
Llama-4-Scout-17B-16E-Instruct-UD-Q2_K_XL.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc66c538213fe9b5aa2efad9e1b020dcd4b01edc90e66c0db3fa744239346862
3
- size 42371712160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c1f864d474b00edc0750b164ead6b503f4dedaee3b9b1161de42e1112365dd8
3
+ size 42376955040
Llama-4-Scout-17B-16E-Instruct-UD-Q3_K_XL.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2589da1ba31d4e823681268ba9c14f7156f8ec53785eb6046e2ea2060e686737
3
- size 48919413920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5adb7a1471db60e37f56b20c89e13028e13d6cd9da34d31b89257e47c973affb
3
+ size 49019028640
Q4_K_M/Llama-4-Scout-17B-16E-Instruct-Q4_K_M-00001-of-00002.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e1a3539e1b4c205d7038cb276d92dc01709babc5c2eb8225bce3753ba64bbed7
3
  size 49848377344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4dd09e77cfd0eb76c36ee9ac62c87243b574e141b05559d11c3e2955c16727b
3
  size 49848377344
Q4_K_M/Llama-4-Scout-17B-16E-Instruct-Q4_K_M-00002-of-00002.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:53d9a61b90e38330daa4bb07afe56aa3e74a3d3aa31d344c053ebbdcfe5d59fe
3
  size 15511520608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7330ae14527f08e8a44a6a573b45084052b8822c4b8a5179c5cad9cd6f6f795
3
  size 15511520608
UD-Q4_K_XL/Llama-4-Scout-17B-16E-Instruct-UD-Q4_K_XL-00001-of-00002.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:421d922dfcd2f1890f0433284495a1ff706b589d6847a756f82d0d8c9f334596
3
- size 49736824224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83a9a4eec72f36441a8765d0756d5342ad85bb15ab99f85248cdcd25436e2e69
3
+ size 49647695264
UD-Q4_K_XL/Llama-4-Scout-17B-16E-Instruct-UD-Q4_K_XL-00002-of-00002.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e901a614c67a0399bab4b82402ef1cdbffedb46e8b554c82472650c4e8b27d1d
3
- size 12261281728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cecd4c41f02bc2437c22e733f38e5db2ebe4a43b161513fb90d1b37fbff77e9
3
+ size 12350410688
UD-Q5_K_XL/Llama-4-Scout-17B-16E-Instruct-UD-Q5_K_XL-00001-of-00002.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:889ee7966f5025afe97fed1628b07536bc55839e769b2745ba50bfcb64fa5f6c
3
+ size 49725225792
UD-Q5_K_XL/Llama-4-Scout-17B-16E-Instruct-UD-Q5_K_XL-00002-of-00002.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ae8384d9e5a24daad6d482d32c18edab355ef4593fb4e0bff4a22cd751d4bc5
3
+ size 29421275680
config.json CHANGED
@@ -6,17 +6,22 @@
6
  "eoi_token_index": 200081,
7
  "image_token_index": 200092,
8
  "model_type": "llama4",
 
9
  "text_config": {
10
  "_attn_implementation_autoset": true,
11
  "attention_bias": false,
12
  "attention_chunk_size": 8192,
13
  "attention_dropout": 0.0,
 
 
14
  "bos_token_id": 200000,
 
15
  "eos_token_id": [
16
  200001,
17
  200007,
18
  200008
19
  ],
 
20
  "for_llm_compressor": false,
21
  "head_dim": 128,
22
  "hidden_act": "silu",
@@ -27,7 +32,106 @@
27
  "intermediate_size_mlp": 16384,
28
  "max_position_embeddings": 10485760,
29
  "model_type": "llama4_text",
30
- "no_rope_layers": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31
  "num_attention_heads": 40,
32
  "num_experts_per_tok": 1,
33
  "num_hidden_layers": 48,
@@ -51,8 +155,10 @@
51
  "use_qk_norm": true,
52
  "vocab_size": 202048
53
  },
 
54
  "torch_dtype": "bfloat16",
55
- "transformers_version": "4.51.0.dev0",
 
56
  "vision_config": {
57
  "_attn_implementation_autoset": true,
58
  "attention_dropout": 0.0,
 
6
  "eoi_token_index": 200081,
7
  "image_token_index": 200092,
8
  "model_type": "llama4",
9
+ "pad_token_id": 200018,
10
  "text_config": {
11
  "_attn_implementation_autoset": true,
12
  "attention_bias": false,
13
  "attention_chunk_size": 8192,
14
  "attention_dropout": 0.0,
15
+ "attn_scale": 0.1,
16
+ "attn_temperature_tuning": 4,
17
  "bos_token_id": 200000,
18
+ "cache_implementation": "hybrid",
19
  "eos_token_id": [
20
  200001,
21
  200007,
22
  200008
23
  ],
24
+ "floor_scale": 8192,
25
  "for_llm_compressor": false,
26
  "head_dim": 128,
27
  "hidden_act": "silu",
 
32
  "intermediate_size_mlp": 16384,
33
  "max_position_embeddings": 10485760,
34
  "model_type": "llama4_text",
35
+ "moe_layers": [
36
+ 0,
37
+ 1,
38
+ 2,
39
+ 3,
40
+ 4,
41
+ 5,
42
+ 6,
43
+ 7,
44
+ 8,
45
+ 9,
46
+ 10,
47
+ 11,
48
+ 12,
49
+ 13,
50
+ 14,
51
+ 15,
52
+ 16,
53
+ 17,
54
+ 18,
55
+ 19,
56
+ 20,
57
+ 21,
58
+ 22,
59
+ 23,
60
+ 24,
61
+ 25,
62
+ 26,
63
+ 27,
64
+ 28,
65
+ 29,
66
+ 30,
67
+ 31,
68
+ 32,
69
+ 33,
70
+ 34,
71
+ 35,
72
+ 36,
73
+ 37,
74
+ 38,
75
+ 39,
76
+ 40,
77
+ 41,
78
+ 42,
79
+ 43,
80
+ 44,
81
+ 45,
82
+ 46,
83
+ 47
84
+ ],
85
+ "no_rope_layers": [
86
+ 1,
87
+ 1,
88
+ 1,
89
+ 0,
90
+ 1,
91
+ 1,
92
+ 1,
93
+ 0,
94
+ 1,
95
+ 1,
96
+ 1,
97
+ 0,
98
+ 1,
99
+ 1,
100
+ 1,
101
+ 0,
102
+ 1,
103
+ 1,
104
+ 1,
105
+ 0,
106
+ 1,
107
+ 1,
108
+ 1,
109
+ 0,
110
+ 1,
111
+ 1,
112
+ 1,
113
+ 0,
114
+ 1,
115
+ 1,
116
+ 1,
117
+ 0,
118
+ 1,
119
+ 1,
120
+ 1,
121
+ 0,
122
+ 1,
123
+ 1,
124
+ 1,
125
+ 0,
126
+ 1,
127
+ 1,
128
+ 1,
129
+ 0,
130
+ 1,
131
+ 1,
132
+ 1,
133
+ 0
134
+ ],
135
  "num_attention_heads": 40,
136
  "num_experts_per_tok": 1,
137
  "num_hidden_layers": 48,
 
155
  "use_qk_norm": true,
156
  "vocab_size": 202048
157
  },
158
+ "tie_word_embeddings": false,
159
  "torch_dtype": "bfloat16",
160
+ "transformers_version": "4.51.3",
161
+ "unsloth_fixed": true,
162
  "vision_config": {
163
  "_attn_implementation_autoset": true,
164
  "attention_dropout": 0.0,