Model Card
Dynamic quantization of DeepSeek-R1-0528 for ik_llama fork, optimized to run with 24GB to 32GB VRAM and 512GB RAM systems while providing the best balance between quality and performance for coding.
THIS MODEL ONLY RUNS ON THE IK_LLAMA FORK!!!
See this detailed guide on how to setup an run ik_llama.
Run
Use the following command lines to run the model (tweak the command to further customize it to your needs).
24GB VRAM
./build/bin/llama-server \
--alias anikifoss/DeepSeek-R1-0528-DQ4_K_R4 \
--model /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00001-of-00010.gguf \
--temp 0.5 --top-k 0 --top-p 1.0 --min-p 0.1 --repeat-penalty 1.0 \
--ctx-size 41000 \
-ctk q8_0 \
-mla 2 -fa \
-amb 512 \
-b 1024 -ub 1024 \
-fmoe \
--n-gpu-layers 99 \
--override-tensor exps=CPU,attn_kv_b=CPU \
--parallel 1 \
--threads 32 \
--host 127.0.0.1 \
--port 8090
32GB VRAM
./build/bin/llama-server \
--alias anikifoss/DeepSeek-R1-0528-DQ4_K_R4 \
--model /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00001-of-00010.gguf \
--temp 0.5 --top-k 0 --top-p 1.0 --min-p 0.1 --repeat-penalty 1.0 \
--ctx-size 75000 \
-ctk f16 \
-mla 2 -fa \
-amb 1024 \
-b 2048 -ub 2048 \
-fmoe \
--n-gpu-layers 99 \
--override-tensor exps=CPU,attn_kv_b=CPU \
--parallel 1 \
--threads 32 \
--host 127.0.0.1 \
--port 8090
Customization
- Replace
/mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4
with the location of the model (where you downloaded it) - Adjust
--threads
to the number of physical cores on your system - Tweak these to your preference
--temp 0.5 --top-k 0 --top-p 1.0 --min-p 0.1 --repeat-penalty 1.0
- Add
--no-mmap
to force the model to be fully loaded into memory (this is especially important when running inference speed benchmarks) - You can increase
--parallel
, but doing so will cause your context buffer (set via--ctx-size
) to be shared between tasks executing in parallel
TODO:
- Experiment with new
-mla 3
(recent ik_llama patches enable new MLA implementation on CUDA) - Re-evaluate
-rtr
(in case Q8_0 can be repacked as Q8_0_R8 after some of the recent patches)
Inference Performance vs VRAM Considerations
You can try the following to squeeze out more context on your system:
- Running with
-ctk q8_0
can save some VRAM, but is a little slower on the target system - Reducing buffers can free up a bit more VRAM at a very minor cost to performance (
-amb 512
and-b 1024 -ub 1024
) - Try
attn_kv_b=CPU
vsattn_k_b=CPU,attn_v_b=CPU
to see which gives you the best performance - Switching to an IQ quant will save some memory at the cost of performance (very very roughly 10% memory savings at the cost of 10% in inference performance)
Optimizing for Coding
Smaller quants, like UD-Q2_K_XL
are much faster when generating tokens, but often produce code that fails to run or contains bugs. Based on empirical observations, coding seems to be strongly affected by the model quantization. So we use larger quantization where it matters to reduce perplexity while remaining within the target system constraints of 24GB-32GB VRAM, 512GB RAM.
Quantization Approach
When running with Flash MLA optimization enabled, ik_llama will unpack attention tensors into Q8_0
, so we match that in our model (similar to ubergarm's ik_llama.cpp quants). We also keep all the other small tensors as Q8_0
while also leaving any F32
tensors untouched. The MoE tensors make up the bulk of the model. The ffn_down_exps tensors are especially sensitive to quantization (we borrow this idea from unsloth
quants), so we quantize them as Q6_K_R4
. Finally, all the other large MoE tensors (ffn_up_exps, ffn_gate_exps) are quantized as Q4_K_R4
Quantization Summary:
- Keep all the small
F32
tensors untouched - Quantize all the attention and related tensors to
Q8_0
- Quantize all the ffn_down_exps tensors to
Q6_K_R4
- Quantize all the ffn_up_exps and ffn_gate_exps tensors to
Q4_K_R4
The attn_kv_b tensors are included in the original model, but they contain the same information as attn_k_b and attn_v_b tensors. Some quants, like unsloth
, remove attn_k_b and attn_v_b tensors altogether. We keep all these tensors for completeness, but push attn_kv_b out of VRAM with attn_kv_b=CPU
, since ik_llama
prefers to use attn_k_b and attn_v_b when all the tensors are available. This behavior may change between releases, so try with attn_k_b=CPU,attn_v_b=CPU
instead and check which option gives you the best performance!
No imatrix
Generally, imatrix is not recommended for Q4 and larger quants. The problem with imatrix is that it will guide what model remembers, while anything not covered by the text sample used to generate the imartrix is more likely to be forgotten. For example, an imatrix derived from wikipedia sample is likely to negatively affect tasks like coding. In other words, while imatrix can improve specific benchmarks, that are similar to the imatrix input sample, it will also skew the model performance towards tasks similar to the imatrix sample at the expense of other tasks.
Benchmarks
Benchmark System: Threadripper Pro 7975WX, 768GB DDR5@5600MHz, RTX 5090 32GB
The following quants were tested:
- Q2_K_R4 (attention -
Q8_0
, all MoE -Q2_K_R4
) - DQ2_K_R4 (attention -
Q8_0
, ffn_down_exps -Q3_K_R4
, ffn_up_exps and ffn_gate_exps -Q2_K_R4
) - Q4_K_R4 (attention -
Q8_0
, all MoE -Q4_K_R4
) - DQ4_K_R4 (attention -
Q8_0
, ffn_down_exps -Q6_K_R4
, ffn_up_exps and ffn_gate_exps -Q4_K_R4
) - Q8_0 (all -
Q8_0
)
Prompt Processing
Token Generation
Perplexity
Absolute Perplexity
Perplexity Relative to Q8_0
Perplexity Numbers
Quant | Final estimate PPL | Difference from Q8_0 |
---|---|---|
Q8_0 | 3.5184 +/- 0.01977 | +0.0000 |
DQ4_K_R4 | 3.5308 +/- 0.01986 | +0.0124 |
Q4_K_R4 | 3.5415 +/- 0.01993 | +0.0231 |
DQ2_K_R4 | 3.8099 +/- 0.02187 | +0.2915 |
Q2_K_R4 | 3.9535 +/- 0.02292 | +0.4351 |
GGUF-DUMP
Click here to see the output of `gguf-dump`
INFO:gguf-dump:* Loading: /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00001-of-00010.gguf
* File is LITTLE endian, script is running on a LITTLE endian host.
* Dumping 49 key/value pair(s)
1: UINT32 | 1 | GGUF.version = 3
2: UINT64 | 1 | GGUF.tensor_count = 186
3: UINT64 | 1 | GGUF.kv_count = 46
4: STRING | 1 | general.architecture = 'deepseek2'
5: STRING | 1 | general.type = 'model'
6: STRING | 1 | general.name = 'DeepSeek R1 0528 Bf16'
7: STRING | 1 | general.size_label = '256x21B'
8: UINT32 | 1 | deepseek2.block_count = 61
9: UINT32 | 1 | deepseek2.context_length = 163840
10: UINT32 | 1 | deepseek2.embedding_length = 7168
11: UINT32 | 1 | deepseek2.feed_forward_length = 18432
12: UINT32 | 1 | deepseek2.attention.head_count = 128
13: UINT32 | 1 | deepseek2.attention.head_count_kv = 128
14: FLOAT32 | 1 | deepseek2.rope.freq_base = 10000.0
15: FLOAT32 | 1 | deepseek2.attention.layer_norm_rms_epsilon = 9.999999974752427e-07
16: UINT32 | 1 | deepseek2.expert_used_count = 8
17: UINT32 | 1 | general.file_type = 214
18: UINT32 | 1 | deepseek2.leading_dense_block_count = 3
19: UINT32 | 1 | deepseek2.vocab_size = 129280
20: UINT32 | 1 | deepseek2.attention.q_lora_rank = 1536
21: UINT32 | 1 | deepseek2.attention.kv_lora_rank = 512
22: UINT32 | 1 | deepseek2.attention.key_length = 192
23: UINT32 | 1 | deepseek2.attention.value_length = 128
24: UINT32 | 1 | deepseek2.expert_feed_forward_length = 2048
25: UINT32 | 1 | deepseek2.expert_count = 256
26: UINT32 | 1 | deepseek2.expert_shared_count = 1
27: FLOAT32 | 1 | deepseek2.expert_weights_scale = 2.5
28: BOOL | 1 | deepseek2.expert_weights_norm = True
29: UINT32 | 1 | deepseek2.expert_gating_func = 2
30: UINT32 | 1 | deepseek2.rope.dimension_count = 64
31: STRING | 1 | deepseek2.rope.scaling.type = 'yarn'
32: FLOAT32 | 1 | deepseek2.rope.scaling.factor = 40.0
33: UINT32 | 1 | deepseek2.rope.scaling.original_context_length = 4096
34: FLOAT32 | 1 | deepseek2.rope.scaling.yarn_log_multiplier = 0.10000000149011612
35: STRING | 1 | tokenizer.ggml.model = 'gpt2'
36: STRING | 1 | tokenizer.ggml.pre = 'deepseek-v3'
37: [STRING] | 129280 | tokenizer.ggml.tokens
38: [INT32] | 129280 | tokenizer.ggml.token_type
39: [STRING] | 127741 | tokenizer.ggml.merges
40: UINT32 | 1 | tokenizer.ggml.bos_token_id = 0
41: UINT32 | 1 | tokenizer.ggml.eos_token_id = 1
42: UINT32 | 1 | tokenizer.ggml.padding_token_id = 1
43: BOOL | 1 | tokenizer.ggml.add_bos_token = True
44: BOOL | 1 | tokenizer.ggml.add_eos_token = False
45: STRING | 1 | tokenizer.chat_template = '{% if not add_generation_prompt is defined %}{% set add_gene'
46: UINT32 | 1 | general.quantization_version = 2
47: UINT16 | 1 | split.no = 0
48: UINT16 | 1 | split.count = 10
49: INT32 | 1 | split.tensors.count = 1147
* Dumping 186 tensor(s)
1: 926679040 | 7168, 129280, 1, 1 | Q8_0 | token_embd.weight
2: 7168 | 7168, 1, 1, 1 | F32 | blk.0.attn_norm.weight
3: 132120576 | 18432, 7168, 1, 1 | Q8_0 | blk.0.ffn_down.weight
4: 132120576 | 7168, 18432, 1, 1 | Q8_0 | blk.0.ffn_gate.weight
5: 132120576 | 7168, 18432, 1, 1 | Q8_0 | blk.0.ffn_up.weight
6: 7168 | 7168, 1, 1, 1 | F32 | blk.0.ffn_norm.weight
7: 512 | 512, 1, 1, 1 | F32 | blk.0.attn_kv_a_norm.weight
8: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.0.attn_kv_a_mqa.weight
9: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.0.attn_kv_b.weight
10: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.0.attn_k_b.weight
11: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.0.attn_v_b.weight
12: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.0.attn_output.weight
13: 1536 | 1536, 1, 1, 1 | F32 | blk.0.attn_q_a_norm.weight
14: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.0.attn_q_a.weight
15: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.0.attn_q_b.weight
16: 7168 | 7168, 1, 1, 1 | F32 | blk.1.attn_norm.weight
17: 132120576 | 18432, 7168, 1, 1 | Q8_0 | blk.1.ffn_down.weight
18: 132120576 | 7168, 18432, 1, 1 | Q8_0 | blk.1.ffn_gate.weight
19: 132120576 | 7168, 18432, 1, 1 | Q8_0 | blk.1.ffn_up.weight
20: 7168 | 7168, 1, 1, 1 | F32 | blk.1.ffn_norm.weight
21: 512 | 512, 1, 1, 1 | F32 | blk.1.attn_kv_a_norm.weight
22: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.1.attn_kv_a_mqa.weight
23: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.1.attn_kv_b.weight
24: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.1.attn_k_b.weight
25: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.1.attn_v_b.weight
26: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.1.attn_output.weight
27: 1536 | 1536, 1, 1, 1 | F32 | blk.1.attn_q_a_norm.weight
28: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.1.attn_q_a.weight
29: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.1.attn_q_b.weight
30: 7168 | 7168, 1, 1, 1 | F32 | blk.2.attn_norm.weight
31: 132120576 | 18432, 7168, 1, 1 | Q8_0 | blk.2.ffn_down.weight
32: 132120576 | 7168, 18432, 1, 1 | Q8_0 | blk.2.ffn_gate.weight
33: 132120576 | 7168, 18432, 1, 1 | Q8_0 | blk.2.ffn_up.weight
34: 7168 | 7168, 1, 1, 1 | F32 | blk.2.ffn_norm.weight
35: 512 | 512, 1, 1, 1 | F32 | blk.2.attn_kv_a_norm.weight
36: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.2.attn_kv_a_mqa.weight
37: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.2.attn_kv_b.weight
38: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.2.attn_k_b.weight
39: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.2.attn_v_b.weight
40: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.2.attn_output.weight
41: 1536 | 1536, 1, 1, 1 | F32 | blk.2.attn_q_a_norm.weight
42: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.2.attn_q_a.weight
43: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.2.attn_q_b.weight
44: 256 | 256, 1, 1, 1 | F32 | blk.3.exp_probs_b.bias
45: 1835008 | 7168, 256, 1, 1 | F32 | blk.3.ffn_gate_inp.weight
46: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.3.ffn_down_shexp.weight
47: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.3.ffn_gate_shexp.weight
48: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.3.ffn_up_shexp.weight
49: 512 | 512, 1, 1, 1 | F32 | blk.3.attn_kv_a_norm.weight
50: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.3.attn_kv_a_mqa.weight
51: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.3.attn_kv_b.weight
52: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.3.attn_k_b.weight
53: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.3.attn_v_b.weight
54: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.3.attn_output.weight
55: 1536 | 1536, 1, 1, 1 | F32 | blk.3.attn_q_a_norm.weight
56: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.3.attn_q_a.weight
57: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.3.attn_q_b.weight
58: 7168 | 7168, 1, 1, 1 | F32 | blk.3.attn_norm.weight
59: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.3.ffn_down_exps.weight
60: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.3.ffn_gate_exps.weight
61: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.3.ffn_up_exps.weight
62: 7168 | 7168, 1, 1, 1 | F32 | blk.3.ffn_norm.weight
63: 256 | 256, 1, 1, 1 | F32 | blk.4.exp_probs_b.bias
64: 1835008 | 7168, 256, 1, 1 | F32 | blk.4.ffn_gate_inp.weight
65: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.4.ffn_down_shexp.weight
66: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.4.ffn_gate_shexp.weight
67: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.4.ffn_up_shexp.weight
68: 512 | 512, 1, 1, 1 | F32 | blk.4.attn_kv_a_norm.weight
69: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.4.attn_kv_a_mqa.weight
70: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.4.attn_kv_b.weight
71: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.4.attn_k_b.weight
72: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.4.attn_v_b.weight
73: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.4.attn_output.weight
74: 1536 | 1536, 1, 1, 1 | F32 | blk.4.attn_q_a_norm.weight
75: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.4.attn_q_a.weight
76: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.4.attn_q_b.weight
77: 7168 | 7168, 1, 1, 1 | F32 | blk.4.attn_norm.weight
78: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.4.ffn_down_exps.weight
79: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.4.ffn_gate_exps.weight
80: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.4.ffn_up_exps.weight
81: 7168 | 7168, 1, 1, 1 | F32 | blk.4.ffn_norm.weight
82: 512 | 512, 1, 1, 1 | F32 | blk.5.attn_kv_a_norm.weight
83: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.5.attn_kv_a_mqa.weight
84: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.5.attn_kv_b.weight
85: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.5.attn_k_b.weight
86: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.5.attn_v_b.weight
87: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.5.attn_output.weight
88: 1536 | 1536, 1, 1, 1 | F32 | blk.5.attn_q_a_norm.weight
89: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.5.attn_q_a.weight
90: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.5.attn_q_b.weight
91: 256 | 256, 1, 1, 1 | F32 | blk.5.exp_probs_b.bias
92: 1835008 | 7168, 256, 1, 1 | F32 | blk.5.ffn_gate_inp.weight
93: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.5.ffn_down_shexp.weight
94: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.5.ffn_gate_shexp.weight
95: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.5.ffn_up_shexp.weight
96: 7168 | 7168, 1, 1, 1 | F32 | blk.5.attn_norm.weight
97: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.5.ffn_down_exps.weight
98: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.5.ffn_gate_exps.weight
99: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.5.ffn_up_exps.weight
100: 7168 | 7168, 1, 1, 1 | F32 | blk.5.ffn_norm.weight
101: 256 | 256, 1, 1, 1 | F32 | blk.6.exp_probs_b.bias
102: 1835008 | 7168, 256, 1, 1 | F32 | blk.6.ffn_gate_inp.weight
103: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.6.ffn_down_shexp.weight
104: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.6.ffn_gate_shexp.weight
105: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.6.ffn_up_shexp.weight
106: 512 | 512, 1, 1, 1 | F32 | blk.6.attn_kv_a_norm.weight
107: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.6.attn_kv_a_mqa.weight
108: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.6.attn_kv_b.weight
109: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.6.attn_k_b.weight
110: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.6.attn_v_b.weight
111: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.6.attn_output.weight
112: 1536 | 1536, 1, 1, 1 | F32 | blk.6.attn_q_a_norm.weight
113: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.6.attn_q_a.weight
114: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.6.attn_q_b.weight
115: 7168 | 7168, 1, 1, 1 | F32 | blk.6.attn_norm.weight
116: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.6.ffn_down_exps.weight
117: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.6.ffn_gate_exps.weight
118: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.6.ffn_up_exps.weight
119: 7168 | 7168, 1, 1, 1 | F32 | blk.6.ffn_norm.weight
120: 256 | 256, 1, 1, 1 | F32 | blk.7.exp_probs_b.bias
121: 1835008 | 7168, 256, 1, 1 | F32 | blk.7.ffn_gate_inp.weight
122: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.7.ffn_down_shexp.weight
123: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.7.ffn_gate_shexp.weight
124: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.7.ffn_up_shexp.weight
125: 512 | 512, 1, 1, 1 | F32 | blk.7.attn_kv_a_norm.weight
126: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.7.attn_kv_a_mqa.weight
127: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.7.attn_kv_b.weight
128: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.7.attn_k_b.weight
129: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.7.attn_v_b.weight
130: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.7.attn_output.weight
131: 1536 | 1536, 1, 1, 1 | F32 | blk.7.attn_q_a_norm.weight
132: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.7.attn_q_a.weight
133: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.7.attn_q_b.weight
134: 7168 | 7168, 1, 1, 1 | F32 | blk.7.attn_norm.weight
135: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.7.ffn_down_exps.weight
136: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.7.ffn_gate_exps.weight
137: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.7.ffn_up_exps.weight
138: 7168 | 7168, 1, 1, 1 | F32 | blk.7.ffn_norm.weight
139: 256 | 256, 1, 1, 1 | F32 | blk.8.exp_probs_b.bias
140: 1835008 | 7168, 256, 1, 1 | F32 | blk.8.ffn_gate_inp.weight
141: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.8.ffn_down_shexp.weight
142: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.8.ffn_gate_shexp.weight
143: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.8.ffn_up_shexp.weight
144: 512 | 512, 1, 1, 1 | F32 | blk.8.attn_kv_a_norm.weight
145: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.8.attn_kv_a_mqa.weight
146: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.8.attn_kv_b.weight
147: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.8.attn_k_b.weight
148: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.8.attn_v_b.weight
149: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.8.attn_output.weight
150: 1536 | 1536, 1, 1, 1 | F32 | blk.8.attn_q_a_norm.weight
151: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.8.attn_q_a.weight
152: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.8.attn_q_b.weight
153: 7168 | 7168, 1, 1, 1 | F32 | blk.8.attn_norm.weight
154: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.8.ffn_down_exps.weight
155: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.8.ffn_gate_exps.weight
156: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.8.ffn_up_exps.weight
157: 7168 | 7168, 1, 1, 1 | F32 | blk.8.ffn_norm.weight
158: 256 | 256, 1, 1, 1 | F32 | blk.9.exp_probs_b.bias
159: 1835008 | 7168, 256, 1, 1 | F32 | blk.9.ffn_gate_inp.weight
160: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.9.ffn_down_shexp.weight
161: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.9.ffn_gate_shexp.weight
162: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.9.ffn_up_shexp.weight
163: 512 | 512, 1, 1, 1 | F32 | blk.9.attn_kv_a_norm.weight
164: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.9.attn_kv_a_mqa.weight
165: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.9.attn_kv_b.weight
166: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.9.attn_k_b.weight
167: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.9.attn_v_b.weight
168: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.9.attn_output.weight
169: 1536 | 1536, 1, 1, 1 | F32 | blk.9.attn_q_a_norm.weight
170: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.9.attn_q_a.weight
171: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.9.attn_q_b.weight
172: 256 | 256, 1, 1, 1 | F32 | blk.10.exp_probs_b.bias
173: 1835008 | 7168, 256, 1, 1 | F32 | blk.10.ffn_gate_inp.weight
174: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.10.ffn_down_shexp.weight
175: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.10.ffn_gate_shexp.weight
176: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.10.ffn_up_shexp.weight
177: 512 | 512, 1, 1, 1 | F32 | blk.10.attn_kv_a_norm.weight
178: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.10.attn_kv_a_mqa.weight
179: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.10.attn_kv_b.weight
180: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.10.attn_k_b.weight
181: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.10.attn_v_b.weight
182: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.10.attn_output.weight
183: 1536 | 1536, 1, 1, 1 | F32 | blk.10.attn_q_a_norm.weight
184: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.10.attn_q_a.weight
185: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.10.attn_q_b.weight
186: 7168 | 7168, 1, 1, 1 | F32 | blk.9.attn_norm.weight
INFO:gguf-dump:* Loading: /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00002-of-00010.gguf
* File is LITTLE endian, script is running on a LITTLE endian host.
* Dumping 6 key/value pair(s)
1: UINT32 | 1 | GGUF.version = 3
2: UINT64 | 1 | GGUF.tensor_count = 101
3: UINT64 | 1 | GGUF.kv_count = 3
4: UINT16 | 1 | split.no = 1
5: UINT16 | 1 | split.count = 10
6: INT32 | 1 | split.tensors.count = 1147
* Dumping 101 tensor(s)
1: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.9.ffn_down_exps.weight
2: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.9.ffn_gate_exps.weight
3: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.9.ffn_up_exps.weight
4: 7168 | 7168, 1, 1, 1 | F32 | blk.9.ffn_norm.weight
5: 7168 | 7168, 1, 1, 1 | F32 | blk.10.attn_norm.weight
6: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.10.ffn_down_exps.weight
7: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.10.ffn_gate_exps.weight
8: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.10.ffn_up_exps.weight
9: 7168 | 7168, 1, 1, 1 | F32 | blk.10.ffn_norm.weight
10: 256 | 256, 1, 1, 1 | F32 | blk.11.exp_probs_b.bias
11: 1835008 | 7168, 256, 1, 1 | F32 | blk.11.ffn_gate_inp.weight
12: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.11.ffn_down_shexp.weight
13: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.11.ffn_gate_shexp.weight
14: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.11.ffn_up_shexp.weight
15: 512 | 512, 1, 1, 1 | F32 | blk.11.attn_kv_a_norm.weight
16: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.11.attn_kv_a_mqa.weight
17: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.11.attn_kv_b.weight
18: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.11.attn_k_b.weight
19: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.11.attn_v_b.weight
20: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.11.attn_output.weight
21: 1536 | 1536, 1, 1, 1 | F32 | blk.11.attn_q_a_norm.weight
22: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.11.attn_q_a.weight
23: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.11.attn_q_b.weight
24: 7168 | 7168, 1, 1, 1 | F32 | blk.11.attn_norm.weight
25: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.11.ffn_down_exps.weight
26: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.11.ffn_gate_exps.weight
27: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.11.ffn_up_exps.weight
28: 7168 | 7168, 1, 1, 1 | F32 | blk.11.ffn_norm.weight
29: 256 | 256, 1, 1, 1 | F32 | blk.12.exp_probs_b.bias
30: 1835008 | 7168, 256, 1, 1 | F32 | blk.12.ffn_gate_inp.weight
31: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.12.ffn_down_shexp.weight
32: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.12.ffn_gate_shexp.weight
33: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.12.ffn_up_shexp.weight
34: 512 | 512, 1, 1, 1 | F32 | blk.12.attn_kv_a_norm.weight
35: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.12.attn_kv_a_mqa.weight
36: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.12.attn_kv_b.weight
37: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.12.attn_k_b.weight
38: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.12.attn_v_b.weight
39: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.12.attn_output.weight
40: 1536 | 1536, 1, 1, 1 | F32 | blk.12.attn_q_a_norm.weight
41: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.12.attn_q_a.weight
42: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.12.attn_q_b.weight
43: 7168 | 7168, 1, 1, 1 | F32 | blk.12.attn_norm.weight
44: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.12.ffn_down_exps.weight
45: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.12.ffn_gate_exps.weight
46: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.12.ffn_up_exps.weight
47: 7168 | 7168, 1, 1, 1 | F32 | blk.12.ffn_norm.weight
48: 256 | 256, 1, 1, 1 | F32 | blk.13.exp_probs_b.bias
49: 1835008 | 7168, 256, 1, 1 | F32 | blk.13.ffn_gate_inp.weight
50: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.13.ffn_down_shexp.weight
51: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.13.ffn_gate_shexp.weight
52: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.13.ffn_up_shexp.weight
53: 512 | 512, 1, 1, 1 | F32 | blk.13.attn_kv_a_norm.weight
54: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.13.attn_kv_a_mqa.weight
55: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.13.attn_kv_b.weight
56: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.13.attn_k_b.weight
57: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.13.attn_v_b.weight
58: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.13.attn_output.weight
59: 1536 | 1536, 1, 1, 1 | F32 | blk.13.attn_q_a_norm.weight
60: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.13.attn_q_a.weight
61: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.13.attn_q_b.weight
62: 7168 | 7168, 1, 1, 1 | F32 | blk.13.attn_norm.weight
63: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.13.ffn_down_exps.weight
64: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.13.ffn_gate_exps.weight
65: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.13.ffn_up_exps.weight
66: 7168 | 7168, 1, 1, 1 | F32 | blk.13.ffn_norm.weight
67: 256 | 256, 1, 1, 1 | F32 | blk.14.exp_probs_b.bias
68: 1835008 | 7168, 256, 1, 1 | F32 | blk.14.ffn_gate_inp.weight
69: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.14.ffn_down_shexp.weight
70: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.14.ffn_gate_shexp.weight
71: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.14.ffn_up_shexp.weight
72: 512 | 512, 1, 1, 1 | F32 | blk.14.attn_kv_a_norm.weight
73: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.14.attn_kv_a_mqa.weight
74: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.14.attn_kv_b.weight
75: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.14.attn_k_b.weight
76: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.14.attn_v_b.weight
77: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.14.attn_output.weight
78: 1536 | 1536, 1, 1, 1 | F32 | blk.14.attn_q_a_norm.weight
79: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.14.attn_q_a.weight
80: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.14.attn_q_b.weight
81: 7168 | 7168, 1, 1, 1 | F32 | blk.14.attn_norm.weight
82: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.14.ffn_down_exps.weight
83: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.14.ffn_gate_exps.weight
84: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.14.ffn_up_exps.weight
85: 7168 | 7168, 1, 1, 1 | F32 | blk.14.ffn_norm.weight
86: 256 | 256, 1, 1, 1 | F32 | blk.15.exp_probs_b.bias
87: 1835008 | 7168, 256, 1, 1 | F32 | blk.15.ffn_gate_inp.weight
88: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.15.ffn_down_shexp.weight
89: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.15.ffn_gate_shexp.weight
90: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.15.ffn_up_shexp.weight
91: 512 | 512, 1, 1, 1 | F32 | blk.15.attn_kv_a_norm.weight
92: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.15.attn_kv_a_mqa.weight
93: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.15.attn_kv_b.weight
94: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.15.attn_k_b.weight
95: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.15.attn_v_b.weight
96: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.15.attn_output.weight
97: 1536 | 1536, 1, 1, 1 | F32 | blk.15.attn_q_a_norm.weight
98: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.15.attn_q_a.weight
99: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.15.attn_q_b.weight
100: 7168 | 7168, 1, 1, 1 | F32 | blk.15.attn_norm.weight
101: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.15.ffn_down_exps.weight
INFO:gguf-dump:* Loading: /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00003-of-00010.gguf
* File is LITTLE endian, script is running on a LITTLE endian host.
* Dumping 6 key/value pair(s)
1: UINT32 | 1 | GGUF.version = 3
2: UINT64 | 1 | GGUF.tensor_count = 132
3: UINT64 | 1 | GGUF.kv_count = 3
4: UINT16 | 1 | split.no = 2
5: UINT16 | 1 | split.count = 10
6: INT32 | 1 | split.tensors.count = 1147
* Dumping 132 tensor(s)
1: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.15.ffn_gate_exps.weight
2: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.15.ffn_up_exps.weight
3: 7168 | 7168, 1, 1, 1 | F32 | blk.15.ffn_norm.weight
4: 256 | 256, 1, 1, 1 | F32 | blk.16.exp_probs_b.bias
5: 1835008 | 7168, 256, 1, 1 | F32 | blk.16.ffn_gate_inp.weight
6: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.16.ffn_down_shexp.weight
7: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.16.ffn_gate_shexp.weight
8: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.16.ffn_up_shexp.weight
9: 512 | 512, 1, 1, 1 | F32 | blk.16.attn_kv_a_norm.weight
10: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.16.attn_kv_a_mqa.weight
11: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.16.attn_kv_b.weight
12: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.16.attn_k_b.weight
13: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.16.attn_v_b.weight
14: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.16.attn_output.weight
15: 1536 | 1536, 1, 1, 1 | F32 | blk.16.attn_q_a_norm.weight
16: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.16.attn_q_a.weight
17: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.16.attn_q_b.weight
18: 7168 | 7168, 1, 1, 1 | F32 | blk.16.attn_norm.weight
19: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.16.ffn_down_exps.weight
20: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.16.ffn_gate_exps.weight
21: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.16.ffn_up_exps.weight
22: 7168 | 7168, 1, 1, 1 | F32 | blk.16.ffn_norm.weight
23: 256 | 256, 1, 1, 1 | F32 | blk.17.exp_probs_b.bias
24: 1835008 | 7168, 256, 1, 1 | F32 | blk.17.ffn_gate_inp.weight
25: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.17.ffn_down_shexp.weight
26: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.17.ffn_gate_shexp.weight
27: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.17.ffn_up_shexp.weight
28: 512 | 512, 1, 1, 1 | F32 | blk.17.attn_kv_a_norm.weight
29: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.17.attn_kv_a_mqa.weight
30: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.17.attn_kv_b.weight
31: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.17.attn_k_b.weight
32: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.17.attn_v_b.weight
33: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.17.attn_output.weight
34: 1536 | 1536, 1, 1, 1 | F32 | blk.17.attn_q_a_norm.weight
35: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.17.attn_q_a.weight
36: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.17.attn_q_b.weight
37: 7168 | 7168, 1, 1, 1 | F32 | blk.17.attn_norm.weight
38: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.17.ffn_down_exps.weight
39: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.17.ffn_gate_exps.weight
40: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.17.ffn_up_exps.weight
41: 7168 | 7168, 1, 1, 1 | F32 | blk.17.ffn_norm.weight
42: 256 | 256, 1, 1, 1 | F32 | blk.18.exp_probs_b.bias
43: 1835008 | 7168, 256, 1, 1 | F32 | blk.18.ffn_gate_inp.weight
44: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.18.ffn_down_shexp.weight
45: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.18.ffn_gate_shexp.weight
46: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.18.ffn_up_shexp.weight
47: 512 | 512, 1, 1, 1 | F32 | blk.18.attn_kv_a_norm.weight
48: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.18.attn_kv_a_mqa.weight
49: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.18.attn_kv_b.weight
50: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.18.attn_k_b.weight
51: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.18.attn_v_b.weight
52: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.18.attn_output.weight
53: 1536 | 1536, 1, 1, 1 | F32 | blk.18.attn_q_a_norm.weight
54: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.18.attn_q_a.weight
55: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.18.attn_q_b.weight
56: 7168 | 7168, 1, 1, 1 | F32 | blk.18.attn_norm.weight
57: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.18.ffn_down_exps.weight
58: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.18.ffn_gate_exps.weight
59: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.18.ffn_up_exps.weight
60: 7168 | 7168, 1, 1, 1 | F32 | blk.18.ffn_norm.weight
61: 256 | 256, 1, 1, 1 | F32 | blk.19.exp_probs_b.bias
62: 1835008 | 7168, 256, 1, 1 | F32 | blk.19.ffn_gate_inp.weight
63: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.19.ffn_down_shexp.weight
64: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.19.ffn_gate_shexp.weight
65: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.19.ffn_up_shexp.weight
66: 512 | 512, 1, 1, 1 | F32 | blk.19.attn_kv_a_norm.weight
67: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.19.attn_kv_a_mqa.weight
68: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.19.attn_kv_b.weight
69: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.19.attn_k_b.weight
70: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.19.attn_v_b.weight
71: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.19.attn_output.weight
72: 1536 | 1536, 1, 1, 1 | F32 | blk.19.attn_q_a_norm.weight
73: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.19.attn_q_a.weight
74: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.19.attn_q_b.weight
75: 7168 | 7168, 1, 1, 1 | F32 | blk.19.attn_norm.weight
76: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.19.ffn_down_exps.weight
77: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.19.ffn_gate_exps.weight
78: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.19.ffn_up_exps.weight
79: 7168 | 7168, 1, 1, 1 | F32 | blk.19.ffn_norm.weight
80: 256 | 256, 1, 1, 1 | F32 | blk.20.exp_probs_b.bias
81: 1835008 | 7168, 256, 1, 1 | F32 | blk.20.ffn_gate_inp.weight
82: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.20.ffn_down_shexp.weight
83: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.20.ffn_gate_shexp.weight
84: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.20.ffn_up_shexp.weight
85: 512 | 512, 1, 1, 1 | F32 | blk.20.attn_kv_a_norm.weight
86: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.20.attn_kv_a_mqa.weight
87: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.20.attn_kv_b.weight
88: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.20.attn_k_b.weight
89: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.20.attn_v_b.weight
90: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.20.attn_output.weight
91: 1536 | 1536, 1, 1, 1 | F32 | blk.20.attn_q_a_norm.weight
92: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.20.attn_q_a.weight
93: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.20.attn_q_b.weight
94: 7168 | 7168, 1, 1, 1 | F32 | blk.20.attn_norm.weight
95: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.20.ffn_down_exps.weight
96: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.20.ffn_gate_exps.weight
97: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.20.ffn_up_exps.weight
98: 7168 | 7168, 1, 1, 1 | F32 | blk.20.ffn_norm.weight
99: 256 | 256, 1, 1, 1 | F32 | blk.21.exp_probs_b.bias
100: 1835008 | 7168, 256, 1, 1 | F32 | blk.21.ffn_gate_inp.weight
101: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.21.ffn_down_shexp.weight
102: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.21.ffn_gate_shexp.weight
103: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.21.ffn_up_shexp.weight
104: 512 | 512, 1, 1, 1 | F32 | blk.21.attn_kv_a_norm.weight
105: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.21.attn_kv_a_mqa.weight
106: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.21.attn_kv_b.weight
107: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.21.attn_k_b.weight
108: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.21.attn_v_b.weight
109: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.21.attn_output.weight
110: 1536 | 1536, 1, 1, 1 | F32 | blk.21.attn_q_a_norm.weight
111: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.21.attn_q_a.weight
112: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.21.attn_q_b.weight
113: 7168 | 7168, 1, 1, 1 | F32 | blk.21.attn_norm.weight
114: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.21.ffn_down_exps.weight
115: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.21.ffn_gate_exps.weight
116: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.21.ffn_up_exps.weight
117: 7168 | 7168, 1, 1, 1 | F32 | blk.21.ffn_norm.weight
118: 256 | 256, 1, 1, 1 | F32 | blk.22.exp_probs_b.bias
119: 1835008 | 7168, 256, 1, 1 | F32 | blk.22.ffn_gate_inp.weight
120: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.22.ffn_down_shexp.weight
121: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.22.ffn_gate_shexp.weight
122: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.22.ffn_up_shexp.weight
123: 512 | 512, 1, 1, 1 | F32 | blk.22.attn_kv_a_norm.weight
124: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.22.attn_kv_a_mqa.weight
125: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.22.attn_kv_b.weight
126: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.22.attn_k_b.weight
127: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.22.attn_v_b.weight
128: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.22.attn_output.weight
129: 1536 | 1536, 1, 1, 1 | F32 | blk.22.attn_q_a_norm.weight
130: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.22.attn_q_a.weight
131: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.22.attn_q_b.weight
132: 7168 | 7168, 1, 1, 1 | F32 | blk.22.attn_norm.weight
INFO:gguf-dump:* Loading: /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00004-of-00010.gguf
* File is LITTLE endian, script is running on a LITTLE endian host.
* Dumping 6 key/value pair(s)
1: UINT32 | 1 | GGUF.version = 3
2: UINT64 | 1 | GGUF.tensor_count = 115
3: UINT64 | 1 | GGUF.kv_count = 3
4: UINT16 | 1 | split.no = 3
5: UINT16 | 1 | split.count = 10
6: INT32 | 1 | split.tensors.count = 1147
* Dumping 115 tensor(s)
1: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.22.ffn_down_exps.weight
2: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.22.ffn_gate_exps.weight
3: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.22.ffn_up_exps.weight
4: 7168 | 7168, 1, 1, 1 | F32 | blk.22.ffn_norm.weight
5: 256 | 256, 1, 1, 1 | F32 | blk.23.exp_probs_b.bias
6: 1835008 | 7168, 256, 1, 1 | F32 | blk.23.ffn_gate_inp.weight
7: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.23.ffn_down_shexp.weight
8: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.23.ffn_gate_shexp.weight
9: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.23.ffn_up_shexp.weight
10: 512 | 512, 1, 1, 1 | F32 | blk.23.attn_kv_a_norm.weight
11: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.23.attn_kv_a_mqa.weight
12: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.23.attn_kv_b.weight
13: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.23.attn_k_b.weight
14: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.23.attn_v_b.weight
15: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.23.attn_output.weight
16: 1536 | 1536, 1, 1, 1 | F32 | blk.23.attn_q_a_norm.weight
17: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.23.attn_q_a.weight
18: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.23.attn_q_b.weight
19: 7168 | 7168, 1, 1, 1 | F32 | blk.23.attn_norm.weight
20: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.23.ffn_down_exps.weight
21: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.23.ffn_gate_exps.weight
22: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.23.ffn_up_exps.weight
23: 7168 | 7168, 1, 1, 1 | F32 | blk.23.ffn_norm.weight
24: 256 | 256, 1, 1, 1 | F32 | blk.24.exp_probs_b.bias
25: 1835008 | 7168, 256, 1, 1 | F32 | blk.24.ffn_gate_inp.weight
26: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.24.ffn_down_shexp.weight
27: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.24.ffn_gate_shexp.weight
28: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.24.ffn_up_shexp.weight
29: 512 | 512, 1, 1, 1 | F32 | blk.24.attn_kv_a_norm.weight
30: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.24.attn_kv_a_mqa.weight
31: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.24.attn_kv_b.weight
32: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.24.attn_k_b.weight
33: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.24.attn_v_b.weight
34: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.24.attn_output.weight
35: 1536 | 1536, 1, 1, 1 | F32 | blk.24.attn_q_a_norm.weight
36: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.24.attn_q_a.weight
37: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.24.attn_q_b.weight
38: 7168 | 7168, 1, 1, 1 | F32 | blk.24.attn_norm.weight
39: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.24.ffn_down_exps.weight
40: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.24.ffn_gate_exps.weight
41: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.24.ffn_up_exps.weight
42: 7168 | 7168, 1, 1, 1 | F32 | blk.24.ffn_norm.weight
43: 256 | 256, 1, 1, 1 | F32 | blk.25.exp_probs_b.bias
44: 1835008 | 7168, 256, 1, 1 | F32 | blk.25.ffn_gate_inp.weight
45: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.25.ffn_down_shexp.weight
46: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.25.ffn_gate_shexp.weight
47: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.25.ffn_up_shexp.weight
48: 512 | 512, 1, 1, 1 | F32 | blk.25.attn_kv_a_norm.weight
49: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.25.attn_kv_a_mqa.weight
50: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.25.attn_kv_b.weight
51: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.25.attn_k_b.weight
52: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.25.attn_v_b.weight
53: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.25.attn_output.weight
54: 1536 | 1536, 1, 1, 1 | F32 | blk.25.attn_q_a_norm.weight
55: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.25.attn_q_a.weight
56: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.25.attn_q_b.weight
57: 7168 | 7168, 1, 1, 1 | F32 | blk.25.attn_norm.weight
58: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.25.ffn_down_exps.weight
59: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.25.ffn_gate_exps.weight
60: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.25.ffn_up_exps.weight
61: 7168 | 7168, 1, 1, 1 | F32 | blk.25.ffn_norm.weight
62: 256 | 256, 1, 1, 1 | F32 | blk.26.exp_probs_b.bias
63: 1835008 | 7168, 256, 1, 1 | F32 | blk.26.ffn_gate_inp.weight
64: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.26.ffn_down_shexp.weight
65: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.26.ffn_gate_shexp.weight
66: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.26.ffn_up_shexp.weight
67: 512 | 512, 1, 1, 1 | F32 | blk.26.attn_kv_a_norm.weight
68: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.26.attn_kv_a_mqa.weight
69: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.26.attn_kv_b.weight
70: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.26.attn_k_b.weight
71: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.26.attn_v_b.weight
72: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.26.attn_output.weight
73: 1536 | 1536, 1, 1, 1 | F32 | blk.26.attn_q_a_norm.weight
74: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.26.attn_q_a.weight
75: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.26.attn_q_b.weight
76: 7168 | 7168, 1, 1, 1 | F32 | blk.26.attn_norm.weight
77: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.26.ffn_down_exps.weight
78: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.26.ffn_gate_exps.weight
79: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.26.ffn_up_exps.weight
80: 7168 | 7168, 1, 1, 1 | F32 | blk.26.ffn_norm.weight
81: 256 | 256, 1, 1, 1 | F32 | blk.27.exp_probs_b.bias
82: 1835008 | 7168, 256, 1, 1 | F32 | blk.27.ffn_gate_inp.weight
83: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.27.ffn_down_shexp.weight
84: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.27.ffn_gate_shexp.weight
85: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.27.ffn_up_shexp.weight
86: 512 | 512, 1, 1, 1 | F32 | blk.27.attn_kv_a_norm.weight
87: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.27.attn_kv_a_mqa.weight
88: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.27.attn_kv_b.weight
89: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.27.attn_k_b.weight
90: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.27.attn_v_b.weight
91: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.27.attn_output.weight
92: 1536 | 1536, 1, 1, 1 | F32 | blk.27.attn_q_a_norm.weight
93: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.27.attn_q_a.weight
94: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.27.attn_q_b.weight
95: 7168 | 7168, 1, 1, 1 | F32 | blk.27.attn_norm.weight
96: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.27.ffn_down_exps.weight
97: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.27.ffn_gate_exps.weight
98: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.27.ffn_up_exps.weight
99: 7168 | 7168, 1, 1, 1 | F32 | blk.27.ffn_norm.weight
100: 256 | 256, 1, 1, 1 | F32 | blk.28.exp_probs_b.bias
101: 1835008 | 7168, 256, 1, 1 | F32 | blk.28.ffn_gate_inp.weight
102: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.28.ffn_down_shexp.weight
103: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.28.ffn_gate_shexp.weight
104: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.28.ffn_up_shexp.weight
105: 512 | 512, 1, 1, 1 | F32 | blk.28.attn_kv_a_norm.weight
106: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.28.attn_kv_a_mqa.weight
107: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.28.attn_kv_b.weight
108: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.28.attn_k_b.weight
109: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.28.attn_v_b.weight
110: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.28.attn_output.weight
111: 1536 | 1536, 1, 1, 1 | F32 | blk.28.attn_q_a_norm.weight
112: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.28.attn_q_a.weight
113: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.28.attn_q_b.weight
114: 7168 | 7168, 1, 1, 1 | F32 | blk.28.attn_norm.weight
115: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.28.ffn_down_exps.weight
INFO:gguf-dump:* Loading: /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00005-of-00010.gguf
* File is LITTLE endian, script is running on a LITTLE endian host.
* Dumping 6 key/value pair(s)
1: UINT32 | 1 | GGUF.version = 3
2: UINT64 | 1 | GGUF.tensor_count = 132
3: UINT64 | 1 | GGUF.kv_count = 3
4: UINT16 | 1 | split.no = 4
5: UINT16 | 1 | split.count = 10
6: INT32 | 1 | split.tensors.count = 1147
* Dumping 132 tensor(s)
1: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.28.ffn_gate_exps.weight
2: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.28.ffn_up_exps.weight
3: 7168 | 7168, 1, 1, 1 | F32 | blk.28.ffn_norm.weight
4: 256 | 256, 1, 1, 1 | F32 | blk.29.exp_probs_b.bias
5: 1835008 | 7168, 256, 1, 1 | F32 | blk.29.ffn_gate_inp.weight
6: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.29.ffn_down_shexp.weight
7: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.29.ffn_gate_shexp.weight
8: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.29.ffn_up_shexp.weight
9: 512 | 512, 1, 1, 1 | F32 | blk.29.attn_kv_a_norm.weight
10: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.29.attn_kv_a_mqa.weight
11: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.29.attn_kv_b.weight
12: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.29.attn_k_b.weight
13: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.29.attn_v_b.weight
14: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.29.attn_output.weight
15: 1536 | 1536, 1, 1, 1 | F32 | blk.29.attn_q_a_norm.weight
16: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.29.attn_q_a.weight
17: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.29.attn_q_b.weight
18: 7168 | 7168, 1, 1, 1 | F32 | blk.29.attn_norm.weight
19: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.29.ffn_down_exps.weight
20: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.29.ffn_gate_exps.weight
21: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.29.ffn_up_exps.weight
22: 7168 | 7168, 1, 1, 1 | F32 | blk.29.ffn_norm.weight
23: 256 | 256, 1, 1, 1 | F32 | blk.30.exp_probs_b.bias
24: 1835008 | 7168, 256, 1, 1 | F32 | blk.30.ffn_gate_inp.weight
25: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.30.ffn_down_shexp.weight
26: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.30.ffn_gate_shexp.weight
27: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.30.ffn_up_shexp.weight
28: 512 | 512, 1, 1, 1 | F32 | blk.30.attn_kv_a_norm.weight
29: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.30.attn_kv_a_mqa.weight
30: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.30.attn_kv_b.weight
31: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.30.attn_k_b.weight
32: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.30.attn_v_b.weight
33: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.30.attn_output.weight
34: 1536 | 1536, 1, 1, 1 | F32 | blk.30.attn_q_a_norm.weight
35: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.30.attn_q_a.weight
36: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.30.attn_q_b.weight
37: 7168 | 7168, 1, 1, 1 | F32 | blk.30.attn_norm.weight
38: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.30.ffn_down_exps.weight
39: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.30.ffn_gate_exps.weight
40: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.30.ffn_up_exps.weight
41: 7168 | 7168, 1, 1, 1 | F32 | blk.30.ffn_norm.weight
42: 256 | 256, 1, 1, 1 | F32 | blk.31.exp_probs_b.bias
43: 1835008 | 7168, 256, 1, 1 | F32 | blk.31.ffn_gate_inp.weight
44: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.31.ffn_down_shexp.weight
45: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.31.ffn_gate_shexp.weight
46: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.31.ffn_up_shexp.weight
47: 512 | 512, 1, 1, 1 | F32 | blk.31.attn_kv_a_norm.weight
48: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.31.attn_kv_a_mqa.weight
49: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.31.attn_kv_b.weight
50: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.31.attn_k_b.weight
51: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.31.attn_v_b.weight
52: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.31.attn_output.weight
53: 1536 | 1536, 1, 1, 1 | F32 | blk.31.attn_q_a_norm.weight
54: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.31.attn_q_a.weight
55: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.31.attn_q_b.weight
56: 7168 | 7168, 1, 1, 1 | F32 | blk.31.attn_norm.weight
57: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.31.ffn_down_exps.weight
58: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.31.ffn_gate_exps.weight
59: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.31.ffn_up_exps.weight
60: 7168 | 7168, 1, 1, 1 | F32 | blk.31.ffn_norm.weight
61: 256 | 256, 1, 1, 1 | F32 | blk.32.exp_probs_b.bias
62: 1835008 | 7168, 256, 1, 1 | F32 | blk.32.ffn_gate_inp.weight
63: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.32.ffn_down_shexp.weight
64: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.32.ffn_gate_shexp.weight
65: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.32.ffn_up_shexp.weight
66: 512 | 512, 1, 1, 1 | F32 | blk.32.attn_kv_a_norm.weight
67: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.32.attn_kv_a_mqa.weight
68: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.32.attn_kv_b.weight
69: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.32.attn_k_b.weight
70: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.32.attn_v_b.weight
71: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.32.attn_output.weight
72: 1536 | 1536, 1, 1, 1 | F32 | blk.32.attn_q_a_norm.weight
73: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.32.attn_q_a.weight
74: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.32.attn_q_b.weight
75: 7168 | 7168, 1, 1, 1 | F32 | blk.32.attn_norm.weight
76: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.32.ffn_down_exps.weight
77: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.32.ffn_gate_exps.weight
78: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.32.ffn_up_exps.weight
79: 7168 | 7168, 1, 1, 1 | F32 | blk.32.ffn_norm.weight
80: 256 | 256, 1, 1, 1 | F32 | blk.33.exp_probs_b.bias
81: 1835008 | 7168, 256, 1, 1 | F32 | blk.33.ffn_gate_inp.weight
82: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.33.ffn_down_shexp.weight
83: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.33.ffn_gate_shexp.weight
84: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.33.ffn_up_shexp.weight
85: 512 | 512, 1, 1, 1 | F32 | blk.33.attn_kv_a_norm.weight
86: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.33.attn_kv_a_mqa.weight
87: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.33.attn_kv_b.weight
88: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.33.attn_k_b.weight
89: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.33.attn_v_b.weight
90: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.33.attn_output.weight
91: 1536 | 1536, 1, 1, 1 | F32 | blk.33.attn_q_a_norm.weight
92: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.33.attn_q_a.weight
93: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.33.attn_q_b.weight
94: 7168 | 7168, 1, 1, 1 | F32 | blk.33.attn_norm.weight
95: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.33.ffn_down_exps.weight
96: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.33.ffn_gate_exps.weight
97: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.33.ffn_up_exps.weight
98: 7168 | 7168, 1, 1, 1 | F32 | blk.33.ffn_norm.weight
99: 256 | 256, 1, 1, 1 | F32 | blk.34.exp_probs_b.bias
100: 1835008 | 7168, 256, 1, 1 | F32 | blk.34.ffn_gate_inp.weight
101: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.34.ffn_down_shexp.weight
102: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.34.ffn_gate_shexp.weight
103: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.34.ffn_up_shexp.weight
104: 512 | 512, 1, 1, 1 | F32 | blk.34.attn_kv_a_norm.weight
105: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.34.attn_kv_a_mqa.weight
106: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.34.attn_kv_b.weight
107: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.34.attn_k_b.weight
108: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.34.attn_v_b.weight
109: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.34.attn_output.weight
110: 1536 | 1536, 1, 1, 1 | F32 | blk.34.attn_q_a_norm.weight
111: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.34.attn_q_a.weight
112: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.34.attn_q_b.weight
113: 7168 | 7168, 1, 1, 1 | F32 | blk.34.attn_norm.weight
114: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.34.ffn_down_exps.weight
115: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.34.ffn_gate_exps.weight
116: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.34.ffn_up_exps.weight
117: 7168 | 7168, 1, 1, 1 | F32 | blk.34.ffn_norm.weight
118: 256 | 256, 1, 1, 1 | F32 | blk.35.exp_probs_b.bias
119: 1835008 | 7168, 256, 1, 1 | F32 | blk.35.ffn_gate_inp.weight
120: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.35.ffn_down_shexp.weight
121: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.35.ffn_gate_shexp.weight
122: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.35.ffn_up_shexp.weight
123: 512 | 512, 1, 1, 1 | F32 | blk.35.attn_kv_a_norm.weight
124: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.35.attn_kv_a_mqa.weight
125: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.35.attn_kv_b.weight
126: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.35.attn_k_b.weight
127: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.35.attn_v_b.weight
128: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.35.attn_output.weight
129: 1536 | 1536, 1, 1, 1 | F32 | blk.35.attn_q_a_norm.weight
130: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.35.attn_q_a.weight
131: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.35.attn_q_b.weight
132: 7168 | 7168, 1, 1, 1 | F32 | blk.35.attn_norm.weight
INFO:gguf-dump:* Loading: /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00006-of-00010.gguf
* File is LITTLE endian, script is running on a LITTLE endian host.
* Dumping 6 key/value pair(s)
1: UINT32 | 1 | GGUF.version = 3
2: UINT64 | 1 | GGUF.tensor_count = 115
3: UINT64 | 1 | GGUF.kv_count = 3
4: UINT16 | 1 | split.no = 5
5: UINT16 | 1 | split.count = 10
6: INT32 | 1 | split.tensors.count = 1147
* Dumping 115 tensor(s)
1: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.35.ffn_down_exps.weight
2: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.35.ffn_gate_exps.weight
3: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.35.ffn_up_exps.weight
4: 7168 | 7168, 1, 1, 1 | F32 | blk.35.ffn_norm.weight
5: 256 | 256, 1, 1, 1 | F32 | blk.36.exp_probs_b.bias
6: 1835008 | 7168, 256, 1, 1 | F32 | blk.36.ffn_gate_inp.weight
7: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.36.ffn_down_shexp.weight
8: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.36.ffn_gate_shexp.weight
9: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.36.ffn_up_shexp.weight
10: 512 | 512, 1, 1, 1 | F32 | blk.36.attn_kv_a_norm.weight
11: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.36.attn_kv_a_mqa.weight
12: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.36.attn_kv_b.weight
13: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.36.attn_k_b.weight
14: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.36.attn_v_b.weight
15: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.36.attn_output.weight
16: 1536 | 1536, 1, 1, 1 | F32 | blk.36.attn_q_a_norm.weight
17: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.36.attn_q_a.weight
18: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.36.attn_q_b.weight
19: 7168 | 7168, 1, 1, 1 | F32 | blk.36.attn_norm.weight
20: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.36.ffn_down_exps.weight
21: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.36.ffn_gate_exps.weight
22: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.36.ffn_up_exps.weight
23: 7168 | 7168, 1, 1, 1 | F32 | blk.36.ffn_norm.weight
24: 256 | 256, 1, 1, 1 | F32 | blk.37.exp_probs_b.bias
25: 1835008 | 7168, 256, 1, 1 | F32 | blk.37.ffn_gate_inp.weight
26: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.37.ffn_down_shexp.weight
27: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.37.ffn_gate_shexp.weight
28: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.37.ffn_up_shexp.weight
29: 512 | 512, 1, 1, 1 | F32 | blk.37.attn_kv_a_norm.weight
30: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.37.attn_kv_a_mqa.weight
31: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.37.attn_kv_b.weight
32: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.37.attn_k_b.weight
33: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.37.attn_v_b.weight
34: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.37.attn_output.weight
35: 1536 | 1536, 1, 1, 1 | F32 | blk.37.attn_q_a_norm.weight
36: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.37.attn_q_a.weight
37: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.37.attn_q_b.weight
38: 7168 | 7168, 1, 1, 1 | F32 | blk.37.attn_norm.weight
39: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.37.ffn_down_exps.weight
40: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.37.ffn_gate_exps.weight
41: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.37.ffn_up_exps.weight
42: 7168 | 7168, 1, 1, 1 | F32 | blk.37.ffn_norm.weight
43: 256 | 256, 1, 1, 1 | F32 | blk.38.exp_probs_b.bias
44: 1835008 | 7168, 256, 1, 1 | F32 | blk.38.ffn_gate_inp.weight
45: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.38.ffn_down_shexp.weight
46: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.38.ffn_gate_shexp.weight
47: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.38.ffn_up_shexp.weight
48: 512 | 512, 1, 1, 1 | F32 | blk.38.attn_kv_a_norm.weight
49: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.38.attn_kv_a_mqa.weight
50: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.38.attn_kv_b.weight
51: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.38.attn_k_b.weight
52: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.38.attn_v_b.weight
53: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.38.attn_output.weight
54: 1536 | 1536, 1, 1, 1 | F32 | blk.38.attn_q_a_norm.weight
55: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.38.attn_q_a.weight
56: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.38.attn_q_b.weight
57: 7168 | 7168, 1, 1, 1 | F32 | blk.38.attn_norm.weight
58: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.38.ffn_down_exps.weight
59: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.38.ffn_gate_exps.weight
60: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.38.ffn_up_exps.weight
61: 7168 | 7168, 1, 1, 1 | F32 | blk.38.ffn_norm.weight
62: 256 | 256, 1, 1, 1 | F32 | blk.39.exp_probs_b.bias
63: 1835008 | 7168, 256, 1, 1 | F32 | blk.39.ffn_gate_inp.weight
64: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.39.ffn_down_shexp.weight
65: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.39.ffn_gate_shexp.weight
66: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.39.ffn_up_shexp.weight
67: 512 | 512, 1, 1, 1 | F32 | blk.39.attn_kv_a_norm.weight
68: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.39.attn_kv_a_mqa.weight
69: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.39.attn_kv_b.weight
70: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.39.attn_k_b.weight
71: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.39.attn_v_b.weight
72: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.39.attn_output.weight
73: 1536 | 1536, 1, 1, 1 | F32 | blk.39.attn_q_a_norm.weight
74: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.39.attn_q_a.weight
75: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.39.attn_q_b.weight
76: 7168 | 7168, 1, 1, 1 | F32 | blk.39.attn_norm.weight
77: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.39.ffn_down_exps.weight
78: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.39.ffn_gate_exps.weight
79: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.39.ffn_up_exps.weight
80: 7168 | 7168, 1, 1, 1 | F32 | blk.39.ffn_norm.weight
81: 256 | 256, 1, 1, 1 | F32 | blk.40.exp_probs_b.bias
82: 1835008 | 7168, 256, 1, 1 | F32 | blk.40.ffn_gate_inp.weight
83: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.40.ffn_down_shexp.weight
84: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.40.ffn_gate_shexp.weight
85: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.40.ffn_up_shexp.weight
86: 512 | 512, 1, 1, 1 | F32 | blk.40.attn_kv_a_norm.weight
87: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.40.attn_kv_a_mqa.weight
88: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.40.attn_kv_b.weight
89: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.40.attn_k_b.weight
90: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.40.attn_v_b.weight
91: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.40.attn_output.weight
92: 1536 | 1536, 1, 1, 1 | F32 | blk.40.attn_q_a_norm.weight
93: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.40.attn_q_a.weight
94: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.40.attn_q_b.weight
95: 7168 | 7168, 1, 1, 1 | F32 | blk.40.attn_norm.weight
96: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.40.ffn_down_exps.weight
97: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.40.ffn_gate_exps.weight
98: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.40.ffn_up_exps.weight
99: 7168 | 7168, 1, 1, 1 | F32 | blk.40.ffn_norm.weight
100: 256 | 256, 1, 1, 1 | F32 | blk.41.exp_probs_b.bias
101: 1835008 | 7168, 256, 1, 1 | F32 | blk.41.ffn_gate_inp.weight
102: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.41.ffn_down_shexp.weight
103: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.41.ffn_gate_shexp.weight
104: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.41.ffn_up_shexp.weight
105: 512 | 512, 1, 1, 1 | F32 | blk.41.attn_kv_a_norm.weight
106: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.41.attn_kv_a_mqa.weight
107: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.41.attn_kv_b.weight
108: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.41.attn_k_b.weight
109: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.41.attn_v_b.weight
110: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.41.attn_output.weight
111: 1536 | 1536, 1, 1, 1 | F32 | blk.41.attn_q_a_norm.weight
112: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.41.attn_q_a.weight
113: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.41.attn_q_b.weight
114: 7168 | 7168, 1, 1, 1 | F32 | blk.41.attn_norm.weight
115: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.41.ffn_down_exps.weight
INFO:gguf-dump:* Loading: /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00007-of-00010.gguf
* File is LITTLE endian, script is running on a LITTLE endian host.
* Dumping 6 key/value pair(s)
1: UINT32 | 1 | GGUF.version = 3
2: UINT64 | 1 | GGUF.tensor_count = 132
3: UINT64 | 1 | GGUF.kv_count = 3
4: UINT16 | 1 | split.no = 6
5: UINT16 | 1 | split.count = 10
6: INT32 | 1 | split.tensors.count = 1147
* Dumping 132 tensor(s)
1: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.41.ffn_gate_exps.weight
2: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.41.ffn_up_exps.weight
3: 7168 | 7168, 1, 1, 1 | F32 | blk.41.ffn_norm.weight
4: 256 | 256, 1, 1, 1 | F32 | blk.42.exp_probs_b.bias
5: 1835008 | 7168, 256, 1, 1 | F32 | blk.42.ffn_gate_inp.weight
6: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.42.ffn_down_shexp.weight
7: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.42.ffn_gate_shexp.weight
8: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.42.ffn_up_shexp.weight
9: 512 | 512, 1, 1, 1 | F32 | blk.42.attn_kv_a_norm.weight
10: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.42.attn_kv_a_mqa.weight
11: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.42.attn_kv_b.weight
12: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.42.attn_k_b.weight
13: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.42.attn_v_b.weight
14: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.42.attn_output.weight
15: 1536 | 1536, 1, 1, 1 | F32 | blk.42.attn_q_a_norm.weight
16: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.42.attn_q_a.weight
17: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.42.attn_q_b.weight
18: 7168 | 7168, 1, 1, 1 | F32 | blk.42.attn_norm.weight
19: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.42.ffn_down_exps.weight
20: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.42.ffn_gate_exps.weight
21: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.42.ffn_up_exps.weight
22: 7168 | 7168, 1, 1, 1 | F32 | blk.42.ffn_norm.weight
23: 256 | 256, 1, 1, 1 | F32 | blk.43.exp_probs_b.bias
24: 1835008 | 7168, 256, 1, 1 | F32 | blk.43.ffn_gate_inp.weight
25: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.43.ffn_down_shexp.weight
26: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.43.ffn_gate_shexp.weight
27: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.43.ffn_up_shexp.weight
28: 512 | 512, 1, 1, 1 | F32 | blk.43.attn_kv_a_norm.weight
29: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.43.attn_kv_a_mqa.weight
30: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.43.attn_kv_b.weight
31: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.43.attn_k_b.weight
32: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.43.attn_v_b.weight
33: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.43.attn_output.weight
34: 1536 | 1536, 1, 1, 1 | F32 | blk.43.attn_q_a_norm.weight
35: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.43.attn_q_a.weight
36: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.43.attn_q_b.weight
37: 7168 | 7168, 1, 1, 1 | F32 | blk.43.attn_norm.weight
38: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.43.ffn_down_exps.weight
39: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.43.ffn_gate_exps.weight
40: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.43.ffn_up_exps.weight
41: 7168 | 7168, 1, 1, 1 | F32 | blk.43.ffn_norm.weight
42: 256 | 256, 1, 1, 1 | F32 | blk.44.exp_probs_b.bias
43: 1835008 | 7168, 256, 1, 1 | F32 | blk.44.ffn_gate_inp.weight
44: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.44.ffn_down_shexp.weight
45: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.44.ffn_gate_shexp.weight
46: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.44.ffn_up_shexp.weight
47: 512 | 512, 1, 1, 1 | F32 | blk.44.attn_kv_a_norm.weight
48: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.44.attn_kv_a_mqa.weight
49: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.44.attn_kv_b.weight
50: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.44.attn_k_b.weight
51: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.44.attn_v_b.weight
52: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.44.attn_output.weight
53: 1536 | 1536, 1, 1, 1 | F32 | blk.44.attn_q_a_norm.weight
54: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.44.attn_q_a.weight
55: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.44.attn_q_b.weight
56: 7168 | 7168, 1, 1, 1 | F32 | blk.44.attn_norm.weight
57: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.44.ffn_down_exps.weight
58: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.44.ffn_gate_exps.weight
59: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.44.ffn_up_exps.weight
60: 7168 | 7168, 1, 1, 1 | F32 | blk.44.ffn_norm.weight
61: 256 | 256, 1, 1, 1 | F32 | blk.45.exp_probs_b.bias
62: 1835008 | 7168, 256, 1, 1 | F32 | blk.45.ffn_gate_inp.weight
63: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.45.ffn_down_shexp.weight
64: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.45.ffn_gate_shexp.weight
65: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.45.ffn_up_shexp.weight
66: 512 | 512, 1, 1, 1 | F32 | blk.45.attn_kv_a_norm.weight
67: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.45.attn_kv_a_mqa.weight
68: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.45.attn_kv_b.weight
69: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.45.attn_k_b.weight
70: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.45.attn_v_b.weight
71: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.45.attn_output.weight
72: 1536 | 1536, 1, 1, 1 | F32 | blk.45.attn_q_a_norm.weight
73: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.45.attn_q_a.weight
74: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.45.attn_q_b.weight
75: 7168 | 7168, 1, 1, 1 | F32 | blk.45.attn_norm.weight
76: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.45.ffn_down_exps.weight
77: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.45.ffn_gate_exps.weight
78: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.45.ffn_up_exps.weight
79: 7168 | 7168, 1, 1, 1 | F32 | blk.45.ffn_norm.weight
80: 256 | 256, 1, 1, 1 | F32 | blk.46.exp_probs_b.bias
81: 1835008 | 7168, 256, 1, 1 | F32 | blk.46.ffn_gate_inp.weight
82: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.46.ffn_down_shexp.weight
83: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.46.ffn_gate_shexp.weight
84: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.46.ffn_up_shexp.weight
85: 512 | 512, 1, 1, 1 | F32 | blk.46.attn_kv_a_norm.weight
86: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.46.attn_kv_a_mqa.weight
87: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.46.attn_kv_b.weight
88: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.46.attn_k_b.weight
89: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.46.attn_v_b.weight
90: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.46.attn_output.weight
91: 1536 | 1536, 1, 1, 1 | F32 | blk.46.attn_q_a_norm.weight
92: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.46.attn_q_a.weight
93: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.46.attn_q_b.weight
94: 7168 | 7168, 1, 1, 1 | F32 | blk.46.attn_norm.weight
95: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.46.ffn_down_exps.weight
96: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.46.ffn_gate_exps.weight
97: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.46.ffn_up_exps.weight
98: 7168 | 7168, 1, 1, 1 | F32 | blk.46.ffn_norm.weight
99: 256 | 256, 1, 1, 1 | F32 | blk.47.exp_probs_b.bias
100: 1835008 | 7168, 256, 1, 1 | F32 | blk.47.ffn_gate_inp.weight
101: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.47.ffn_down_shexp.weight
102: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.47.ffn_gate_shexp.weight
103: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.47.ffn_up_shexp.weight
104: 512 | 512, 1, 1, 1 | F32 | blk.47.attn_kv_a_norm.weight
105: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.47.attn_kv_a_mqa.weight
106: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.47.attn_kv_b.weight
107: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.47.attn_k_b.weight
108: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.47.attn_v_b.weight
109: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.47.attn_output.weight
110: 1536 | 1536, 1, 1, 1 | F32 | blk.47.attn_q_a_norm.weight
111: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.47.attn_q_a.weight
112: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.47.attn_q_b.weight
113: 7168 | 7168, 1, 1, 1 | F32 | blk.47.attn_norm.weight
114: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.47.ffn_down_exps.weight
115: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.47.ffn_gate_exps.weight
116: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.47.ffn_up_exps.weight
117: 7168 | 7168, 1, 1, 1 | F32 | blk.47.ffn_norm.weight
118: 256 | 256, 1, 1, 1 | F32 | blk.48.exp_probs_b.bias
119: 1835008 | 7168, 256, 1, 1 | F32 | blk.48.ffn_gate_inp.weight
120: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.48.ffn_down_shexp.weight
121: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.48.ffn_gate_shexp.weight
122: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.48.ffn_up_shexp.weight
123: 512 | 512, 1, 1, 1 | F32 | blk.48.attn_kv_a_norm.weight
124: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.48.attn_kv_a_mqa.weight
125: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.48.attn_kv_b.weight
126: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.48.attn_k_b.weight
127: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.48.attn_v_b.weight
128: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.48.attn_output.weight
129: 1536 | 1536, 1, 1, 1 | F32 | blk.48.attn_q_a_norm.weight
130: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.48.attn_q_a.weight
131: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.48.attn_q_b.weight
132: 7168 | 7168, 1, 1, 1 | F32 | blk.48.attn_norm.weight
INFO:gguf-dump:* Loading: /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00008-of-00010.gguf
* File is LITTLE endian, script is running on a LITTLE endian host.
* Dumping 6 key/value pair(s)
1: UINT32 | 1 | GGUF.version = 3
2: UINT64 | 1 | GGUF.tensor_count = 115
3: UINT64 | 1 | GGUF.kv_count = 3
4: UINT16 | 1 | split.no = 7
5: UINT16 | 1 | split.count = 10
6: INT32 | 1 | split.tensors.count = 1147
* Dumping 115 tensor(s)
1: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.48.ffn_down_exps.weight
2: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.48.ffn_gate_exps.weight
3: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.48.ffn_up_exps.weight
4: 7168 | 7168, 1, 1, 1 | F32 | blk.48.ffn_norm.weight
5: 256 | 256, 1, 1, 1 | F32 | blk.49.exp_probs_b.bias
6: 1835008 | 7168, 256, 1, 1 | F32 | blk.49.ffn_gate_inp.weight
7: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.49.ffn_down_shexp.weight
8: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.49.ffn_gate_shexp.weight
9: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.49.ffn_up_shexp.weight
10: 512 | 512, 1, 1, 1 | F32 | blk.49.attn_kv_a_norm.weight
11: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.49.attn_kv_a_mqa.weight
12: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.49.attn_kv_b.weight
13: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.49.attn_k_b.weight
14: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.49.attn_v_b.weight
15: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.49.attn_output.weight
16: 1536 | 1536, 1, 1, 1 | F32 | blk.49.attn_q_a_norm.weight
17: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.49.attn_q_a.weight
18: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.49.attn_q_b.weight
19: 7168 | 7168, 1, 1, 1 | F32 | blk.49.attn_norm.weight
20: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.49.ffn_down_exps.weight
21: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.49.ffn_gate_exps.weight
22: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.49.ffn_up_exps.weight
23: 7168 | 7168, 1, 1, 1 | F32 | blk.49.ffn_norm.weight
24: 256 | 256, 1, 1, 1 | F32 | blk.50.exp_probs_b.bias
25: 1835008 | 7168, 256, 1, 1 | F32 | blk.50.ffn_gate_inp.weight
26: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.50.ffn_down_shexp.weight
27: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.50.ffn_gate_shexp.weight
28: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.50.ffn_up_shexp.weight
29: 512 | 512, 1, 1, 1 | F32 | blk.50.attn_kv_a_norm.weight
30: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.50.attn_kv_a_mqa.weight
31: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.50.attn_kv_b.weight
32: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.50.attn_k_b.weight
33: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.50.attn_v_b.weight
34: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.50.attn_output.weight
35: 1536 | 1536, 1, 1, 1 | F32 | blk.50.attn_q_a_norm.weight
36: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.50.attn_q_a.weight
37: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.50.attn_q_b.weight
38: 7168 | 7168, 1, 1, 1 | F32 | blk.50.attn_norm.weight
39: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.50.ffn_down_exps.weight
40: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.50.ffn_gate_exps.weight
41: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.50.ffn_up_exps.weight
42: 7168 | 7168, 1, 1, 1 | F32 | blk.50.ffn_norm.weight
43: 256 | 256, 1, 1, 1 | F32 | blk.51.exp_probs_b.bias
44: 1835008 | 7168, 256, 1, 1 | F32 | blk.51.ffn_gate_inp.weight
45: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.51.ffn_down_shexp.weight
46: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.51.ffn_gate_shexp.weight
47: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.51.ffn_up_shexp.weight
48: 512 | 512, 1, 1, 1 | F32 | blk.51.attn_kv_a_norm.weight
49: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.51.attn_kv_a_mqa.weight
50: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.51.attn_kv_b.weight
51: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.51.attn_k_b.weight
52: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.51.attn_v_b.weight
53: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.51.attn_output.weight
54: 1536 | 1536, 1, 1, 1 | F32 | blk.51.attn_q_a_norm.weight
55: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.51.attn_q_a.weight
56: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.51.attn_q_b.weight
57: 7168 | 7168, 1, 1, 1 | F32 | blk.51.attn_norm.weight
58: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.51.ffn_down_exps.weight
59: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.51.ffn_gate_exps.weight
60: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.51.ffn_up_exps.weight
61: 7168 | 7168, 1, 1, 1 | F32 | blk.51.ffn_norm.weight
62: 256 | 256, 1, 1, 1 | F32 | blk.52.exp_probs_b.bias
63: 1835008 | 7168, 256, 1, 1 | F32 | blk.52.ffn_gate_inp.weight
64: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.52.ffn_down_shexp.weight
65: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.52.ffn_gate_shexp.weight
66: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.52.ffn_up_shexp.weight
67: 512 | 512, 1, 1, 1 | F32 | blk.52.attn_kv_a_norm.weight
68: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.52.attn_kv_a_mqa.weight
69: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.52.attn_kv_b.weight
70: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.52.attn_k_b.weight
71: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.52.attn_v_b.weight
72: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.52.attn_output.weight
73: 1536 | 1536, 1, 1, 1 | F32 | blk.52.attn_q_a_norm.weight
74: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.52.attn_q_a.weight
75: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.52.attn_q_b.weight
76: 7168 | 7168, 1, 1, 1 | F32 | blk.52.attn_norm.weight
77: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.52.ffn_down_exps.weight
78: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.52.ffn_gate_exps.weight
79: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.52.ffn_up_exps.weight
80: 7168 | 7168, 1, 1, 1 | F32 | blk.52.ffn_norm.weight
81: 256 | 256, 1, 1, 1 | F32 | blk.53.exp_probs_b.bias
82: 1835008 | 7168, 256, 1, 1 | F32 | blk.53.ffn_gate_inp.weight
83: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.53.ffn_down_shexp.weight
84: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.53.ffn_gate_shexp.weight
85: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.53.ffn_up_shexp.weight
86: 512 | 512, 1, 1, 1 | F32 | blk.53.attn_kv_a_norm.weight
87: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.53.attn_kv_a_mqa.weight
88: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.53.attn_kv_b.weight
89: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.53.attn_k_b.weight
90: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.53.attn_v_b.weight
91: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.53.attn_output.weight
92: 1536 | 1536, 1, 1, 1 | F32 | blk.53.attn_q_a_norm.weight
93: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.53.attn_q_a.weight
94: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.53.attn_q_b.weight
95: 7168 | 7168, 1, 1, 1 | F32 | blk.53.attn_norm.weight
96: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.53.ffn_down_exps.weight
97: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.53.ffn_gate_exps.weight
98: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.53.ffn_up_exps.weight
99: 7168 | 7168, 1, 1, 1 | F32 | blk.53.ffn_norm.weight
100: 256 | 256, 1, 1, 1 | F32 | blk.54.exp_probs_b.bias
101: 1835008 | 7168, 256, 1, 1 | F32 | blk.54.ffn_gate_inp.weight
102: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.54.ffn_down_shexp.weight
103: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.54.ffn_gate_shexp.weight
104: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.54.ffn_up_shexp.weight
105: 512 | 512, 1, 1, 1 | F32 | blk.54.attn_kv_a_norm.weight
106: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.54.attn_kv_a_mqa.weight
107: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.54.attn_kv_b.weight
108: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.54.attn_k_b.weight
109: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.54.attn_v_b.weight
110: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.54.attn_output.weight
111: 1536 | 1536, 1, 1, 1 | F32 | blk.54.attn_q_a_norm.weight
112: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.54.attn_q_a.weight
113: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.54.attn_q_b.weight
114: 7168 | 7168, 1, 1, 1 | F32 | blk.54.attn_norm.weight
115: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.54.ffn_down_exps.weight
INFO:gguf-dump:* Loading: /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00009-of-00010.gguf
* File is LITTLE endian, script is running on a LITTLE endian host.
* Dumping 6 key/value pair(s)
1: UINT32 | 1 | GGUF.version = 3
2: UINT64 | 1 | GGUF.tensor_count = 116
3: UINT64 | 1 | GGUF.kv_count = 3
4: UINT16 | 1 | split.no = 8
5: UINT16 | 1 | split.count = 10
6: INT32 | 1 | split.tensors.count = 1147
* Dumping 116 tensor(s)
1: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.54.ffn_gate_exps.weight
2: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.54.ffn_up_exps.weight
3: 7168 | 7168, 1, 1, 1 | F32 | blk.54.ffn_norm.weight
4: 256 | 256, 1, 1, 1 | F32 | blk.55.exp_probs_b.bias
5: 1835008 | 7168, 256, 1, 1 | F32 | blk.55.ffn_gate_inp.weight
6: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.55.ffn_down_shexp.weight
7: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.55.ffn_gate_shexp.weight
8: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.55.ffn_up_shexp.weight
9: 512 | 512, 1, 1, 1 | F32 | blk.55.attn_kv_a_norm.weight
10: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.55.attn_kv_a_mqa.weight
11: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.55.attn_kv_b.weight
12: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.55.attn_k_b.weight
13: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.55.attn_v_b.weight
14: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.55.attn_output.weight
15: 1536 | 1536, 1, 1, 1 | F32 | blk.55.attn_q_a_norm.weight
16: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.55.attn_q_a.weight
17: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.55.attn_q_b.weight
18: 7168 | 7168, 1, 1, 1 | F32 | blk.55.attn_norm.weight
19: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.55.ffn_down_exps.weight
20: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.55.ffn_gate_exps.weight
21: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.55.ffn_up_exps.weight
22: 7168 | 7168, 1, 1, 1 | F32 | blk.55.ffn_norm.weight
23: 256 | 256, 1, 1, 1 | F32 | blk.56.exp_probs_b.bias
24: 1835008 | 7168, 256, 1, 1 | F32 | blk.56.ffn_gate_inp.weight
25: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.56.ffn_down_shexp.weight
26: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.56.ffn_gate_shexp.weight
27: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.56.ffn_up_shexp.weight
28: 512 | 512, 1, 1, 1 | F32 | blk.56.attn_kv_a_norm.weight
29: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.56.attn_kv_a_mqa.weight
30: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.56.attn_kv_b.weight
31: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.56.attn_k_b.weight
32: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.56.attn_v_b.weight
33: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.56.attn_output.weight
34: 1536 | 1536, 1, 1, 1 | F32 | blk.56.attn_q_a_norm.weight
35: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.56.attn_q_a.weight
36: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.56.attn_q_b.weight
37: 7168 | 7168, 1, 1, 1 | F32 | blk.56.attn_norm.weight
38: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.56.ffn_down_exps.weight
39: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.56.ffn_gate_exps.weight
40: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.56.ffn_up_exps.weight
41: 7168 | 7168, 1, 1, 1 | F32 | blk.56.ffn_norm.weight
42: 256 | 256, 1, 1, 1 | F32 | blk.57.exp_probs_b.bias
43: 1835008 | 7168, 256, 1, 1 | F32 | blk.57.ffn_gate_inp.weight
44: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.57.ffn_down_shexp.weight
45: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.57.ffn_gate_shexp.weight
46: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.57.ffn_up_shexp.weight
47: 512 | 512, 1, 1, 1 | F32 | blk.57.attn_kv_a_norm.weight
48: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.57.attn_kv_a_mqa.weight
49: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.57.attn_kv_b.weight
50: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.57.attn_k_b.weight
51: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.57.attn_v_b.weight
52: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.57.attn_output.weight
53: 1536 | 1536, 1, 1, 1 | F32 | blk.57.attn_q_a_norm.weight
54: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.57.attn_q_a.weight
55: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.57.attn_q_b.weight
56: 7168 | 7168, 1, 1, 1 | F32 | blk.57.attn_norm.weight
57: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.57.ffn_down_exps.weight
58: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.57.ffn_gate_exps.weight
59: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.57.ffn_up_exps.weight
60: 7168 | 7168, 1, 1, 1 | F32 | blk.57.ffn_norm.weight
61: 256 | 256, 1, 1, 1 | F32 | blk.58.exp_probs_b.bias
62: 1835008 | 7168, 256, 1, 1 | F32 | blk.58.ffn_gate_inp.weight
63: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.58.ffn_down_shexp.weight
64: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.58.ffn_gate_shexp.weight
65: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.58.ffn_up_shexp.weight
66: 512 | 512, 1, 1, 1 | F32 | blk.58.attn_kv_a_norm.weight
67: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.58.attn_kv_a_mqa.weight
68: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.58.attn_kv_b.weight
69: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.58.attn_k_b.weight
70: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.58.attn_v_b.weight
71: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.58.attn_output.weight
72: 1536 | 1536, 1, 1, 1 | F32 | blk.58.attn_q_a_norm.weight
73: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.58.attn_q_a.weight
74: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.58.attn_q_b.weight
75: 7168 | 7168, 1, 1, 1 | F32 | blk.58.attn_norm.weight
76: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.58.ffn_down_exps.weight
77: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.58.ffn_gate_exps.weight
78: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.58.ffn_up_exps.weight
79: 7168 | 7168, 1, 1, 1 | F32 | blk.58.ffn_norm.weight
80: 256 | 256, 1, 1, 1 | F32 | blk.59.exp_probs_b.bias
81: 1835008 | 7168, 256, 1, 1 | F32 | blk.59.ffn_gate_inp.weight
82: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.59.ffn_down_shexp.weight
83: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.59.ffn_gate_shexp.weight
84: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.59.ffn_up_shexp.weight
85: 512 | 512, 1, 1, 1 | F32 | blk.59.attn_kv_a_norm.weight
86: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.59.attn_kv_a_mqa.weight
87: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.59.attn_kv_b.weight
88: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.59.attn_k_b.weight
89: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.59.attn_v_b.weight
90: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.59.attn_output.weight
91: 1536 | 1536, 1, 1, 1 | F32 | blk.59.attn_q_a_norm.weight
92: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.59.attn_q_a.weight
93: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.59.attn_q_b.weight
94: 7168 | 7168, 1, 1, 1 | F32 | blk.59.attn_norm.weight
95: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.59.ffn_down_exps.weight
96: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.59.ffn_gate_exps.weight
97: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.59.ffn_up_exps.weight
98: 7168 | 7168, 1, 1, 1 | F32 | blk.59.ffn_norm.weight
99: 256 | 256, 1, 1, 1 | F32 | blk.60.exp_probs_b.bias
100: 1835008 | 7168, 256, 1, 1 | F32 | blk.60.ffn_gate_inp.weight
101: 14680064 | 2048, 7168, 1, 1 | Q8_0 | blk.60.ffn_down_shexp.weight
102: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.60.ffn_gate_shexp.weight
103: 14680064 | 7168, 2048, 1, 1 | Q8_0 | blk.60.ffn_up_shexp.weight
104: 512 | 512, 1, 1, 1 | F32 | blk.60.attn_kv_a_norm.weight
105: 4128768 | 7168, 576, 1, 1 | Q8_0 | blk.60.attn_kv_a_mqa.weight
106: 16777216 | 512, 32768, 1, 1 | Q8_0 | blk.60.attn_kv_b.weight
107: 8388608 | 128, 65536, 1, 1 | Q8_0 | blk.60.attn_k_b.weight
108: 8388608 | 512, 16384, 1, 1 | Q8_0 | blk.60.attn_v_b.weight
109: 117440512 | 16384, 7168, 1, 1 | Q8_0 | blk.60.attn_output.weight
110: 1536 | 1536, 1, 1, 1 | F32 | blk.60.attn_q_a_norm.weight
111: 11010048 | 7168, 1536, 1, 1 | Q8_0 | blk.60.attn_q_a.weight
112: 37748736 | 1536, 24576, 1, 1 | Q8_0 | blk.60.attn_q_b.weight
113: 926679040 | 7168, 129280, 1, 1 | Q8_0 | output.weight
114: 7168 | 7168, 1, 1, 1 | F32 | blk.60.attn_norm.weight
115: 3758096384 | 2048, 7168, 256, 1 | Q6_K_R4 | blk.60.ffn_down_exps.weight
116: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.60.ffn_gate_exps.weight
INFO:gguf-dump:* Loading: /mnt/data/Models/anikifoss/DeepSeek-R1-0528-DQ4_K_R4/DeepSeek-R1-0528-DQ4_K_R4-00010-of-00010.gguf
* File is LITTLE endian, script is running on a LITTLE endian host.
* Dumping 6 key/value pair(s)
1: UINT32 | 1 | GGUF.version = 3
2: UINT64 | 1 | GGUF.tensor_count = 3
3: UINT64 | 1 | GGUF.kv_count = 3
4: UINT16 | 1 | split.no = 9
5: UINT16 | 1 | split.count = 10
6: INT32 | 1 | split.tensors.count = 1147
* Dumping 3 tensor(s)
1: 3758096384 | 7168, 2048, 256, 1 | Q4_K_R4 | blk.60.ffn_up_exps.weight
2: 7168 | 7168, 1, 1, 1 | F32 | blk.60.ffn_norm.weight
3: 7168 | 7168, 1, 1, 1 | F32 | output_norm.weight
- Downloads last month
- 66
Model tree for anikifoss/DeepSeek-R1-0528-DQ4_K_R4
Base model
deepseek-ai/DeepSeek-R1-0528