roleplaiapp
commited on
update
Browse files
README.md
CHANGED
@@ -35,12 +35,12 @@ tags:
|
|
35 |
**Organization:** `PowerInfer`
|
36 |
**Quantized File:** `smallthinker-3b-preview-iq3_m-imat.gguf`
|
37 |
**Quantization:** `GGUF`
|
38 |
-
**Quantization Method:** `
|
39 |
**Use Imatrix:** `True`
|
40 |
**Split Model:** `False`
|
41 |
|
42 |
## Overview
|
43 |
-
This is an GGUF
|
44 |
|
45 |
## Quantization By
|
46 |
I often have idle A100 GPUs while building/testing and training the RP app, so I put them to use quantizing models.
|
|
|
35 |
**Organization:** `PowerInfer`
|
36 |
**Quantized File:** `smallthinker-3b-preview-iq3_m-imat.gguf`
|
37 |
**Quantization:** `GGUF`
|
38 |
+
**Quantization Method:** `IQ3_M`
|
39 |
**Use Imatrix:** `True`
|
40 |
**Split Model:** `False`
|
41 |
|
42 |
## Overview
|
43 |
+
This is an imatrix GGUF IQ3_M quantized version of [SmallThinker-3B-Preview](https://huggingface.co/PowerInfer/SmallThinker-3B-Preview).
|
44 |
|
45 |
## Quantization By
|
46 |
I often have idle A100 GPUs while building/testing and training the RP app, so I put them to use quantizing models.
|