Update README.md
Browse files
README.md
CHANGED
@@ -28,7 +28,8 @@ By default, this model caps out at 32K context. Additional configuration is requ
|
|
28 |
"factor": 4.0,
|
29 |
"original_max_position_embeddings": 32768,
|
30 |
"type": "yarn"
|
31 |
-
}
|
|
|
32 |
|
33 |
Once this is done, you can push the model to 64K context at Q4 KV cache quantization on a single 24GB VRAM card with minimal loss of accuracy.
|
34 |
|
|
|
28 |
"factor": 4.0,
|
29 |
"original_max_position_embeddings": 32768,
|
30 |
"type": "yarn"
|
31 |
+
}
|
32 |
+
```
|
33 |
|
34 |
Once this is done, you can push the model to 64K context at Q4 KV cache quantization on a single 24GB VRAM card with minimal loss of accuracy.
|
35 |
|