dinerburger commited on
Commit
270a50c
·
verified ·
1 Parent(s): 0051aa1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -28,7 +28,8 @@ By default, this model caps out at 32K context. Additional configuration is requ
28
  "factor": 4.0,
29
  "original_max_position_embeddings": 32768,
30
  "type": "yarn"
31
- }```
 
32
 
33
  Once this is done, you can push the model to 64K context at Q4 KV cache quantization on a single 24GB VRAM card with minimal loss of accuracy.
34
 
 
28
  "factor": 4.0,
29
  "original_max_position_embeddings": 32768,
30
  "type": "yarn"
31
+ }
32
+ ```
33
 
34
  Once this is done, you can push the model to 64K context at Q4 KV cache quantization on a single 24GB VRAM card with minimal loss of accuracy.
35