File size: 1,195 Bytes
a099f72 53f327e a099f72 84d2042 2252f2d a099f72 7f1b7a6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 |
---
license: other
tags:
- chatbot
- gptq
- storywriting
---
# chronos-13b-8K-4bit
The original Chronos-13B model was merged with a LoRA trained on a majority of 1500 samples in the 8000 token range in the same style, with a cutoff of 8k tokens in full 8bit. It is meant to be used standalone, but if you would like the LoRA to merge/combine on your own, you can find it here https://huggingface.co/ZeusLabs/chronos-13b-8k-lora
The `config.json` includes modifications allowing extended context so you will need to use it with `trust_remote_code` if not using Exllama.
4bit (int4) quantized version using `true-sequential` and `groupsize 128` of https://huggingface.co/elinas/chronos-13b plus https://huggingface.co/ZeusLabs/chronos-13b-8k-lora
This model is primarily focused on chat, roleplay, and storywriting, but can accomplish other tasks such as simple reasoning and coding.
Chronos generates very long outputs with coherent text, largely due to the human inputs it was trained on.
This model uses Alpaca formatting, so for optimal model performance, use:
```
### Instruction:
Your instruction or question here.
### Response:
```
[Zeus Labs Discord](https://discord.gg/76e2HBzRKD) |