DESCRIPTION
UPDATE: 2025-02-12
Velvet-14B converted to GGUF format (F32) with fbuciuni90/llama.cpp fork and quantized with ggerganov/llama.cpp commit b4689.
NOTE: The Velvet tokenizer is not yet compatible with ggerganov/llama.cpp. Please wait for pull request #11716 to be merged, or compile it yourself.
Original Model: https://huggingface.co/Almawave/Velvet-14B
PROMPT FORMAT
Basic prompt format:
<s><instruction>{prompt}</instruction>
Prompt format with system message:
<s><instruction>{system_prompt}\n\n{prompt}</instruction>
DOWNLOAD
Quant | Link |
---|---|
Q3_K_S | Velvet-14B-Q3_K_S.gguf |
Q4_K_S | Velvet-14B-Q4_K_S.gguf |
BYE :3
- Downloads last month
- 31
Hardware compatibility
Log In
to view the estimation
3-bit
4-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
HF Inference deployability: The model has no library tag.
Model tree for DagMeow/Velvet-14B-GGUF
Base model
Almawave/Velvet-14B