DESCRIPTION

UPDATE: 2025-02-12

Velvet-14B converted to GGUF format (F32) with fbuciuni90/llama.cpp fork and quantized with ggerganov/llama.cpp commit b4689.

NOTE: The Velvet tokenizer is not yet compatible with ggerganov/llama.cpp. Please wait for pull request #11716 to be merged, or compile it yourself.

Original Model: https://huggingface.co/Almawave/Velvet-14B

PROMPT FORMAT

Basic prompt format:

<s><instruction>{prompt}</instruction>

Prompt format with system message:

<s><instruction>{system_prompt}\n\n{prompt}</instruction>

DOWNLOAD

BYE :3

Downloads last month
31
GGUF
Model size
14.1B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

3-bit

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for DagMeow/Velvet-14B-GGUF

Quantized
(7)
this model