NOTE: Requires 128k context support -> https://github.com/ggerganov/llama.cpp/releases/tag/b2961.
GGUF llama.cpp quantized version of:
- Original model: Microsoft/Phi-3-medium-128k-instruct
- Model creator: Microsoft
- License
Recommended Prompt Format (Chat Format)
<|user|>
Provide some context and/or instructions to the model.<|end|>
<|assistant|>
AI message goes here<|end|>
<|user|>
The user’s message goes here<|end|>
<|assistant|>
- Downloads last month
- 9
Hardware compatibility
Log In
to view the estimation
5-bit
16-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
HF Inference deployability: The model has no library tag.