license: cc-by-nc-4.0 license_name: cc-by-nc-4.0
π§ MythoMax-L2-13B - GGUF FP16 (Unquantized)
This is a GGUF-converted, float16 version of Gryphe's MythoMax-L2-13B, designed for local inference with full quality on high-VRAM GPUs.
ποΈ Converted & shared by: Sandra Weidmann
π οΈ Tested with: RTX 3090, text-generation-webui
+ llama.cpp
π Original Model: Gryphe/MythoMax-L2-13B
β¨ Why this model?
This model was converted to preserve full precision (float16) for use in:
- π§ fine-tuned instruction tasks
- π roleplay and creative writing
- π¬ emotionally nuanced dialogue
- π§ͺ experimentation with full-context outputs (4096+ tokens)
π¦ Model Details
Property | Value |
---|---|
Format | GGUF |
Precision | float16 (f16) |
Context Size | 4096 |
Tensor Count | 363 |
File Size | ~26.0β―GB |
Original Format | Transformers (.bin ) |
Converted Using | convert_hf_to_gguf.py |
π§° Usage (with llama.cpp
)
./main -m mythomax-l2-13b-f16.gguf -c 4096 -n 512 --color
Or via text-generation-webui:
Backend: llama.cpp
Load model: mythomax-l2-13b-f16.gguf
Set context: 4096+
π Notes
This GGUF build is shared for non-commercial, experimental, and educational use.
Full credit to the original model author Gryphe.
If this version helped you, consider giving it a β and sharing feedback.
Sandra β¨
py-sandy
https://samedia.app/dev
- Downloads last month
- 414
Hardware compatibility
Log In
to view the estimation
16-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for py-sandy/MythoMax-L2-13B-GGUF-FP16
Base model
Gryphe/MythoMax-L2-13b