license: cc-by-nc-4.0 license_name: cc-by-nc-4.0

🧠 MythoMax-L2-13B - GGUF FP16 (Unquantized)

This is a GGUF-converted, float16 version of Gryphe's MythoMax-L2-13B, designed for local inference with full quality on high-VRAM GPUs.

πŸŽ™οΈ Converted & shared by: Sandra Weidmann
πŸ› οΈ Tested with: RTX 3090, text-generation-webui + llama.cpp
πŸ”— Original Model: Gryphe/MythoMax-L2-13B


✨ Why this model?

This model was converted to preserve full precision (float16) for use in:

  • 🧠 fine-tuned instruction tasks
  • 🎭 roleplay and creative writing
  • πŸ’¬ emotionally nuanced dialogue
  • πŸ§ͺ experimentation with full-context outputs (4096+ tokens)

πŸ“¦ Model Details

Property Value
Format GGUF
Precision float16 (f16)
Context Size 4096
Tensor Count 363
File Size ~26.0β€―GB
Original Format Transformers (.bin)
Converted Using convert_hf_to_gguf.py

🧰 Usage (with llama.cpp)

./main -m mythomax-l2-13b-f16.gguf -c 4096 -n 512 --color
Or via text-generation-webui:

Backend: llama.cpp

Load model: mythomax-l2-13b-f16.gguf

Set context: 4096+

πŸ’™ Notes
This GGUF build is shared for non-commercial, experimental, and educational use.
Full credit to the original model author Gryphe.
If this version helped you, consider giving it a ⭐ and sharing feedback.

Sandra ✨
py-sandy
https://samedia.app/dev
Downloads last month
414
GGUF
Model size
13B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for py-sandy/MythoMax-L2-13B-GGUF-FP16

Quantized
(14)
this model