Mixture of Quantized Experts (MoQE): Complementary Effect of Low-bit Quantization and Robustness Paper • 2310.02410 • Published Oct 3, 2023 • 3
Llama 4 Collection Meta's new Llama 4 multimodal models, Scout & Maverick. Includes Dynamic GGUFs, 16-bit & Dynamic 4-bit uploads. Run & fine-tune them with Unsloth! • 15 items • Updated 5 days ago • 44
Running 2.55k 2.55k The Ultra-Scale Playbook 🌌 The ultimate guide to training LLM on large GPU Clusters