Skywork-R1V2 Collection Multimodal Hybrid Reinforcement Learning for Reasoning • 4 items • Updated 11 days ago • 10
Unsloth Dynamic 2.0 Quants Collection New 2.0 version of our Dynamic GGUF + Quants. Dynamic 2.0 achieves superior accuracy & outperforms all leading quantization methods. • 29 items • Updated 9 days ago • 85
xLAM-2 Collection A family of Large Action Model for multi-turn conversation and tool-use • 10 items • Updated 5 days ago • 13
Granite 3.3 Language Models Collection Our latest language models licensed under Apache 2.0 license. • 4 items • Updated 8 days ago • 33
PRIMA.CPP: Speeding Up 70B-Scale LLM Inference on Low-Resource Everyday Home Clusters Paper • 2504.08791 • Published Apr 7 • 130
Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention Paper • 2502.11089 • Published Feb 16 • 157
DeepSeek-R1 Thoughtology: Let's <think> about LLM Reasoning Paper • 2504.07128 • Published Apr 2 • 83
Apriel Collection ServiceNow Language Modeling Lab's first model family series • 3 items • Updated 3 days ago • 9
Kimi-VL-A3B Collection Moonshot's efficient MoE VLMs, exceptional on agent, long-context, and thinking • 6 items • Updated 28 days ago • 66
ReSearch Collection Trained models as described in the paper "ReSearch: Learning to Reason with Search for LLMs via Reinforcement Learning" • 5 items • Updated Mar 27 • 5
CoRNStack Collection State-of-the-art code retrieval and re-ranking models and datasets • 9 items • Updated Mar 26 • 17
VisualWebInstruct: Scaling up Multimodal Instruction Data through Web Search Paper • 2503.10582 • Published Mar 13 • 23
Cohere Labs Aya Vision Collection Aya Vision is a state-of-the-art family of vision models that brings multimodal capabilities to 23 languages. • 5 items • Updated 24 days ago • 68
Llama Nemotron Collection Open, Production-ready Enterprise Models • 6 items • Updated about 6 hours ago • 51