Collections

Discover the best community collections!

Collections including paper arxiv:2507.04886
Nemo models [pretrain]
Proof-of-concept: SOTA tokenizers can be used for compatible precomputed embeddings, industry can repeat with their tokenizers
Max models [pretrain]
Multilingual language model collection with frozen, unified Unicode-based embeddings. Includes Russian, Chinese, and their MoE fusion.