Gemini 2.5 Pro, thinking by default! We excited launch our best Gemini model for reasoning, multimodal and coding yet! #1 on LMSYS, Humanity’s Last Exam, AIME and GPQA and more!
TL;DR: - 💻 Best Gemini coding model yet, particularly for web development (excels on LiveCodeBench). - 🧠 Default "Thinking" with up to 64k token output - 🌌 1 Million multimodal input context for text, image, video, audio, and pdf - 🛠️ Function calling, structured output, google search & code execution. - 🏆 #1 on LMArena & sota on AIME, GPQA, Humanity's Last Exam - 💡 Knowledge cut of January 2025 - 🤗 Available for free as Experimental in AI Studio, Gemini API & Gemini APP - 🚀 Rate limits - Free 2 RPM 50 req/day
For Inference Providers who have built support for our Billing API (currently: Fal, Novita, HF-Inference – with more coming soon), we've started enabling Pay as you go (=PAYG)
What this means is that you can use those Inference Providers beyond the free included credits, and they're charged to your HF account.
You can see it on this view: any provider that does not have a "Billing disabled" badge, is PAYG-compatible.
After some heated discussion 🔥, we clarify our intent re. storage limits on the Hub
TL;DR: - public storage is free, and (unless blatant abuse) unlimited. We do ask that you consider upgrading to PRO and/or Enterprise Hub if possible - private storage is paid above a significant free tier (1TB if you have a paid account, 100GB otherwise)
INTELLECT-1 is the first collaboratively trained 10 billion parameter language model trained from scratch on 1 trillion tokens of English text and code.
Misc models: 🦖T-Rex2, a very powerful object detection model for many applications https://github.com/IDEA-Research/T-Rex 👀 CT-RATE : A 3D dataset paired with text reports ibrahimhamamci/CT-RATE 🐙Octopus v2: a Gemma-based model trained for Android API - extremely fast, better than Llama+RAG, great results NexaAIDev/Octopus-v2
New state-of-the-art open LLM! 🚀 Databricks just released DBRX, a 132B MoE trained on 12T tokens. Claiming to surpass OpenAI GPT-3.5 and is competitive with Google Gemini 1.0 Pro. 🤯
TL;DR 🧮 132B MoE with 16 experts with 4 active in generation 🪟 32 000 context window 📈 Outperforms open LLMs on common benchmarks, including MMLU 🚀 Up to 2x faster inference than Llama 2 70B 💻 Trained on 12T tokens 🔡 Uses the GPT-4 tokenizer 📜 Custom License, commercially useable
🌏Models and datasets around the world - Tess-70B, a MiQu-70B fine-tune with high-quality data migtissera/Tess-70B-v1.6 - UNI, a model trained on 100 million pathology images from 100k+ slides MahmoodLab/UNI - CONCH, a VLM trained on 1.17 million pathology image-text pairs MahmoodLab/CONCH