Aurélien-Morgan CLAUDON

Aurelien-Morgan

AI & ML interests

None yet

Recent Activity

Organizations

ONNXConfig for all's profile picture Gradio-Blocks-Party's profile picture Keras Dreambooth Event's profile picture Blog-explorers's profile picture OpenLLM France's profile picture huggingPartyParis's profile picture ZeroGPU Explorers's profile picture LocalLLaMA's profile picture Cohere Labs Community's profile picture Open RL Leaderboard's profile picture Chinese LLMs on Hugging Face's profile picture Paris AI Running Club's profile picture cvmistralparis's profile picture Hugging Face Discord Community's profile picture Hugging Face Party @ PyTorch Conference's profile picture Nerdy Face's profile picture retrain-pipelines's profile picture

Aurelien-Morgan's activity

New activity in blog-explorers/README about 15 hours ago
liked a Space 1 day ago
upvoted an article 3 days ago
upvoted an article 11 days ago
view article
Article

Tiny Agents in Python: a MCP-powered agent in ~70 lines of code

By celinah and 3 others
122
reacted to cbensimon's post with 👀 19 days ago
view post
Post
5729
🚀 ZeroGPU medium size is now available as a power-user feature

Nothing too fancy for now—ZeroGPU Spaces still default to large (70GB VRAM)—but this paves the way for:
- 💰 size-based quotas / pricing (medium will offer significantly more usage than large)
- 🦣 the upcoming xlarge size (141GB VRAM)

You can as of now control GPU size via a Space variable. Accepted values:
- auto (future default)
- medium
- large (current default)

The auto mode checks total CUDA tensor size during startup:
- More than 30GB → large
- Otherwise → medium
·
reacted to ordagan's post with ❤️ 19 days ago
view post
Post
2197
Excited to introduce Jamba by AI21
ai21labs/Jamba-v0.1

We are thrilled to announce Jamba, the world’s first production-grade Mamba based model.

Key Features:
- First production-grade Mamba based model built on a novel SSM-Transformer hybrid architecture
- 3X throughput on long contexts compared to Mixtral 8x7B
- Democratizes access to a massive 256K context window
- The only model in its size class that fits up to 140K context on a single GPU

Jamba is based on a novel architecture that combines Mamba and Transformer. While our initial results show great efficiency gains, we expect this to be further explored and improved with the help of the community.

Check out our blog post for more info: https://ai21-labs.webflow.io/blog/announcing-jamba
  • 2 replies
·
posted an update 21 days ago
view reply

Lots of insights and wisdom. Very nice. Articulating these ideas and notions has much value and will pay off. Thank you.

upvoted an article 25 days ago
view article
Article

LeRobot Community Datasets: The “ImageNet” of Robotics — When and How?

By danaaubakirova and 6 others
56