When you encounter "RuntimeError: cuDNN Frontend error: [cudnn_frontend] Error: No execution plans support the graph."
We can use workaround like this:
torch.backends.cuda.enable_cudnn_sdp(False)
but this slow downs the performance gain from PyTorch 2.5.
Although it is fixed(not "fixed" but default option is turn-off the cuDNN SDPA) at here -- https://github.com/pytorch/pytorch/pull/138587 , but not released yet. (you need to install directly from source)
Check this out: I trained an AI on huggingface posts! all of these are AI generated: ---------- Hello!
I'm excited to share that my colleague @felipeebert and I have released the largest Spanish LLM benchmark to date.
We've developed the Spanish LLM Evaluation Benchmark (SLAB), a set of benchmarks designed to evaluate the ability of language models to understand, generate and translate in Spanish.
SLAB includes five different benchmarks: - Sentiment Analysis: evaluate models' ability to detect and describe sentiment in natural language - Fact Checking: evaluate models' ability to detect and refute factual errors in text - Question Answering: evaluate models' ability to answer questions in Spanish - Open-ended Questions: evaluate models' ability to generate coherent responses in Spanish - Translation: evaluate models' ability to translate in Spanish
SLAB is aligned with the latest Spanish LLM industry developments and includes the most recent models available on the market. We aim to keep our benchmarks up-to-date and relevant to the Spanish language ecosystem.
A new family of models that brings the power of transformer AI to the masses.
This model is designed to be accessible and easy to use, while still offering high-quality results.
Key features: - Small model size: only 23M parameters - Supports text generation, image generation, and text-to-image tasks - Data-efficient training with a lightweight tokenizer - Optimized for efficient on-device usage - Uses the powerful transformer architecture to deliver high-quality results
zamal/Molmo-4bit Thrilled to announce that the Molmo 7B 4-bit Space is now live! ๐ The model size has been reduced by six times with almost no performance loss, and the results will leave you amazed!
It runs on zero GPU, making it incredibly accessible for everyone!
I'm excited to share my latest project that combines my passion for deep learning and racing cars. I recently created a simple method to predict Formula 1 lap times using machine learning . This is the first solution of its kind in the open-source community, and I'm thrilled to present it to you all.
๐๏ธ The project leverages historical telemetry data to predict lap times, providing a new tool for race strategy and performance analysis. You can check out the notebook on Kaggle here https://www.kaggle.com/code/lucasdraichi/hamilton-lap-time-prediction and see the detailed breakdown of the model and its predictions.
I invite you all to take a look at the lap time predictor, provide feedback, and join the discussion. Your insights and participation would be invaluable as we continue to develop and enhance these tools.
Let's push the boundaries of what's possible with AI in motorsports together!