Mads Kuhlmann-Joergensen

maalber

AI & ML interests

None yet

Recent Activity

Organizations

Blog-explorers's profile picture Rapidata's profile picture

maalber's activity

reacted to jasoncorkill's post with 👀 15 days ago
view post
Post
3592
Benchmark Update: @google Veo3 (Text-to-Video)

Two months ago, we benchmarked @google ’s Veo2 model. It fell short, struggling with style consistency and temporal coherence, trailing behind Runway, Pika, @tencent , and even @alibaba-pai .

That’s changed.

We just wrapped up benchmarking Veo3, and the improvements are substantial. It outperformed every other model by a wide margin across all key metrics. Not just better, dominating across style, coherence, and prompt adherence. It's rare to see such a clear lead in today’s hyper-competitive T2V landscape.

Dataset coming soon. Stay tuned.
·
reacted to jasoncorkill's post with 🔥 27 days ago
view post
Post
2868
🔥 Hidream I1 is online! 🔥

We just added Hidream I1 to our T2I leaderboard (https://www.rapidata.ai/leaderboard/image-models) benchmarked using 195k+ human responses from 38k+ annotators, all collected in under 24 hours.

It landed #3 overall, right behind:
- @openai 4o
- @black-forest-labs Flux 1 Pro
...and just ahead of @black-forest-labs Flux 1.1 Pro, @xai-org Aurora and @google Imagen3.

Want to dig into the data? Check out our dataset here:
Rapidata/Hidream_t2i_human_preference

What model should we benchmark next?
replied to jasoncorkill's post about 2 months ago
view reply

Thank you! If you also like the dataset itself that will make a big impact!

reacted to jasoncorkill's post with 🚀🔥 about 2 months ago
view post
Post
3283
🚀 We tried something new!

We just published a dataset using a new (for us) preference modality: direct ranking based on aesthetic preference. We ranked a couple of thousand images from most to least preferred, all sampled from the Open Image Preferences v1 dataset by the amazing @data-is-better-together team.

📊 Check it out here:
Rapidata/2k-ranked-images-open-image-preferences-v1

We're really curious to hear your thoughts!
Is this kind of ranking interesting or useful to you? Let us know! 💬

If it is, please consider leaving a ❤️ and if we hit 30 ❤️s, we’ll go ahead and rank the full 17k image dataset!
·
reacted to jasoncorkill's post with 🔥 2 months ago
view post
Post
3066
🔥 Yesterday was a fire day!
We dropped two brand-new datasets capturing Human Preferences for text-to-video and text-to-image generations powered by our own crowdsourcing tool!

Whether you're working on model evaluation, alignment, or fine-tuning, this is for you.

1. Text-to-Video Dataset (Pika 2.2 model):
Rapidata/text-2-video-human-preferences-pika2.2

2. Text-to-Image Dataset (Reve-AI Halfmoon):
Rapidata/Reve-AI-Halfmoon_t2i_human_preference

Let’s train AI on AI-generated content with humans in the loop.
Let’s make generative models that actually get us.
reacted to jasoncorkill's post with 🚀🧠❤️ 2 months ago
view post
Post
2383
🚀 Rapidata: Setting the Standard for Model Evaluation

Rapidata is proud to announce our first independent appearance in academic research, featured in the Lumina-Image 2.0 paper. This marks the beginning of our journey to become the standard for testing text-to-image and generative models. Our expertise in large-scale human annotations allows researchers to refine their models with accurate, real-world feedback.

As we continue to establish ourselves as a key player in model evaluation, we’re here to support researchers with high-quality annotations at scale. Reach out to [email protected] to see how we can help.

Lumina-Image 2.0: A Unified and Efficient Image Generative Framework (2503.21758)