The Collectionists
AI & ML interests
None defined yet.
the-collectionists's activity
Post
2024
PSA for anyone using
Both of these themes have been updated to fix some of the long-standing inconsistencies ever since the transition to Gradio v5. Textboxes are no longer bright green and
If your space is already using one of these themes, you just need to restart your space to get the latest version. No code changes needed.
Nymbo/Nymbo_Theme
or Nymbo/Nymbo_Theme_5
in a Gradio space ~Both of these themes have been updated to fix some of the long-standing inconsistencies ever since the transition to Gradio v5. Textboxes are no longer bright green and
in-line code
is readable now! Both themes are now visually identical across versions.If your space is already using one of these themes, you just need to restart your space to get the latest version. No code changes needed.
Post
4636
Post
13131
I was thinking i need to step up my game on training Flux LoRas models, time to have some fun ! ☀️
Expect a new drop per week on aesthetics that catched my attention, here are 3 of them that worked really well !
fffiloni/cute-comic-800
fffiloni/carbo-800
fffiloni/oniric-750
Expect a new drop per week on aesthetics that catched my attention, here are 3 of them that worked really well !
fffiloni/cute-comic-800
fffiloni/carbo-800
fffiloni/oniric-750
Post
6134
Hey everyone, we've given https://hf.co/spaces page a fresh update!
Smart Search: Now just type what you want to do—like "make a viral meme" or "generate music"—and our search gets it.
New Categories: Check out the cool new filter bar with icons to help you pick a category fast.
Redesigned Space Cards: Reworked a bit to really show off the app descriptions, so you know what each Space does at a glance.
Random Prompt: Need ideas? Hit the dice button for a burst of inspiration.
We’d love to hear what you think—drop us some feedback plz!
Smart Search: Now just type what you want to do—like "make a viral meme" or "generate music"—and our search gets it.
New Categories: Check out the cool new filter bar with icons to help you pick a category fast.
Redesigned Space Cards: Reworked a bit to really show off the app descriptions, so you know what each Space does at a glance.
Random Prompt: Need ideas? Hit the dice button for a burst of inspiration.
We’d love to hear what you think—drop us some feedback plz!
Post
3570
Explain like i'm 5 the last take from
@thomwolf
on X about Dario's essay on DeepSeek:
—› Open-source AI is like a big cookbook that everyone can read and improve. Instead of a few chefs keeping their recipes secret, anyone can cook, test, and invent new things.
If only one company controls AI, everything stops if they have a problem—like when the internet goes down. With open-source, many people can help, making sure it keeps running smoothly.
AI isn’t just a race between two countries; it’s a team effort around the world. By sharing, we move faster and create safer technology for everyone.
—
🤗
—› Open-source AI is like a big cookbook that everyone can read and improve. Instead of a few chefs keeping their recipes secret, anyone can cook, test, and invent new things.
If only one company controls AI, everything stops if they have a problem—like when the internet goes down. With open-source, many people can help, making sure it keeps running smoothly.
AI isn’t just a race between two countries; it’s a team effort around the world. By sharing, we move faster and create safer technology for everyone.
—
🤗
Post
3221
Finally, an open-source AI that turns your lyrics into full songs is here—meet YuE! Unlike other tools that only create short clips, YuE can make entire songs (up to 5 minutes) with vocals, melody, and instruments all working together. Letsss go!
m-a-p/YuE-s1-7B-anneal-en-cot
m-a-p/YuE-s1-7B-anneal-en-cot

jbilcke-hf
posted
an
update
6 months ago
Post
15343
Doing some testing with HunyuanVideo on the Hugging Face Inference Endpoints 🤗
prompt: "a Shiba Inu is acting as a DJ, he wears sunglasses and is mixing and scratching with vinyl discs at a Ibiza sunny sand beach party"
1280x720, 22 steps, 121 frames
There are still some things to iron out regarding speed and memory usage, right now it takes 20min on an A100 (see attached charts)
but you can check it out here:
https://huggingface.co/jbilcke-hf/HunyuanVideo-for-InferenceEndpoints
There are various things I want to try like the 100% diffusers version and other models (LTX-Video..)
prompt: "a Shiba Inu is acting as a DJ, he wears sunglasses and is mixing and scratching with vinyl discs at a Ibiza sunny sand beach party"
1280x720, 22 steps, 121 frames
There are still some things to iron out regarding speed and memory usage, right now it takes 20min on an A100 (see attached charts)
but you can check it out here:
https://huggingface.co/jbilcke-hf/HunyuanVideo-for-InferenceEndpoints
There are various things I want to try like the 100% diffusers version and other models (LTX-Video..)
Post
2229
Qwen/QwQ-32B-Preview shows us the future (and it's going to be exciting)...
I tested it against some really challenging reasoning prompts and the results are amazing 🤯.
Check this dataset for the results: victor/qwq-misguided-attention
I tested it against some really challenging reasoning prompts and the results are amazing 🤯.
Check this dataset for the results: victor/qwq-misguided-attention
Post
2615
Perfect example of why
Qwen/Qwen2.5-Coder-32B-Instruct is insane?
Introducing: AI Video Composer 🔥
huggingface-projects/ai-video-composer
Drag and drop your assets (images/videos/audios) to create any video you want using natural language!
It works by asking the model to output a valid FFMPEG and this can be quite complex but most of the time Qwen2.5-Coder-32B gets it right (that thing is a beast). It's an update of an old project made with GPT4 and it was almost impossible to make it work with open models back then (~1.5 years ago), but not anymore, let's go open weights 🚀.
Introducing: AI Video Composer 🔥
huggingface-projects/ai-video-composer
Drag and drop your assets (images/videos/audios) to create any video you want using natural language!
It works by asking the model to output a valid FFMPEG and this can be quite complex but most of the time Qwen2.5-Coder-32B gets it right (that thing is a beast). It's an update of an old project made with GPT4 and it was almost impossible to make it work with open models back then (~1.5 years ago), but not anymore, let's go open weights 🚀.
Post
1863
Qwen2.5-72B is now the default HuggingChat model.
This model is so good that you must try it! I often get better results on rephrasing with it than Sonnet or GPT-4!!
This model is so good that you must try it! I often get better results on rephrasing with it than Sonnet or GPT-4!!
Post
13990
DimensionX is out for you to try and duplicate 🤗
—> fffiloni/DimensionX
Discuss Paper: DimensionX: Create Any 3D and 4D Scenes from a Single Image with Controllable Video Diffusion (2411.04928)
Examples by the amazing William Lamkin @phanes
—> fffiloni/DimensionX
Discuss Paper: DimensionX: Create Any 3D and 4D Scenes from a Single Image with Controllable Video Diffusion (2411.04928)
Examples by the amazing William Lamkin @phanes
Post
1683
nvidia/Llama-3.1-Nemotron-70B-Instruct-HF is now available in HuggingChat if you want to try it!
https://huggingface.co/chat/models/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
https://huggingface.co/chat/models/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
Post
2694
NEW - Inference Playground
Maybe like me you have always wanted a super easy way to compare llama3.2-1B vs. llama3.2-3B? or the same model with different temperatures?
Trying and comparing warm Inference API models has never been easier!
Just go to https://hf.co/playground, set your token and you're ready to go.
We'll keep improving, feedback welcome 😊
Maybe like me you have always wanted a super easy way to compare llama3.2-1B vs. llama3.2-3B? or the same model with different temperatures?
Trying and comparing warm Inference API models has never been easier!
Just go to https://hf.co/playground, set your token and you're ready to go.
We'll keep improving, feedback welcome 😊
Post
10908
Realtime Whisper Large v3 Turbo Demo:
It transcribes audio in about 0.3 seconds.
KingNish/Realtime-whisper-large-v3-turbo
It transcribes audio in about 0.3 seconds.
KingNish/Realtime-whisper-large-v3-turbo
Post
20025
Visionary Walter Murch (editor for Francis Ford Coppola), in 1999:
“ So let's suppose a technical apotheosis some time in the middle of the 21st century, when it somehow becomes possible for one person to make an entire feature film, with virtual actors. Would this be a good thing?
If the history of oil painting is any guide, the broadest answer would be yes, with the obvious caution to keep a wary eye on the destabilizing effect of following too intently a hermetically personal vision. One need only look at the unraveling of painting or classical music in the 20th century to see the risks.
Let's go even further, and force the issue to its ultimate conclusion by supposing the diabolical invention of a black box that could directly convert a single person's thoughts into a viewable cinematic reality. You would attach a series of electrodes to various points on your skull and simply think the film into existence.
And since we are time-traveling, let us present this hypothetical invention as a Faustian bargain to the future filmmakers of the 21st century. If this box were offered by some mysterious cloaked figure in exchange for your eternal soul, would you take it?
The kind of filmmakers who would accept, even leap, at the offer are driven by the desire to see their own vision on screen in as pure a form as possible. They accept present levels of collaboration as the evil necessary to achieve this vision. Alfred Hitchcock, I imagine, would be one of them, judging from his description of the creative process: "The film is already made in my head before we start shooting."”
—
Read "A Digital Cinema of the Mind? Could Be" by Walter Murch: https://archive.nytimes.com/www.nytimes.com/library/film/050299future-film.html
“ So let's suppose a technical apotheosis some time in the middle of the 21st century, when it somehow becomes possible for one person to make an entire feature film, with virtual actors. Would this be a good thing?
If the history of oil painting is any guide, the broadest answer would be yes, with the obvious caution to keep a wary eye on the destabilizing effect of following too intently a hermetically personal vision. One need only look at the unraveling of painting or classical music in the 20th century to see the risks.
Let's go even further, and force the issue to its ultimate conclusion by supposing the diabolical invention of a black box that could directly convert a single person's thoughts into a viewable cinematic reality. You would attach a series of electrodes to various points on your skull and simply think the film into existence.
And since we are time-traveling, let us present this hypothetical invention as a Faustian bargain to the future filmmakers of the 21st century. If this box were offered by some mysterious cloaked figure in exchange for your eternal soul, would you take it?
The kind of filmmakers who would accept, even leap, at the offer are driven by the desire to see their own vision on screen in as pure a form as possible. They accept present levels of collaboration as the evil necessary to achieve this vision. Alfred Hitchcock, I imagine, would be one of them, judging from his description of the creative process: "The film is already made in my head before we start shooting."”
—
Read "A Digital Cinema of the Mind? Could Be" by Walter Murch: https://archive.nytimes.com/www.nytimes.com/library/film/050299future-film.html
Post
8277
Exciting news! Introducing super-fast AI video assistant, currently in beta. With a minimum latency of under 500ms and an average latency of just 600ms.
DEMO LINK:
KingNish/Live-Video-Chat
DEMO LINK:
KingNish/Live-Video-Chat
Post
3673
A super good and fast image inpainting demo is here.
Its' super cool and realistic.
Demo by @OzzyGT (Must try):
OzzyGT/diffusers-fast-inpaint
Its' super cool and realistic.
Demo by @OzzyGT (Must try):
OzzyGT/diffusers-fast-inpaint
Post
3607
Mistral Nemo is better than many models in 1st grader level reasoning.
Post
3931
I am experimenting with Flux and trying to push it to its limits without training (as I am GPU-poor 😅).
I found some flaws in the pipelines, which I resolved, and now I am able to generate an approx similar quality image as Flux Schnell 4 steps in just 1 step.
Demo Link:
KingNish/Realtime-FLUX
I found some flaws in the pipelines, which I resolved, and now I am able to generate an approx similar quality image as Flux Schnell 4 steps in just 1 step.
Demo Link:
KingNish/Realtime-FLUX