AI & ML interests
None defined yet.
Post
2296
Open Source Avengers, Assemble! Ask an expert AI agent team to solve complex problems together ๐ฅ
Consilium brings together multiple agents that debate and use live research (web, arXiv, SEC) to reach a consensus. You set the strategy, they find the answer.
Credit to @azettl for this awesome demo: Agents-MCP-Hackathon/consilium_mcp
Consilium brings together multiple agents that debate and use live research (web, arXiv, SEC) to reach a consensus. You set the strategy, they find the answer.
Credit to @azettl for this awesome demo: Agents-MCP-Hackathon/consilium_mcp

jbilcke-hfย
posted
an
update
21 days ago
Post
2620
Did you know that there is a UI wrapper around https://github.com/a-r-r-o-w/finetrainers which is a great library made by
@a-r-r-o-w
for finetuning AI video models?
The UI is called VideoModelStudio (or VMS in casual chat)
All you have to do is to duplicate this space:
jbilcke-hf/VideoModelStudio
The UI is called VideoModelStudio (or VMS in casual chat)
All you have to do is to duplicate this space:
jbilcke-hf/VideoModelStudio

jbilcke-hfย
posted
an
update
21 days ago
Post
1816
Hi everyone,
I've seen some unsuccessful attempts at running Wan2GP inside a Hugging Face Space, which is a shame as it is a great Gradio app!
So here is a fork that you can use, with some instructions on how to do this:
jbilcke-hf/Wan2GP_you_must_clone_this_space_to_use_it#1
Note : some things like persistent models/storage/custom LoRAs might not be fully working out of the box. If you need those, you might have to dig into the Wan2GP codebase, see how to tweak the storage folder. Happy hacking!
I've seen some unsuccessful attempts at running Wan2GP inside a Hugging Face Space, which is a shame as it is a great Gradio app!
So here is a fork that you can use, with some instructions on how to do this:
jbilcke-hf/Wan2GP_you_must_clone_this_space_to_use_it#1
Note : some things like persistent models/storage/custom LoRAs might not be fully working out of the box. If you need those, you might have to dig into the Wan2GP codebase, see how to tweak the storage folder. Happy hacking!
Post
2631
PSA for anyone using
Both of these themes have been updated to fix some of the long-standing inconsistencies ever since the transition to Gradio v5. Textboxes are no longer bright green and
If your space is already using one of these themes, you just need to restart your space to get the latest version. No code changes needed.
Nymbo/Nymbo_Theme
or Nymbo/Nymbo_Theme_5
in a Gradio space ~Both of these themes have been updated to fix some of the long-standing inconsistencies ever since the transition to Gradio v5. Textboxes are no longer bright green and
in-line code
is readable now! Both themes are now visually identical across versions.If your space is already using one of these themes, you just need to restart your space to get the latest version. No code changes needed.
Post
4867
Post
15044
I was thinking i need to step up my game on training Flux LoRas models, time to have some fun ! โ๏ธ
Expect a new drop per week on aesthetics that catched my attention, here are 3 of them that worked really well !
fffiloni/cute-comic-800
fffiloni/carbo-800
fffiloni/oniric-750
Expect a new drop per week on aesthetics that catched my attention, here are 3 of them that worked really well !
fffiloni/cute-comic-800
fffiloni/carbo-800
fffiloni/oniric-750
Post
6183
Hey everyone, we've given https://hf.co/spaces page a fresh update!
Smart Search: Now just type what you want to doโlike "make a viral meme" or "generate music"โand our search gets it.
New Categories: Check out the cool new filter bar with icons to help you pick a category fast.
Redesigned Space Cards: Reworked a bit to really show off the app descriptions, so you know what each Space does at a glance.
Random Prompt: Need ideas? Hit the dice button for a burst of inspiration.
Weโd love to hear what you thinkโdrop us some feedback plz!
Smart Search: Now just type what you want to doโlike "make a viral meme" or "generate music"โand our search gets it.
New Categories: Check out the cool new filter bar with icons to help you pick a category fast.
Redesigned Space Cards: Reworked a bit to really show off the app descriptions, so you know what each Space does at a glance.
Random Prompt: Need ideas? Hit the dice button for a burst of inspiration.
Weโd love to hear what you thinkโdrop us some feedback plz!
Post
3581
Explain like i'm 5 the last take from
@thomwolf
on X about Dario's essay on DeepSeek:
โโบ Open-source AI is like a big cookbook that everyone can read and improve. Instead of a few chefs keeping their recipes secret, anyone can cook, test, and invent new things.
If only one company controls AI, everything stops if they have a problemโlike when the internet goes down. With open-source, many people can help, making sure it keeps running smoothly.
AI isnโt just a race between two countries; itโs a team effort around the world. By sharing, we move faster and create safer technology for everyone.
โ
๐ค
โโบ Open-source AI is like a big cookbook that everyone can read and improve. Instead of a few chefs keeping their recipes secret, anyone can cook, test, and invent new things.
If only one company controls AI, everything stops if they have a problemโlike when the internet goes down. With open-source, many people can help, making sure it keeps running smoothly.
AI isnโt just a race between two countries; itโs a team effort around the world. By sharing, we move faster and create safer technology for everyone.
โ
๐ค
Post
3268
Finally, an open-source AI that turns your lyrics into full songs is hereโmeet YuE! Unlike other tools that only create short clips, YuE can make entire songs (up to 5 minutes) with vocals, melody, and instruments all working together. Letsss go!
m-a-p/YuE-s1-7B-anneal-en-cot
m-a-p/YuE-s1-7B-anneal-en-cot

jbilcke-hfย
posted
an
update
6 months ago
Post
15410
Doing some testing with HunyuanVideo on the Hugging Face Inference Endpoints ๐ค
prompt: "a Shiba Inu is acting as a DJ, he wears sunglasses and is mixing and scratching with vinyl discs at a Ibiza sunny sand beach party"
1280x720, 22 steps, 121 frames
There are still some things to iron out regarding speed and memory usage, right now it takes 20min on an A100 (see attached charts)
but you can check it out here:
https://huggingface.co/jbilcke-hf/HunyuanVideo-for-InferenceEndpoints
There are various things I want to try like the 100% diffusers version and other models (LTX-Video..)
prompt: "a Shiba Inu is acting as a DJ, he wears sunglasses and is mixing and scratching with vinyl discs at a Ibiza sunny sand beach party"
1280x720, 22 steps, 121 frames
There are still some things to iron out regarding speed and memory usage, right now it takes 20min on an A100 (see attached charts)
but you can check it out here:
https://huggingface.co/jbilcke-hf/HunyuanVideo-for-InferenceEndpoints
There are various things I want to try like the 100% diffusers version and other models (LTX-Video..)
Post
2230
Qwen/QwQ-32B-Preview shows us the future (and it's going to be exciting)...
I tested it against some really challenging reasoning prompts and the results are amazing ๐คฏ.
Check this dataset for the results: victor/qwq-misguided-attention
I tested it against some really challenging reasoning prompts and the results are amazing ๐คฏ.
Check this dataset for the results: victor/qwq-misguided-attention
Post
2615
Perfect example of why
Qwen/Qwen2.5-Coder-32B-Instruct is insane?
Introducing: AI Video Composer ๐ฅ
huggingface-projects/ai-video-composer
Drag and drop your assets (images/videos/audios) to create any video you want using natural language!
It works by asking the model to output a valid FFMPEG and this can be quite complex but most of the time Qwen2.5-Coder-32B gets it right (that thing is a beast). It's an update of an old project made with GPT4 and it was almost impossible to make it work with open models back then (~1.5 years ago), but not anymore, let's go open weights ๐.
Introducing: AI Video Composer ๐ฅ
huggingface-projects/ai-video-composer
Drag and drop your assets (images/videos/audios) to create any video you want using natural language!
It works by asking the model to output a valid FFMPEG and this can be quite complex but most of the time Qwen2.5-Coder-32B gets it right (that thing is a beast). It's an update of an old project made with GPT4 and it was almost impossible to make it work with open models back then (~1.5 years ago), but not anymore, let's go open weights ๐.
Post
1863
Qwen2.5-72B is now the default HuggingChat model.
This model is so good that you must try it! I often get better results on rephrasing with it than Sonnet or GPT-4!!
This model is so good that you must try it! I often get better results on rephrasing with it than Sonnet or GPT-4!!
Post
13996
DimensionX is out for you to try and duplicate ๐ค
โ> fffiloni/DimensionX
Discuss Paper: DimensionX: Create Any 3D and 4D Scenes from a Single Image with Controllable Video Diffusion (2411.04928)
Examples by the amazing William Lamkin @phanes
โ> fffiloni/DimensionX
Discuss Paper: DimensionX: Create Any 3D and 4D Scenes from a Single Image with Controllable Video Diffusion (2411.04928)
Examples by the amazing William Lamkin @phanes
Post
1683
nvidia/Llama-3.1-Nemotron-70B-Instruct-HF is now available in HuggingChat if you want to try it!
https://huggingface.co/chat/models/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
https://huggingface.co/chat/models/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF
Post
2694
NEW - Inference Playground
Maybe like me you have always wanted a super easy way to compare llama3.2-1B vs. llama3.2-3B? or the same model with different temperatures?
Trying and comparing warm Inference API models has never been easier!
Just go to https://hf.co/playground, set your token and you're ready to go.
We'll keep improving, feedback welcome ๐
Maybe like me you have always wanted a super easy way to compare llama3.2-1B vs. llama3.2-3B? or the same model with different temperatures?
Trying and comparing warm Inference API models has never been easier!
Just go to https://hf.co/playground, set your token and you're ready to go.
We'll keep improving, feedback welcome ๐
Post
11024
Realtime Whisper Large v3 Turbo Demo:
It transcribes audio in about 0.3 seconds.
KingNish/Realtime-whisper-large-v3-turbo
It transcribes audio in about 0.3 seconds.
KingNish/Realtime-whisper-large-v3-turbo
Post
20027
Visionary Walter Murch (editor for Francis Ford Coppola), in 1999:
โ So let's suppose a technical apotheosis some time in the middle of the 21st century, when it somehow becomes possible for one person to make an entire feature film, with virtual actors. Would this be a good thing?
If the history of oil painting is any guide, the broadest answer would be yes, with the obvious caution to keep a wary eye on the destabilizing effect of following too intently a hermetically personal vision. One need only look at the unraveling of painting or classical music in the 20th century to see the risks.
Let's go even further, and force the issue to its ultimate conclusion by supposing the diabolical invention of a black box that could directly convert a single person's thoughts into a viewable cinematic reality. You would attach a series of electrodes to various points on your skull and simply think the film into existence.
And since we are time-traveling, let us present this hypothetical invention as a Faustian bargain to the future filmmakers of the 21st century. If this box were offered by some mysterious cloaked figure in exchange for your eternal soul, would you take it?
The kind of filmmakers who would accept, even leap, at the offer are driven by the desire to see their own vision on screen in as pure a form as possible. They accept present levels of collaboration as the evil necessary to achieve this vision. Alfred Hitchcock, I imagine, would be one of them, judging from his description of the creative process: "The film is already made in my head before we start shooting."โ
โ
Read "A Digital Cinema of the Mind? Could Be" by Walter Murch: https://archive.nytimes.com/www.nytimes.com/library/film/050299future-film.html
โ So let's suppose a technical apotheosis some time in the middle of the 21st century, when it somehow becomes possible for one person to make an entire feature film, with virtual actors. Would this be a good thing?
If the history of oil painting is any guide, the broadest answer would be yes, with the obvious caution to keep a wary eye on the destabilizing effect of following too intently a hermetically personal vision. One need only look at the unraveling of painting or classical music in the 20th century to see the risks.
Let's go even further, and force the issue to its ultimate conclusion by supposing the diabolical invention of a black box that could directly convert a single person's thoughts into a viewable cinematic reality. You would attach a series of electrodes to various points on your skull and simply think the film into existence.
And since we are time-traveling, let us present this hypothetical invention as a Faustian bargain to the future filmmakers of the 21st century. If this box were offered by some mysterious cloaked figure in exchange for your eternal soul, would you take it?
The kind of filmmakers who would accept, even leap, at the offer are driven by the desire to see their own vision on screen in as pure a form as possible. They accept present levels of collaboration as the evil necessary to achieve this vision. Alfred Hitchcock, I imagine, would be one of them, judging from his description of the creative process: "The film is already made in my head before we start shooting."โ
โ
Read "A Digital Cinema of the Mind? Could Be" by Walter Murch: https://archive.nytimes.com/www.nytimes.com/library/film/050299future-film.html