Victor Mustar's picture

Victor Mustar PRO

victor

AI & ML interests

Building the UX of this website

Recent Activity

Organizations

Hugging Face's profile picture Google's profile picture Safetensors's profile picture Competitions's profile picture 21 RNN's profile picture Spaces-explorers's profile picture Text Generation Inference's profile picture Spaces Examples's profile picture CVPR Demo Track's profile picture Hugging Chat's profile picture Webhooks Explorers (BETA)'s profile picture lora concepts library's profile picture Huggingface Projects's profile picture Scanned Tokens's profile picture hf admins's profile picture Hugging Face OSS Metrics's profile picture Stable Diffusion Dreambooth Concepts Library's profile picture Core ML Projects's profile picture temp-org's profile picture Blog-explorers's profile picture Mustarz's profile picture Open LLM Leaderboard's profile picture Enterprise Explorers's profile picture The Collectionists's profile picture ZeroGPU Explorers's profile picture Hugging Face Tools's profile picture TstOrg141's profile picture Stable Video benchmark's profile picture Social Post Explorers's profile picture Dev Mode Explorers's profile picture LLHF's profile picture SLLHF's profile picture Self-serve FTW's profile picture Inference Explorers's profile picture

victor's activity

reacted to csabakecskemeti's post with 👍 about 1 hour ago
view post
Post
1569
I'm collecting llama-bench results for inference with a llama 3.1 8B q4 and q8 reference models on varoius GPUs. The results are average of 5 executions.
The system varies (different motherboard and CPU ... but that probably that has little effect on the inference performance).

https://devquasar.com/gpu-gguf-inference-comparison/
the exact models user are in the page

I'd welcome results from other GPUs is you have access do anything else you've need in the post. Hopefully this is useful information everyone.
reacted to onekq's post with 👍 about 1 hour ago
reacted to smirki's post with 👍 about 1 hour ago
view post
Post
434
Introducing a SMALL Reasoning React Model with State!
We did this by introducing a new form of reasoning that aligns with UI principles to do a layer of testing. For example:
"Looking back at all these pieces, we've considered state management, data structures, core functionalities etc"
And it comes in all sizes. Great for agents!
Tesslate/tessa-t1-react-reasoning-model-67e0fb72ca23e04473885c0e
Tesslate/Tessa-T1-14B
https://huggingface.co/smirki/Tessa-T1-14B-Q8_0-GGUF
reacted to MikeDoes's post with 🔥 about 1 hour ago
New activity in deepseek-ai/DeepSeek-V3 about 3 hours ago

Point to latest checkpoint

#96 opened about 3 hours ago by
victor
updated a Space about 4 hours ago
published a Space about 4 hours ago