Matricardi Fabio

FM-1976

AI & ML interests

control system engineering, AI, LLM with python. ThePoorGPUguy on substack

Recent Activity

Organizations

None yet

FM-1976's activity

reacted to Xenova's post with πŸ”₯ 2 days ago
view post
Post
2003
NEW: Real-time conversational AI models can now run 100% locally in your browser! 🀯

πŸ” Privacy by design (no data leaves your device)
πŸ’° Completely free... forever
πŸ“¦ Zero installation required, just visit a website
⚑️ Blazingly-fast WebGPU-accelerated inference

Try it out: webml-community/conversational-webgpu

For those interested, here's how it works:
- Silero VAD for voice activity detection
- Whisper for speech recognition
- SmolLM2-1.7B for text generation
- Kokoro for text to speech

Powered by Transformers.js and ONNX Runtime Web! πŸ€— I hope you like it!
  • 2 replies
Β·
reacted to yeonseok-zeticai's post with πŸ”₯ 3 days ago
view post
Post
1992
πŸš€ NEW DROP: run your own on-device LLMβ€”in minutes, on any phone
Today we’re open-sourcing everything you need to put Qwen3-0.6B straight into a production-ready mobile app:

πŸŽ₯ Watch Qwen3-0.6B chat in real time on any smartphones!

πŸ“Š TPS benchmarks – slides comparing token-per-second across heterogeneous mobile devices

πŸ’» Plug-and-play source – Just Copy & Run the source to your project for Android (Kotlin & Java) and iOS (Swift).

🀞 Cross-platform, one pipeline – ZETIC.MLange auto-tunes kernels for every different devices, we’ve tested.

πŸ‘¨β€πŸ’» Ready for production – swap in your own model, re-benchmark with one command, publish.

Get started
Just Sign-up and check the playground project, QWEN-0.6B
- https://mlange.zetic.ai/p/zetic-example/Qwen3-0.6B

We built this to show that cloud-free LLMs are ready today. Dive in, fork it, and tag ZETIC.ai when you launch your own on-device assistant, game NPC, or offline content generatorβ€”we’ll spotlight the best projects.