Pablo Carrera's picture

Pablo Carrera

pabloce

AI & ML interests

Hello!

Recent Activity

liked a Space 2 days ago
Qwen/Qwen2.5-Max-Demo
liked a model 2 days ago
m-a-p/YuE-s1-7B-anneal-en-cot
liked a model 3 days ago
deepseek-ai/DeepSeek-R1
View all activity

Organizations

ZeroGPU Explorers's profile picture Replete-AI's profile picture SolidRusT Networks's profile picture Poscye's profile picture NeuroLattice's profile picture Cognitive Computations's profile picture

pabloce's activity

reacted to mitkox's post with šŸš€ 5 days ago
view post
Post
2047
llama.cpp is 26.8% faster than ollama.
I have upgraded both, and using the same settings, I am running the same DeepSeek R1 Distill 1.5B on the same hardware. It's an Apples to Apples comparison.

Total duration:
llama.cpp 6.85 sec <- 26.8% faster
ollama 8.69 sec

Breakdown by phase:
Model loading
llama.cpp 241 ms <- 2x faster
ollama 553 ms

Prompt processing
llama.cpp 416.04 tokens/s with an eval time 45.67 ms <- 10x faster
ollama 42.17 tokens/s with an eval time of 498 ms

Token generation
llama.cpp 137.79 tokens/s with an eval time 6.62 sec <- 13% faster
ollama 122.07 tokens/s with an eval time 7.64 sec

llama.cpp is LLM inference in C/C++; ollama adds abstraction layers and marketing.

Make sure you own your AI. AI in the cloud is not aligned with you; it's aligned with the company that owns it.
Ā·
reacted to cbensimon's post with šŸ¤— about 2 months ago
view post
Post
4418
Hello everybody,

We've rolled out a major update to ZeroGPU! All the Spaces are now running on it.

Major improvements:

1. GPU cold starts about twice as fast!
2. RAM usage reduced by two-thirds, allowing more effective resource usage, meaning more GPUs for the community!
3. ZeroGPU initializations (coldstarts) can now be tracked and displayed (use progress=gr.Progress(track_tqdm=True))
4. Improved compatibility and PyTorch integration, increasing ZeroGPU compatible spaces without requiring any modifications!

Feel free to answer in the post if you have any questions

šŸ¤— Best regards,
Charles