Fidite Nemini's picture

Fidite Nemini PRO

FiditeNemini

AI & ML interests

Prompt engineering, unalignment, MLX, model merging, diffusion models

Recent Activity

Organizations

Fidite Nemini Open Source's profile picture MLX Community's profile picture Cognitive Computations's profile picture

FiditeNemini's activity

reacted to bartowski's post with πŸ‘ 29 days ago
view post
Post
23694
Access requests enabled for latest GLM models

While a fix is being implemented (https://github.com/ggml-org/llama.cpp/pull/12957) I want to leave the models up for visibility and continued discussion, but want to prevent accidental downloads of known broken models (even though there are settings that could fix it at runtime for now)

With this goal, I've enabled access requests. I don't really want your data, so I'm sorry that I don't think there's a way around that? But that's what I'm gonna do for now, and I'll remove the gate when a fix is up and verified and I have a chance to re-convert and quantize!

Hope you don't mind in the mean time :D
  • 1 reply
Β·
reacted to merterbak's post with πŸ”₯ about 1 month ago
view post
Post
3016
Meta has unveiled its Llama 4 πŸ¦™ family of models, featuring native multimodality and mixture-of-experts architecture. Two model families are available now:
ModelsπŸ€—: meta-llama/llama-4-67f0c30d9fe03840bc9d0164
Blog Post: https://ai.meta.com/blog/llama-4-multimodal-intelligence/
HF's Blog Post: https://huggingface.co/blog/llama4-release

- 🧠 Native Multimodality - Process text and images in a unified architecture
- πŸ” Mixture-of-Experts - First Llama models using MoE for incredible efficiency
- πŸ“ Super Long Context - Up to 10M tokens
- 🌐 Multilingual Power - Trained on 200 languages with 10x more multilingual tokens than Llama 3 (including over 100 languages with over 1 billion tokens each)

πŸ”Ή Llama 4 Scout
- 17B active parameters (109B total)
- 16 experts architecture
- 10M context window
- Fits on a single H100 GPU
- Beats Gemma 3, Gemini 2.0 Flash-Lite, and Mistral 3.1

πŸ”Ή Llama 4 Maverick
- 17B active parameters (400B total)
- 128 experts architecture
- It can fit perfectly on DGX H100(8x H100)
- 1M context window
- Outperforms GPT-4o and Gemini 2.0 Flash
- ELO score of 1417 on LMArena currently second best model on arena

πŸ”Ή Llama 4 Behemoth (Coming Soon)
- 288B active parameters (2T total)
- 16 experts architecture
- Teacher model for Scout and Maverick
- Outperforms GPT-4.5, Claude Sonnet 3.7, and Gemini 2.0 Pro on STEM benchmarks
reacted to clem's post with πŸ”₯ about 1 month ago
view post
Post
2002
Llama models (arguably the most successful open AI models of all times) just represented 3% of total model downloads on Hugging Face in March.

People and media like stories of winner takes all & one model/company to rule them all but the reality is much more nuanced than this!

Kudos to all the small AI builders out there!
  • 2 replies
Β·
reacted to mlabonne's post with πŸ”₯ about 2 months ago
reacted to AdinaY's post with πŸ‘€ about 2 months ago
view post
Post
2105
Skywork-R1VπŸš€ 38B open multimodal reasoning model with advanced visual CoT capabilities, released by Skywork.

Skywork/Skywork-R1V-38B

✨ Visual Reasoning: Breaks down complex images step by step.
✨ Math & Science: Solves visual problems with high precision.
✨ Combines text & images for deeper understanding.

New activity in TheDrummer/Fallen-Llama-3.3-R1-70B-v1-GGUF 2 months ago

Wrong gguf's in repo?

3
#1 opened 3 months ago by
FiditeNemini
reacted to samihalawa's post with πŸ‘ 3 months ago
view post
Post
1782
βœ…BEST DEBUG PROMPT
Language: Any. πŸŒ€Project Type: Any

What prompt if sent to you  will make you  detect and fix all the code crashing  issues in the COMPLETE codebase so I don't have to ask you to fix again and again ?
Step 1. Gimme such prompt
Step 2 Follow it yourself quietly and COMPLETELY
Step3: State  that if you are asked again about finding fatal bugs and logic issues and inconsistencies in the current codebase you would not be able  to  find more. (You cannot lie, so you must make all the code adjustments necessary prior to  such statement).