Blog, Articles, and discussions
Community Articles
view all
Community Articles
view allTopic 28: What is Mixture-of-Mamba?
By
and 1 other
•
•
2Mimicking Consciousness in LLMs: Ascending the Dimensions of Thought with Recurrent Processing
By
•
结合Deepseek代码探讨MLA的改进及收益
By
•
•
2大模型偏好优化技术:DPO及其变种
By
•
•
2Grok 3 ai : Best AI model now!
By
•
•
4Argunauts Training Phase II: Selfplay Finetuning Line-By-Line
By
•
•
2Synthetic Face Embeddings: Research Notes and Methodology
By
and 1 other
•
•
1How to use Sentient’s Dobby-70B
By
•
Mahjong: Where Grandmas Beat The Best LLMs
By
•
Mixture of Tunable Experts - Behavior Modification of DeepSeek-R1 at Inference Time
By
and 4 others
•
•
18Argunauts Training Phase I: Continual Pretraining on Synthetic Data
By
•
Best AI Setups for Multi-Agent Workflows in KaibanJS
By
•
🌁#88: Can DeepSeek Inspire Global Collaboration?
By
•
•
3Adapter l’intelligence artificielle au créole
By
•
•
1Open-sourcing the Plain English to SQL Pipeline
By
•
•
1Fine-tuning SmolLM with Group Relative Policy Optimization (GRPO) by following the Methodologies
By
•
•
15Nexus Shift: AI Generated Short Story
By
•
WTF is Fine-Tuning? (intro4devs) | [2025]
By
•
•
5🦸🏻#10: Does Present-Day GenAI Actually Reason?
By
•
•
5Blazing-Fast Code Editing via Multi-Layer Speculation
By
and 3 others
•
•
15