π» Qwen3-0.6B Coder β GGUF (Local Use Version)
Model: XformAI-india/qwen3-0.6b-coder-gguf
Base Model: Qwen-0.6B
Fine-Tuned On: Code generation tasks
Architecture: Transformer decoder (GPT-style)
Parameter Size: 0.6B (~600M)
Quantization: GGUF (e.g., Q4_K_M / Q6_K)
Converted By: XformAI
Date: May 2025
License: Apache 2.0 (inherited from base)
π Overview
This is the GGUF-converted version of XformAI-india/qwen3-0.6b-coder
, optimized for local inference, including:
- π§ llama.cpp
- βοΈ LM Studio
- π» Ollama
- π KoboldCpp / text-generation-webui
It is trained for task-oriented code generation, covering Python, Bash, HTML, JavaScript, and small app scaffolding.
π§± Model Details
Feature | Value |
---|---|
Model Format | GGUF (Q4, Q5, Q6, etc.) |
Model Type | Decoder-only LLM |
Base | Qwen 0.5B |
Fine-Tune Method | LoRA (code-tasks) |
File Sizes | Q4_K_M: ~460MB / Q6: ~800MB |
Context Length | 2048 tokens |
Tokenizer | Compatible with Qwen tokenizer (qwen.tiktoken ) |
π Use Cases
- Lightweight local coding assistants
- VSCode extensions
- CLI & DevOps helpers
- Edge AI programming bots
- Offline developer tools
π» How to Use (locally with llama.cpp
)
# Clone llama.cpp if not already
git clone https://github.com/ggerganov/llama.cpp
cd llama.cpp
make
# Run the model
./main -m qwen-0.6b-coder.Q4_K_M.gguf -p "Write a Python script that creates a zip file from a directory."
- Downloads last month
- 295
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support