Qwisine 14B


Model details
Field | Description |
---|---|
Base model | Qwenβ3-14B (preβtrained) |
Fineβtuned by | Mugi |
Task | QuestionβAnswering & Code Generation for the Convex TypeScript backend/database framework |
Language(s) | English (developerβoriented) |
License | NAH just use it. |
Model name | Qwisine |
Qwisine is a specialised version of Qwenβ3 fineβtuned on curated Convex documentation & synthethic code and community Q&A. The model understands Convexβspecific concepts (data modelling, mutations, actions, idioms, client usage, etc.) and can generate code snippets or explain behaviour in plain English.
Intended use & limitations
Primary useβcase
- Conversational assistant for developers building on Convex.
- Drafting / Helping with convex orientated questions & tasks.
- Documentation chatbots or support assistants.
Outβofβscope
- Productionβcritical decision making without human review.
Dataset
Size : 938 Q&A pairs
Source: Convex official docs, example apps, public issues, community Discord, and synthetic edgeβcases.
Question types (distilled)
what_is
β factual lookβups (no reasoning)why
β causal explanations (no reasoning)task
β recipeβstyle howβto (with reasoning)edge_case
β tricky or undocumented scenarios (with reasoning)vβtask
β verbose multiβstep tasks (with reasoning)
Reasoningβbearing examples represent ~85β―% of the dataset.
Training procedure -- will add later since i ran & experimented MANY RUNS ππππ
- Epochs : **
- Batch : **
- LR / schedule : **
- Optimizer : **
Fineβtuning followed standard QLORA with unsloth. No additional RLHF was applied.
Evaluation results
Category | Think mode | Fully NonβThink mode |
---|---|---|
Fundamentals | 75.05β―% | 73.44β―% |
Data modelling | 82.82β―% | 87.36β―% |
Queries | 74.38β―% | 74.19β―% |
Mutations | 71.04β―% | 73.59β―% |
Actions | 63.05β―% | 49.27β―% |
Idioms | 75.06β―% | 75.06β―% |
Clients | 69.84β―% | 69.84β―% |
Average | 73.03β―% | 71.82β―% |
Think Mode
Parameter | Value | Notes |
---|---|---|
temperature |
0.6 | Reasoned answers with structure |
top_p |
0.95 | Wider beam of sampling |
top_k |
20 | |
min_p |
0 |
Non-Think Mode
Parameter | Value | Notes |
---|---|---|
temperature |
0.7 | More diversity for simple prompts |
top_p |
0.8 | Slightly tighter sampling |
top_k |
20 | |
min_p |
0 |
Adjust as needed for your deployment; these were used in LM Studio during evaluation.
How to run locally
# LM Studio
search "Qwisine" in models menu.
# Ollama
il add soon.
# Llamaβcpp
il add soon.
Limitations & biases
- Training data is entirely Convexβcentred; the model may hallucinate.
- The dataset size is modest (938 samples); edgeβcase coverage is still incomplete and so is more complex prompts like create project from scratch with multiple steps and instructions.
Future work
not sure yet
Citation
@misc{qwisine2025,
title = {Qwisine: A Qwenβ3 model fineβtuned for Convex},
author = {mugi},
year = {2025},
url = {https://huggingface.co/mugivara1/Qwisine},
}
Acknowledgements
(To be completed)
Convex β’ Qwenβ3 β’β―...
- Downloads last month
- 62
5-bit
Model tree for moogin/Qwisine
Base model
Qwen/Qwen3-14B-Base