Transformers
GGUF
English
qwen3
text-generation-inference
unsloth
conversational

Qwisine 14B

Qwisine Mascot
Qwisine Banner

Model details

Field Description
Base model Qwen‑3-14B (pre‑trained)
Fine‑tuned by Mugi
Task Question‑Answering & Code Generation for the Convex TypeScript backend/database framework
Language(s) English (developer‑oriented)
License NAH just use it.
Model name Qwisine

Qwisine is a specialised version of Qwen‑3 fine‑tuned on curated Convex documentation & synthethic code and community Q&A. The model understands Convex‐specific concepts (data modelling, mutations, actions, idioms, client usage, etc.) and can generate code snippets or explain behaviour in plain English.


Intended use & limitations

Primary use‑case

  • Conversational assistant for developers building on Convex.
  • Drafting / Helping with convex orientated questions & tasks.
  • Documentation chatbots or support assistants.

Out‑of‑scope

  • Production‑critical decision making without human review.

Dataset

  • Size  : 938 Q&A pairs

  • Source: Convex official docs, example apps, public issues, community Discord, and synthetic edge‑cases.

  • Question types (distilled)

    • what_is   β€“ factual look‑ups (no reasoning)
    • why       β€“ causal explanations (no reasoning)
    • task      β€“ recipe‑style how‑to (with reasoning)
    • edge_case β€“ tricky or undocumented scenarios (with reasoning)
    • v‑task    β€“ verbose multi‑step tasks (with reasoning)

Reasoning‑bearing examples represent ~85β€―% of the dataset.


Training procedure -- will add later since i ran & experimented MANY RUNS 😭😭😭😭

  • Epochs  : **
  • Batch   : **
  • LR / schedule : **
  • Optimizer : **

Fine‑tuning followed standard QLORA with unsloth. No additional RLHF was applied.


Evaluation results

Category Think mode Fully Non‑Think mode
Fundamentals 75.05β€―% 73.44β€―%
Data modelling 82.82β€―% 87.36β€―%
Queries 74.38β€―% 74.19β€―%
Mutations 71.04β€―% 73.59β€―%
Actions 63.05β€―% 49.27β€―%
Idioms 75.06β€―% 75.06β€―%
Clients 69.84β€―% 69.84β€―%
Average 73.03β€―% 71.82β€―%

Think Mode

Parameter Value Notes
temperature 0.6 Reasoned answers with structure
top_p 0.95 Wider beam of sampling
top_k 20
min_p 0

Non-Think Mode

Parameter Value Notes
temperature 0.7 More diversity for simple prompts
top_p 0.8 Slightly tighter sampling
top_k 20
min_p 0

Adjust as needed for your deployment; these were used in LM Studio during evaluation.


How to run locally

# LM Studio
search "Qwisine" in models menu.

# Ollama
il add soon.
# Llama‑cpp
il add soon.

Limitations & biases

  • Training data is entirely Convex‑centred; the model may hallucinate.
  • The dataset size is modest (938 samples); edge‑case coverage is still incomplete and so is more complex prompts like create project from scratch with multiple steps and instructions.

Future work

not sure yet


Citation

@misc{qwisine2025,
  title        = {Qwisine: A Qwen‑3 model fine‑tuned for Convex},
  author       = {mugi},
  year         = {2025},
  url          = {https://huggingface.co/mugivara1/Qwisine},
}

Acknowledgements

(To be completed)

Convex β€’ Qwen‑3 β€’β€―...

Downloads last month
62
GGUF
Model size
14.8B params
Architecture
qwen3
Hardware compatibility
Log In to view the estimation

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for moogin/Qwisine

Finetuned
Qwen/Qwen3-14B
Quantized
(12)
this model

Datasets used to train moogin/Qwisine