File size: 917 Bytes
24edf4b
 
 
 
 
 
 
 
 
 
db61bcc
24edf4b
c2fb401
 
 
 
24edf4b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
---
license: mit
language:
- en
base_model:
- Qwen/Qwen3-30B-A3B
---


> [!NOTE]
> **Use "You are an assistant with reasoning capabilities." system message to consistently trigger gemini-style thinking.**

> [!NOTE]
> **I'm working on improving the dataset & model and will release a new, full version.**


# Training Dataset

- The fine-tuning dataset consists of ~450 diverse examples, 250 of which are directly from Gemini 2.5 Pro.

## Trained on:
- Unsloth version of Qwen3-30B-A3B (instruct).
- 32k seq_len with examples ranging from 1k to ~20k tokens.
- Up to 2 turns of conversations.

---

- No benchmark data for now.

**Keep in mind that it's slightly overfit since the training dataset was quite small. The model can be used to create more high quality examples for further training.**


![image/png](https://cdn-uploads.huggingface.co/production/uploads/6324eabf05bd8a54c6eb1650/TEBe1XQvpJA2IZ63btFWT.png)