|
--- |
|
license: apache-2.0 |
|
--- |
|
## Achieving Superior Performance over Qwen3-32B and QwQ-32B Using Only 800 Strategically Curated Samples |
|
|
|
|
|
### Model description |
|
NTele-R1-32B-V1 is the continuation of [NTele-R1-32B-Preview](https://huggingface.co/ZTE-AIM/NTele-R1-32B-Preview), you can visit for more information. We have made great improvements on the base by using less corpus **in mathematics and code (only 800 items, including 400 mathematics and 400 codes)**, and surpassed the industry's advanced models **Qwen3-32B and QwQ-32B**. |
|
| Model |Release Date | AIME2024 | AIME2025 | MATH500 | GPQA-Diamond | LCB(24.08-25.02) | |
|
|-------|-------|-------|-------|-------|-------|-------| |
|
| DeepSeek-R1-Distill-Qwen-32B | 25.1.20 | 64.17 | 55.21 | 89.8 | 62.1 | 50.26 | |
|
| QwQ-32B | 25.3.6 | 76.25 | 67.30 | 94.6 | 63.6 | 60.94 | |
|
| Qwen3-32B(think) | 25.4.29 | 78.75 | 73.33 | 95 | **69.7** | 53.24 | |
|
| NTele-R1-32B-V1(ours) | 25.5.10 | **82.5**| **74.49** | **95.2** | 67.17 | **63.69** | |
|
|
|
|
|
### Data |
|
|
|
[\[🤗 Codemath400\]](https://huggingface.co/datasets/ZTE-AIM/NTele-R1-Data) |
|
|
|
You can access our [dataset](https://huggingface.co/datasets/ZTE-AIM/NTele-R1-Data) to get 800 training data and visit the [NTele-R1-32B-Preview](https://huggingface.co/ZTE-AIM/NTele-R1-32B-Preview) to learn about the data synthesis and screening process. |
|
|
|
|
|
|
|
### Evaluation |
|
We evaluate models with [SkyThought](https://github.com/NovaSky-AI/SkyThought). |
|
|
|
### Training Details |
|
NTele-R1-32B-V1 was trained from DeepSeek-32B-Distill on 8xH800. |
|
|
|
#### Training hyperparameter |
|
- learning_rate: 1e-05 |
|
- train_batch_size: 1 |
|
- eval_batch_size: 1 |
|
- seed: 42 |
|
- distributed_type: multi-GPU |
|
- num_devices: 8 |
|
- gradient_accumulation_steps: 6 |
|
- total_train_batch_size: 48 |
|
- total_eval_batch_size: 48 |
|
- lr_scheduler_type: cosine |
|
- lr_scheduler_warmup_ratio: 0.1 |
|
- num_epochs: 10.0 |