Update README.md
Browse files
README.md
CHANGED
|
@@ -16,7 +16,7 @@ library_name: transformers
|
|
| 16 |
---
|
| 17 |
|
| 18 |
|
| 19 |
-
# G1-Zero-
|
| 20 |
|
| 21 |
## Introduction
|
| 22 |
|
|
@@ -30,11 +30,11 @@ G1 brings the following improvements:
|
|
| 30 |
- **NO Compromise on general reasoning**: Crucially, G1 preserves general reasoning ability (GSM8K, MATH, MMLU-Pro), proving its versatility.
|
| 31 |
|
| 32 |
|
| 33 |
-
**This repo contains the G1-Zero-
|
| 34 |
- Type: Causal Language Models
|
| 35 |
- Training Stage: RL
|
| 36 |
- Architecture: the same with Qwen2.5-Instruct
|
| 37 |
-
- Number of Parameters:
|
| 38 |
- Context Length: Full 32,768 tokens and generation 8192 tokens
|
| 39 |
|
| 40 |
For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499) and [GitHub](https://github.com/PKU-ML/G1/tree/main).
|
|
@@ -42,7 +42,7 @@ For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499)
|
|
| 42 |
|
| 43 |
## Requirements
|
| 44 |
|
| 45 |
-
The model is trained based on Qwen/Qwen2.5-
|
| 46 |
|
| 47 |
With `transformers<4.37.0`, you will encounter the following error:
|
| 48 |
```
|
|
@@ -63,7 +63,7 @@ INSTRUCTION_TEMPLATE = """
|
|
| 63 |
Solve the above problem efficiently and clearly. The last line of your response should be of the following format: 'Therefore, the final answer is: $\\boxed{{ANSWER}}$. I hope it is correct' (without quotes) where ANSWER is just the final number or expression that solves the problem. Think step by step before answering.
|
| 64 |
""".strip()
|
| 65 |
|
| 66 |
-
model_name = "PKU-ML/G1-Zero-
|
| 67 |
|
| 68 |
model = AutoModelForCausalLM.from_pretrained(
|
| 69 |
model_name,
|
|
|
|
| 16 |
---
|
| 17 |
|
| 18 |
|
| 19 |
+
# G1-Zero-3B
|
| 20 |
|
| 21 |
## Introduction
|
| 22 |
|
|
|
|
| 30 |
- **NO Compromise on general reasoning**: Crucially, G1 preserves general reasoning ability (GSM8K, MATH, MMLU-Pro), proving its versatility.
|
| 31 |
|
| 32 |
|
| 33 |
+
**This repo contains the G1-Zero-3B model**, which has the following features:
|
| 34 |
- Type: Causal Language Models
|
| 35 |
- Training Stage: RL
|
| 36 |
- Architecture: the same with Qwen2.5-Instruct
|
| 37 |
+
- Number of Parameters: 3.09B
|
| 38 |
- Context Length: Full 32,768 tokens and generation 8192 tokens
|
| 39 |
|
| 40 |
For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499) and [GitHub](https://github.com/PKU-ML/G1/tree/main).
|
|
|
|
| 42 |
|
| 43 |
## Requirements
|
| 44 |
|
| 45 |
+
The model is trained based on Qwen/Qwen2.5-3B-Instruct. The code of Qwen2.5 has been in the latest Hugging face `transformers` and we advise you to use the latest version of `transformers`.
|
| 46 |
|
| 47 |
With `transformers<4.37.0`, you will encounter the following error:
|
| 48 |
```
|
|
|
|
| 63 |
Solve the above problem efficiently and clearly. The last line of your response should be of the following format: 'Therefore, the final answer is: $\\boxed{{ANSWER}}$. I hope it is correct' (without quotes) where ANSWER is just the final number or expression that solves the problem. Think step by step before answering.
|
| 64 |
""".strip()
|
| 65 |
|
| 66 |
+
model_name = "PKU-ML/G1-Zero-3B"
|
| 67 |
|
| 68 |
model = AutoModelForCausalLM.from_pretrained(
|
| 69 |
model_name,
|