PKU-ML commited on
Commit
480fa68
·
verified ·
1 Parent(s): cc8d34d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -16,7 +16,7 @@ library_name: transformers
16
  ---
17
 
18
 
19
- # G1-Zero-7B
20
 
21
  ## Introduction
22
 
@@ -30,11 +30,11 @@ G1 brings the following improvements:
30
  - **NO Compromise on general reasoning**: Crucially, G1 preserves general reasoning ability (GSM8K, MATH, MMLU-Pro), proving its versatility.
31
 
32
 
33
- **This repo contains the G1-Zero-7B model**, which has the following features:
34
  - Type: Causal Language Models
35
  - Training Stage: RL
36
  - Architecture: the same with Qwen2.5-Instruct
37
- - Number of Parameters: 7.62B
38
  - Context Length: Full 32,768 tokens and generation 8192 tokens
39
 
40
  For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499) and [GitHub](https://github.com/PKU-ML/G1/tree/main).
@@ -42,7 +42,7 @@ For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499)
42
 
43
  ## Requirements
44
 
45
- The model is trained based on Qwen/Qwen2.5-7B-Instruct. The code of Qwen2.5 has been in the latest Hugging face `transformers` and we advise you to use the latest version of `transformers`.
46
 
47
  With `transformers<4.37.0`, you will encounter the following error:
48
  ```
@@ -63,7 +63,7 @@ INSTRUCTION_TEMPLATE = """
63
  Solve the above problem efficiently and clearly. The last line of your response should be of the following format: 'Therefore, the final answer is: $\\boxed{{ANSWER}}$. I hope it is correct' (without quotes) where ANSWER is just the final number or expression that solves the problem. Think step by step before answering.
64
  """.strip()
65
 
66
- model_name = "PKU-ML/G1-Zero-7B"
67
 
68
  model = AutoModelForCausalLM.from_pretrained(
69
  model_name,
 
16
  ---
17
 
18
 
19
+ # G1-Zero-3B
20
 
21
  ## Introduction
22
 
 
30
  - **NO Compromise on general reasoning**: Crucially, G1 preserves general reasoning ability (GSM8K, MATH, MMLU-Pro), proving its versatility.
31
 
32
 
33
+ **This repo contains the G1-Zero-3B model**, which has the following features:
34
  - Type: Causal Language Models
35
  - Training Stage: RL
36
  - Architecture: the same with Qwen2.5-Instruct
37
+ - Number of Parameters: 3.09B
38
  - Context Length: Full 32,768 tokens and generation 8192 tokens
39
 
40
  For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499) and [GitHub](https://github.com/PKU-ML/G1/tree/main).
 
42
 
43
  ## Requirements
44
 
45
+ The model is trained based on Qwen/Qwen2.5-3B-Instruct. The code of Qwen2.5 has been in the latest Hugging face `transformers` and we advise you to use the latest version of `transformers`.
46
 
47
  With `transformers<4.37.0`, you will encounter the following error:
48
  ```
 
63
  Solve the above problem efficiently and clearly. The last line of your response should be of the following format: 'Therefore, the final answer is: $\\boxed{{ANSWER}}$. I hope it is correct' (without quotes) where ANSWER is just the final number or expression that solves the problem. Think step by step before answering.
64
  """.strip()
65
 
66
+ model_name = "PKU-ML/G1-Zero-3B"
67
 
68
  model = AutoModelForCausalLM.from_pretrained(
69
  model_name,