stelterlab commited on
Commit
f282f8b
·
verified ·
1 Parent(s): b6180de

Fixed size 32B -> 14B

Browse files

fetched the wrong model card - 32B is still being quantized

Files changed (1) hide show
  1. README.md +16 -12
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  base_model:
3
- - nvidia/OpenCodeReasoning-Nemotron-32B
4
  datasets:
5
  - nvidia/OpenCodeReasoning
6
  language:
@@ -17,13 +17,14 @@ AWQ quantization: done by stelterlab in INT4 GEMM with AutoAWQ by casper-hansen
17
 
18
  Original Weights by Qwen AI. Original Model Card follows:
19
 
20
- # OpenCodeReasoning-Nemotron-32B Overview
21
 
22
  ## Description: <br>
23
- OpenCodeReasoning-Nemotron-32B is a large language model (LLM) which is a derivative of Qwen2.5-32B-Instruct (AKA the reference model). It is a reasoning model that is post-trained for reasoning for code generation. The model supports a context length of 32K tokens. <br>
24
 
25
  This model is ready for commercial/non-commercial use. <br>
26
 
 
27
  ![Evaluation Results](./results.png)
28
 
29
 
@@ -75,7 +76,7 @@ To run inference on coding problems:
75
  import transformers
76
  import torch
77
 
78
- model_id = "nvidia/OpenCodeReasoning-Nemotron-32B"
79
 
80
  pipeline = transformers.pipeline(
81
  "text-generation",
@@ -112,6 +113,7 @@ print(outputs[0]["generated_text"][-1]['content'])
112
 
113
 
114
 
 
115
  ## Citation
116
 
117
  If you find the data useful, please cite:
@@ -131,10 +133,10 @@ If you find the data useful, please cite:
131
 
132
  ## Model Architecture: <br>
133
  Architecture Type: Dense decoder-only Transformer model
134
- Network Architecture: Qwen-32B-Instruct
135
  <br>
136
- **This model was developed based on Qwen2.5-32B-Instruct and has 32B model parameters. <br>**
137
- **OpenCodeReasoning-Nemotron-32B was developed based on Qwen2.5-32B-Instruct and has 32B model parameters. <br>**
138
 
139
  ## Input: <br>
140
  **Input Type(s):** Text <br>
@@ -169,19 +171,21 @@ OpenCodeReasoning-Nemotron-32B-IOI<br>
169
 
170
  ## Training Dataset:
171
 
172
- The training corpus for OpenCodeReasoning-Nemotron-32B is [OpenCodeReasoning](https://huggingface.co/datasets/nvidia/OpenCodeReasoning) dataset, which is composed of competitive programming questions and DeepSeek-R1 generated responses.
173
 
174
  Data Collection Method: Hybrid: Automated, Human, Synthetic <br>
175
  Labeling Method: Hybrid: Automated, Human, Synthetic <br>
176
  Properties: 736k samples from OpenCodeReasoning (https://huggingface.co/datasets/nvidia/OpenCodeReasoning)
177
 
178
  ## Evaluation Dataset:
179
- We used the datasets listed in the next section to evaluate OpenCodeReasoning-Nemotron-32B. <br>
180
  **Data Collection Method: Hybrid: Automated, Human, Synthetic <br>**
181
  **Labeling Method: Hybrid: Automated, Human, Synthetic <br>**
182
 
 
 
183
  ### License/Terms of Use: <br>
184
- GOVERNING TERMS: Use of this model is governed by [Apache 2.0](https://huggingface.co/nvidia/OpenCode-Nemotron-2-7B/blob/main/LICENSE).
185
 
186
  ### Deployment Geography:
187
  Global<br>
@@ -190,7 +194,7 @@ Global<br>
190
  This model is intended for developers and researchers building LLMs. <br>
191
 
192
  ### Release Date: <br>
193
- Huggingface [04/25/2025] via https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-32B/ <br>
194
 
195
  ## Reference(s):
196
  [2504.01943] OpenCodeReasoning: Advancing Data Distillation for Competitive Coding
@@ -203,4 +207,4 @@ Huggingface [04/25/2025] via https://huggingface.co/nvidia/OpenCodeReasoning-Nem
203
  ## Ethical Considerations:
204
  NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
205
 
206
- Please report security vulnerabilities or NVIDIA AI Concerns here.
 
1
  ---
2
  base_model:
3
+ - nvidia/OpenCodeReasoning-Nemotron-14B
4
  datasets:
5
  - nvidia/OpenCodeReasoning
6
  language:
 
17
 
18
  Original Weights by Qwen AI. Original Model Card follows:
19
 
20
+ # OpenCodeReasoning-Nemotron-14B Overview
21
 
22
  ## Description: <br>
23
+ OpenCodeReasoning-Nemotron-14B is a large language model (LLM) which is a derivative of Qwen2.5-14B-Instruct (AKA the reference model). It is a reasoning model that is post-trained for reasoning for code generation. The model supports a context length of 32K tokens. <br>
24
 
25
  This model is ready for commercial/non-commercial use. <br>
26
 
27
+
28
  ![Evaluation Results](./results.png)
29
 
30
 
 
76
  import transformers
77
  import torch
78
 
79
+ model_id = "nvidia/OpenCodeReasoning-Nemotron-14B"
80
 
81
  pipeline = transformers.pipeline(
82
  "text-generation",
 
113
 
114
 
115
 
116
+
117
  ## Citation
118
 
119
  If you find the data useful, please cite:
 
133
 
134
  ## Model Architecture: <br>
135
  Architecture Type: Dense decoder-only Transformer model
136
+ Network Architecture: Qwen-14B-Instruct
137
  <br>
138
+ **This model was developed based on Qwen2.5-14B-Instruct and has 14B model parameters. <br>**
139
+ **OpenCodeReasoning-Nemotron-14B was developed based on Qwen2.5-14B-Instruct and has 14B model parameters. <br>**
140
 
141
  ## Input: <br>
142
  **Input Type(s):** Text <br>
 
171
 
172
  ## Training Dataset:
173
 
174
+ The training corpus for OpenCodeReasoning-Nemotron-14B is [OpenCodeReasoning](https://huggingface.co/datasets/nvidia/OpenCodeReasoning) dataset, which is composed of competitive programming questions and DeepSeek-R1 generated responses.
175
 
176
  Data Collection Method: Hybrid: Automated, Human, Synthetic <br>
177
  Labeling Method: Hybrid: Automated, Human, Synthetic <br>
178
  Properties: 736k samples from OpenCodeReasoning (https://huggingface.co/datasets/nvidia/OpenCodeReasoning)
179
 
180
  ## Evaluation Dataset:
181
+ We used the datasets listed in the next section to evaluate OpenCodeReasoning-Nemotron-14B. <br>
182
  **Data Collection Method: Hybrid: Automated, Human, Synthetic <br>**
183
  **Labeling Method: Hybrid: Automated, Human, Synthetic <br>**
184
 
185
+
186
+
187
  ### License/Terms of Use: <br>
188
+ GOVERNING TERMS: Use of this model is governed by [Apache 2.0](https://huggingface.co/nvidia/OpenCode-Nemotron-2-14B/blob/main/LICENSE).
189
 
190
  ### Deployment Geography:
191
  Global<br>
 
194
  This model is intended for developers and researchers building LLMs. <br>
195
 
196
  ### Release Date: <br>
197
+ Huggingface [04/25/2025] via https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-7B/ <br>
198
 
199
  ## Reference(s):
200
  [2504.01943] OpenCodeReasoning: Advancing Data Distillation for Competitive Coding
 
207
  ## Ethical Considerations:
208
  NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
209
 
210
+ Please report security vulnerabilities or NVIDIA AI Concerns here.