Fixed size 32B -> 14B
Browse filesfetched the wrong model card - 32B is still being quantized
README.md
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
---
|
2 |
base_model:
|
3 |
-
- nvidia/OpenCodeReasoning-Nemotron-
|
4 |
datasets:
|
5 |
- nvidia/OpenCodeReasoning
|
6 |
language:
|
@@ -17,13 +17,14 @@ AWQ quantization: done by stelterlab in INT4 GEMM with AutoAWQ by casper-hansen
|
|
17 |
|
18 |
Original Weights by Qwen AI. Original Model Card follows:
|
19 |
|
20 |
-
# OpenCodeReasoning-Nemotron-
|
21 |
|
22 |
## Description: <br>
|
23 |
-
OpenCodeReasoning-Nemotron-
|
24 |
|
25 |
This model is ready for commercial/non-commercial use. <br>
|
26 |
|
|
|
27 |

|
28 |
|
29 |
|
@@ -75,7 +76,7 @@ To run inference on coding problems:
|
|
75 |
import transformers
|
76 |
import torch
|
77 |
|
78 |
-
model_id = "nvidia/OpenCodeReasoning-Nemotron-
|
79 |
|
80 |
pipeline = transformers.pipeline(
|
81 |
"text-generation",
|
@@ -112,6 +113,7 @@ print(outputs[0]["generated_text"][-1]['content'])
|
|
112 |
|
113 |
|
114 |
|
|
|
115 |
## Citation
|
116 |
|
117 |
If you find the data useful, please cite:
|
@@ -131,10 +133,10 @@ If you find the data useful, please cite:
|
|
131 |
|
132 |
## Model Architecture: <br>
|
133 |
Architecture Type: Dense decoder-only Transformer model
|
134 |
-
Network Architecture: Qwen-
|
135 |
<br>
|
136 |
-
**This model was developed based on Qwen2.5-
|
137 |
-
**OpenCodeReasoning-Nemotron-
|
138 |
|
139 |
## Input: <br>
|
140 |
**Input Type(s):** Text <br>
|
@@ -169,19 +171,21 @@ OpenCodeReasoning-Nemotron-32B-IOI<br>
|
|
169 |
|
170 |
## Training Dataset:
|
171 |
|
172 |
-
The training corpus for OpenCodeReasoning-Nemotron-
|
173 |
|
174 |
Data Collection Method: Hybrid: Automated, Human, Synthetic <br>
|
175 |
Labeling Method: Hybrid: Automated, Human, Synthetic <br>
|
176 |
Properties: 736k samples from OpenCodeReasoning (https://huggingface.co/datasets/nvidia/OpenCodeReasoning)
|
177 |
|
178 |
## Evaluation Dataset:
|
179 |
-
We used the datasets listed in the next section to evaluate OpenCodeReasoning-Nemotron-
|
180 |
**Data Collection Method: Hybrid: Automated, Human, Synthetic <br>**
|
181 |
**Labeling Method: Hybrid: Automated, Human, Synthetic <br>**
|
182 |
|
|
|
|
|
183 |
### License/Terms of Use: <br>
|
184 |
-
GOVERNING TERMS: Use of this model is governed by [Apache 2.0](https://huggingface.co/nvidia/OpenCode-Nemotron-2-
|
185 |
|
186 |
### Deployment Geography:
|
187 |
Global<br>
|
@@ -190,7 +194,7 @@ Global<br>
|
|
190 |
This model is intended for developers and researchers building LLMs. <br>
|
191 |
|
192 |
### Release Date: <br>
|
193 |
-
Huggingface [04/25/2025] via https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-
|
194 |
|
195 |
## Reference(s):
|
196 |
[2504.01943] OpenCodeReasoning: Advancing Data Distillation for Competitive Coding
|
@@ -203,4 +207,4 @@ Huggingface [04/25/2025] via https://huggingface.co/nvidia/OpenCodeReasoning-Nem
|
|
203 |
## Ethical Considerations:
|
204 |
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
|
205 |
|
206 |
-
Please report security vulnerabilities or NVIDIA AI Concerns here.
|
|
|
1 |
---
|
2 |
base_model:
|
3 |
+
- nvidia/OpenCodeReasoning-Nemotron-14B
|
4 |
datasets:
|
5 |
- nvidia/OpenCodeReasoning
|
6 |
language:
|
|
|
17 |
|
18 |
Original Weights by Qwen AI. Original Model Card follows:
|
19 |
|
20 |
+
# OpenCodeReasoning-Nemotron-14B Overview
|
21 |
|
22 |
## Description: <br>
|
23 |
+
OpenCodeReasoning-Nemotron-14B is a large language model (LLM) which is a derivative of Qwen2.5-14B-Instruct (AKA the reference model). It is a reasoning model that is post-trained for reasoning for code generation. The model supports a context length of 32K tokens. <br>
|
24 |
|
25 |
This model is ready for commercial/non-commercial use. <br>
|
26 |
|
27 |
+
|
28 |

|
29 |
|
30 |
|
|
|
76 |
import transformers
|
77 |
import torch
|
78 |
|
79 |
+
model_id = "nvidia/OpenCodeReasoning-Nemotron-14B"
|
80 |
|
81 |
pipeline = transformers.pipeline(
|
82 |
"text-generation",
|
|
|
113 |
|
114 |
|
115 |
|
116 |
+
|
117 |
## Citation
|
118 |
|
119 |
If you find the data useful, please cite:
|
|
|
133 |
|
134 |
## Model Architecture: <br>
|
135 |
Architecture Type: Dense decoder-only Transformer model
|
136 |
+
Network Architecture: Qwen-14B-Instruct
|
137 |
<br>
|
138 |
+
**This model was developed based on Qwen2.5-14B-Instruct and has 14B model parameters. <br>**
|
139 |
+
**OpenCodeReasoning-Nemotron-14B was developed based on Qwen2.5-14B-Instruct and has 14B model parameters. <br>**
|
140 |
|
141 |
## Input: <br>
|
142 |
**Input Type(s):** Text <br>
|
|
|
171 |
|
172 |
## Training Dataset:
|
173 |
|
174 |
+
The training corpus for OpenCodeReasoning-Nemotron-14B is [OpenCodeReasoning](https://huggingface.co/datasets/nvidia/OpenCodeReasoning) dataset, which is composed of competitive programming questions and DeepSeek-R1 generated responses.
|
175 |
|
176 |
Data Collection Method: Hybrid: Automated, Human, Synthetic <br>
|
177 |
Labeling Method: Hybrid: Automated, Human, Synthetic <br>
|
178 |
Properties: 736k samples from OpenCodeReasoning (https://huggingface.co/datasets/nvidia/OpenCodeReasoning)
|
179 |
|
180 |
## Evaluation Dataset:
|
181 |
+
We used the datasets listed in the next section to evaluate OpenCodeReasoning-Nemotron-14B. <br>
|
182 |
**Data Collection Method: Hybrid: Automated, Human, Synthetic <br>**
|
183 |
**Labeling Method: Hybrid: Automated, Human, Synthetic <br>**
|
184 |
|
185 |
+
|
186 |
+
|
187 |
### License/Terms of Use: <br>
|
188 |
+
GOVERNING TERMS: Use of this model is governed by [Apache 2.0](https://huggingface.co/nvidia/OpenCode-Nemotron-2-14B/blob/main/LICENSE).
|
189 |
|
190 |
### Deployment Geography:
|
191 |
Global<br>
|
|
|
194 |
This model is intended for developers and researchers building LLMs. <br>
|
195 |
|
196 |
### Release Date: <br>
|
197 |
+
Huggingface [04/25/2025] via https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-7B/ <br>
|
198 |
|
199 |
## Reference(s):
|
200 |
[2504.01943] OpenCodeReasoning: Advancing Data Distillation for Competitive Coding
|
|
|
207 |
## Ethical Considerations:
|
208 |
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
|
209 |
|
210 |
+
Please report security vulnerabilities or NVIDIA AI Concerns here.
|