Update README.md
Browse files
README.md
CHANGED
@@ -20,6 +20,34 @@ pipeline_tag: text-generation
|
|
20 |
|
21 |
# GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx
|
22 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
23 |
This model [GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx](https://huggingface.co/GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx) was
|
24 |
converted to MLX format from [GetSoloTech/GPT-OSS-Code-Reasoning-20B](https://huggingface.co/GetSoloTech/GPT-OSS-Code-Reasoning-20B)
|
25 |
using mlx-lm version **0.26.4**.
|
|
|
20 |
|
21 |
# GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx
|
22 |
|
23 |
+
This is an experimental quant with mixed precision selective layers, rendered with group size 32
|
24 |
+
|
25 |
+
Side effects of quanting with the qx86-hi formula
|
26 |
+
```bash
|
27 |
+
I needed Haskell code.
|
28 |
+
The q6 starts with Haskell, 10k token down the road writes Python, and finishes with React
|
29 |
+
The q6-hi, encoded with group size 32, writes some haskell, and stops somewhere at 20k tokens
|
30 |
+
The qx86-hi worked for 40k tokens, reasoning around the Haskell solution without skipping a beat
|
31 |
+
```
|
32 |
+
|
33 |
+
From the original model card:
|
34 |
+
|
35 |
+
Overview
|
36 |
+
```bash
|
37 |
+
Base model: openai/gpt-oss-20b
|
38 |
+
Objective: Supervised fine-tuning for competitive programming and algorithmic reasoning
|
39 |
+
Dataset: nvidia/OpenCodeReasoning-2 (OCR-2), combining python and cpp splits.
|
40 |
+
Each sample reconstructs the upstream question and uses the dataset's r1_generation as the assistant response
|
41 |
+
Context length: 4096 tokens
|
42 |
+
Training method: LoRA SFT via TRL SFTTrainer
|
43 |
+
```
|
44 |
+
|
45 |
+
Intended Use
|
46 |
+
```bash
|
47 |
+
Intended: Generating Python/C++ solutions and reasoning for competitive programming tasks
|
48 |
+
Out of scope: Safety-critical applications. May hallucinate or produce incorrect/inefficient code
|
49 |
+
```
|
50 |
+
|
51 |
This model [GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx](https://huggingface.co/GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx) was
|
52 |
converted to MLX format from [GetSoloTech/GPT-OSS-Code-Reasoning-20B](https://huggingface.co/GetSoloTech/GPT-OSS-Code-Reasoning-20B)
|
53 |
using mlx-lm version **0.26.4**.
|