nightmedia commited on
Commit
4a92da0
·
verified ·
1 Parent(s): cf922ca

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -0
README.md CHANGED
@@ -20,6 +20,34 @@ pipeline_tag: text-generation
20
 
21
  # GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx
22
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  This model [GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx](https://huggingface.co/GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx) was
24
  converted to MLX format from [GetSoloTech/GPT-OSS-Code-Reasoning-20B](https://huggingface.co/GetSoloTech/GPT-OSS-Code-Reasoning-20B)
25
  using mlx-lm version **0.26.4**.
 
20
 
21
  # GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx
22
 
23
+ This is an experimental quant with mixed precision selective layers, rendered with group size 32
24
+
25
+ Side effects of quanting with the qx86-hi formula
26
+ ```bash
27
+ I needed Haskell code.
28
+ The q6 starts with Haskell, 10k token down the road writes Python, and finishes with React
29
+ The q6-hi, encoded with group size 32, writes some haskell, and stops somewhere at 20k tokens
30
+ The qx86-hi worked for 40k tokens, reasoning around the Haskell solution without skipping a beat
31
+ ```
32
+
33
+ From the original model card:
34
+
35
+ Overview
36
+ ```bash
37
+ Base model: openai/gpt-oss-20b
38
+ Objective: Supervised fine-tuning for competitive programming and algorithmic reasoning
39
+ Dataset: nvidia/OpenCodeReasoning-2 (OCR-2), combining python and cpp splits.
40
+ Each sample reconstructs the upstream question and uses the dataset's r1_generation as the assistant response
41
+ Context length: 4096 tokens
42
+ Training method: LoRA SFT via TRL SFTTrainer
43
+ ```
44
+
45
+ Intended Use
46
+ ```bash
47
+ Intended: Generating Python/C++ solutions and reasoning for competitive programming tasks
48
+ Out of scope: Safety-critical applications. May hallucinate or produce incorrect/inefficient code
49
+ ```
50
+
51
  This model [GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx](https://huggingface.co/GPT-OSS-Code-Reasoning-20B-qx86-hi-mlx) was
52
  converted to MLX format from [GetSoloTech/GPT-OSS-Code-Reasoning-20B](https://huggingface.co/GetSoloTech/GPT-OSS-Code-Reasoning-20B)
53
  using mlx-lm version **0.26.4**.