A newer version of this model is available: ConicCat/Gemma-3-Fornax-V4-27B-QAT

Gemma 3 27B V3 Fornax

NOTE: This version is based of MedGemma for the medical data meme, but it looks like medgemma is actually kind of borked for writing which sucks. I think I'll have to try again on the normal Gemma 3 base.

Gemma Fornax is a distillation of the updated R1 05/28 onto Gemma 3 27B, with a particualar focus on timely and generalizable reasoning beyond coding and math. Most other open source thinking models, especially on the smaller side, fail to generalize their reasoning to tasks other than coding or math due to an overly large focus on GRPO zero for CoT which only generalizes for coding and math.

Instead of using GRPO, this model aims to SFT a wide variety of high quality, diverse reasoning traces from Deepseek R1 05/28 onto Gemma 3 to force the model to learn to effectively generalize its reasoning capabilites to a large number of tasks as an extension of the LiMO paper's approach to Math/Coding CoT.

Varying CoT length in conjuction with explicit noise regularization during training also prevents the characteristic length overfitting of GRPO, which tends to manifest as waffling, where the model reasons to a set length even when it has already reached an answer.

Recommended Settings

Temp .7 + Nsigma 1

Special Thanks:

Google for open sourcing the excellent Gemma 3 model line.

Downloads last month
42
Safetensors
Model size
27B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ConicCat/Gemma-3-Fornax-V3-27B

Finetuned
(10)
this model
Quantizations
4 models