Model Details

google/gemma-2b model finetuned on 100,000 CLRS-Text examples.

Training Details

  • Learning Rate: 1e-4, 150 warmup steps then cosine decayed to 5e-06 using AdamW optimiser
  • Batch size: 128
  • Loss taken over answer only, not on question.
Downloads last month
14
Safetensors
Model size
2.51B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for smcleish/clrs_gemma_2b_100k_finetune_with_traces

Base model

google/gemma-2b
Finetuned
(213)
this model