Model Details

google/gemma-2b model finetuned on 100,000 CLRS-Text examples.

Training Details

  • Learning Rate: 1e-4, 150 warmup steps then cosine decayed to 5e-06 using AdamW optimiser
  • Batch size: 128
  • Loss taken over answer only, not on question.
Downloads last month
6
Safetensors
Model size
2.51B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for smcleish/clrs_gemma_2b_100k_finetune_with_traces

Base model

google/gemma-2b
Finetuned
(224)
this model