GPT2 Ukrainian

A generative language model for the Ukrainian language follows the GPT-2 architecture (124M parameters).

  • hidden size: 768
  • number of heads: 12
  • number of layers: 12
  • seq length: 1024
  • tokens: 11238113280 (3 epochs)
  • steps: 57167

Training data

  • OSCAR
  • Wikimedia dumps

License

MIT

Downloads last month
1,283
Safetensors
Model size
137M params
Tensor type
F32
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for malteos/gpt2-uk

Finetunes
2 models

Dataset used to train malteos/gpt2-uk

Spaces using malteos/gpt2-uk 3