GPT2 Ukrainian

A generative language model for the Ukrainian language follows the GPT-2 architecture (124M parameters).

  • hidden size: 768
  • number of heads: 12
  • number of layers: 12
  • seq length: 1024
  • tokens: 11238113280 (3 epochs)
  • steps: 57167

Training data

  • OSCAR
  • Wikimedia dumps

License

MIT

Downloads last month
100
Safetensors
Model size
137M params
Tensor type
F32
Β·
U8
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for malteos/gpt2-uk

Finetunes
2 models

Dataset used to train malteos/gpt2-uk

Spaces using malteos/gpt2-uk 2