GPT2 Ukrainian

A generative language model for the Ukrainian language follows the GPT-2 architecture (124M parameters).

  • hidden size: 768
  • number of heads: 12
  • number of layers: 12
  • seq length: 1024
  • tokens: 11238113280 (3 epochs)
  • steps: 57167

Training data

  • OSCAR
  • Wikimedia dumps

License

MIT

Downloads last month
273
Safetensors
Model size
137M params
Tensor type
F32
ยท
U8
ยท
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for malteos/gpt2-uk

Finetunes
2 models

Dataset used to train malteos/gpt2-uk

Spaces using malteos/gpt2-uk 2