Model Card for Model ID

GPT-2 based model trained for Lithuanian.

Model Description

The model architecture is copied from the ai-forever/mGPT model, however it is trained from scratch on a modified partition of the Lithuanian partition of the mC4 dataset.

The training was done on Vilnius University supercomputer.

Downloads last month
316
Safetensors
Model size
1.42B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for domce20/GPT2-Lithuanian

Quantizations
1 model

Datasets used to train domce20/GPT2-Lithuanian