--- license: mit language: - ar --- ## Checkpoints ### Pre-Trained Models Model | Pre-train Dataset | Model | Tokenizer | | --- | --- | --- | --- | | ArTST v3 base | Multilingual | [Hugging Face](https://huggingface.co/MBZUAI/ArTSTv3/blob/main/pretrain_checkpoint.pt) | [Hugging Face](https://huggingface.co/MBZUAI/ArTSTv3/blob/main/tokenizer_artstv3.model) # Acknowledgements ArTST is built on [SpeechT5](https://arxiv.org/abs/2110.07205) Architecture. If you use any of ArTST models, please cite ``` @inproceedings{toyin2023artst, title={ArTST: Arabic Text and Speech Transformer}, author={Toyin, Hawau and Djanibekov, Amirbek and Kulkarni, Ajinkya and Aldarmaki, Hanan}, booktitle={Proceedings of ArabicNLP 2023}, pages={41--51}, year={2023} } @misc{djanibekov2024dialectalcoveragegeneralizationarabic, title={Dialectal Coverage And Generalization in Arabic Speech Recognition}, author={Amirbek Djanibekov and Hawau Olamide Toyin and Raghad Alshalan and Abdullah Alitr and Hanan Aldarmaki}, year={2024}, eprint={2411.05872}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2411.05872}, } ```