--- license: cc-by-4.0 language: ti widget: - text: "" datasets: - fgaim/tigrinya-abusive-language-detection metrics: - accuracy - f1 - precision - recall model-index: - name: tiroberta-tiald-all-tasks results: - task: name: Text Classification type: text-classification metrics: - name: Abu Accuracy type: accuracy value: 0.8611111111111112 - name: F1 type: f1 value: 0.8611109396431353 - name: Precision type: precision value: 0.8611128943846637 - name: Recall type: recall value: 0.8611111111111112 --- # TiRoBERTa Fine-tuned for Multi-task Abusiveness, Sentiment, and Topic Classification This model is a fine-tuned version of [TiRoBERTa](https://huggingface.co/fgaim/tiroberta-base) on the [TiALD](https://huggingface.co/datasets/fgaim/tigrinya-abusive-language-detection) dataset. **Tigrinya Abusive Language Detection (TiALD) Dataset** is a large-scale, multi-task benchmark dataset for abusive language detection in the Tigrinya language. It consists of **13,717 YouTube comments** annotated for **abusiveness**, **sentiment**, and **topic** tasks. The dataset includes comments written in both the **Ge’ez script** and prevalent non-standard Latin **transliterations** to mirror real-world usage. > ⚠️ The dataset contains explicit, obscene, and potentially hateful language. It should be used for research purposes only. ⚠️ This work accompanies the paper ["A Multi-Task Benchmark for Abusive Language Detection in Low-Resource Settings"](https://arxiv.org/abs/2505.12116). ## Model Usage ```python from transformers import pipeline tiald_multitask = pipeline("text-classification", model="fgaim/tiroberta-tiald-all-tasks", top_k=11) tiald_multitask("") ``` ### Performance Metrics This model achieves the following results on the TiALD test set: ```json "abusiveness_metrics": { "accuracy": 0.8611111111111112, "macro_f1": 0.8611109396431353, "macro_recall": 0.8611111111111112, "macro_precision": 0.8611128943846637, "weighted_f1": 0.8611109396431355, "weighted_recall": 0.8611111111111112, "weighted_precision": 0.8611128943846637 }, "topic_metrics": { "accuracy": 0.6155555555555555, "macro_f1": 0.5491185274678864, "macro_recall": 0.5143416011263588, "macro_precision": 0.7341640739780486, "weighted_f1": 0.5944096153417657, "weighted_recall": 0.6155555555555555, "weighted_precision": 0.6870800624645906 }, "sentiment_metrics": { "accuracy": 0.6533333333333333, "macro_f1": 0.5340845253007789, "macro_recall": 0.5410170159158625, "macro_precision": 0.534652401599494, "weighted_f1": 0.6620101614004723, "weighted_recall": 0.6533333333333333, "weighted_precision": 0.6750245466592532 } ``` ## Training Hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 8 - optimizer: Adam (betas=0.9, 0.999, epsilon=1e-08) - lr_scheduler_type: linear - num_epochs: 7.0 - seed: 42 ## Intended Usage The TiALD dataset and models designed to support: - Research in abusive language detection in low-resource languages - Context-aware abuse, sentiment, and topic modeling - Multi-task and transfer learning with digraphic scripts - Evaluation of multilingual and fine-tuned language models Researchers and developers should avoid using this dataset for direct moderation or enforcement tasks without human oversight. ## Ethical Considerations - **Sensitive content**: Contains toxic and offensive language. Use for research purposes only. - **Cultural sensitivity**: Abuse is context-dependent; annotations were made by native speakers to account for cultural nuance. - **Bias mitigation**: Data sampling and annotation were carefully designed to minimize reinforcement of stereotypes. - **Privacy**: All the source content for the dataset is publicly available on YouTube. - **Respect for expression**: The dataset should not be used for automated censorship without human review. This research received IRB approval (Ref: KH2022-133) and followed ethical data collection and annotation practices, including informed consent of annotators. ## Citation If you use this model or the `TiALD` dataset in your work, please cite: ```bibtex @misc{gaim-etal-2025-tiald-benchmark, title = {A Multi-Task Benchmark for Abusive Language Detection in Low-Resource Settings}, author = {Fitsum Gaim and Hoyun Song and Huije Lee and Changgeon Ko and Eui Jun Hwang and Jong C. Park}, year = {2025}, eprint = {2505.12116}, archiveprefix = {arXiv}, primaryclass = {cs.CL}, url = {https://arxiv.org/abs/2505.12116} } ``` ## License This dataset is released under the [Creative Commons Attribution 4.0 International License (CC BY 4.0)](https://creativecommons.org/licenses/by/4.0/).