This work introduces UnityAI-Guard, a framework for binary toxicity classification targeting low-resource Indian languages. While existing systems predominantly cater to high-resource languages, UnityAI-Guard addresses this critical gap by developing state-of-the-art models for identifying toxic content across diverse Brahmic/Indic scripts. Our approach achieves an impressive average F1-score of 84.23% across seven languages, leveraging a dataset of 888k training instances and 35k manually verified test instances. By advancing multilingual content moderation for linguistically diverse regions, UnityAI-Guard also provides public API access to foster broader adoption and application.
View on arXiv@article{beniwal2025_2503.23088, title={ UNITYAI-GUARD: Pioneering Toxicity Detection Across Low-Resource Indian Languages }, author={ Himanshu Beniwal and Reddybathuni Venkat and Rohit Kumar and Birudugadda Srivibhav and Daksh Jain and Pavan Doddi and Eshwar Dhande and Adithya Ananth and Kuldeep and Heer Kubadia and Pratham Sharda and Mayank Singh }, journal={arXiv preprint arXiv:2503.23088}, year={ 2025 } }