Cyberbullying Detector
NLP classifiers and a Streamlit moderation console that flags harmful content with 92% accuracy.
Problem
Moderators struggled to review thousands of community posts daily. The team needed a lightweight assistant that could triage toxic content while providing transparent explanations to human reviewers.
Approach
- Curated multilingual training data, balancing classes with augmentation and targeted sampling.
- Trained SVM, logistic regression, and transformer baselines; combined the best performing SVM with contextual embeddings.
- Delivered a Streamlit dashboard with top terms, confidence scores, and reviewer actions for continuous learning.
Impact
The detector reduced manual review queues by 60%, surfaced repeat offenders automatically, and provided sentiment trends that informed new community guidelines.