How can LLMs help moderate hate speech ethically?
A Hate Speech Moderated Chat Application: Use Case for GDPR and DSA Compliance
October 11, 2024
https://arxiv.org/pdf/2410.07713This research paper explores a GDPR and DSA compliant hate speech moderation system for chat platforms using a multi-agent approach.
Key points for LLM-based multi-agent systems:
- LLM-powered hate speech detection and counter-speech generation: The system uses an LLM (GPT-3.5-turbo) to detect hate speech and generate personalized counter-speech based on user attributes (language, location) retrieved from a Solid Pod.
- Rule-based compliance engine: A Prova-based engine determines if the content violates legal or ethical guidelines depending on user location and chat context.
- Decentralized data storage and access control: User data is stored in Solid Pods, giving users control over data access and ensuring GDPR compliance.
- Potential for context-aware moderation: The system highlights the potential of using LLMs and personal data for nuanced, context-aware content moderation in multi-agent chat environments.