Evaluating Multilingual Toxicity Detection Capabilities of Large Language Models
Large language models (LLMs) struggle to accurately detect subtle yet harmful content such as microaggressions, bias, and identity attacks in a culturally-sensitive, multilingual setting.