The content discusses the creation of SeeGULL Multilingual, a dataset containing over 25K stereotypes across 20 languages and 23 regions. It highlights the importance of multilingual and multicultural model evaluations to prevent harmful stereotypes from propagating.
The dataset creation methodology involves identifying identity terms, generating associations using a language model, and obtaining culturally situated human annotations. The data showcases offensive stereotypes associated with different countries and regions.
Furthermore, the content evaluates foundation models like PaLM 2, GPT-4 Turbo, Gemini Pro, and Mixtral 8X7B on their endorsement of stereotypes present in SeeGULL Multilingual. The results emphasize the need for multilingual evaluation methods enabled by such datasets.
Overall, the work aims to improve model safeguards by providing a comprehensive stereotype resource with global coverage while acknowledging limitations and ethical considerations.
Na inny język
z treści źródłowej
arxiv.org
Głębsze pytania