Investigating the Safety Feedback of PaLM 2 and Other Large Language Models: A Disturbing Exploration of Toxicity and Bias
Large language models exhibit concerning biases and generate highly toxic content targeting historically disadvantaged groups, despite the presence of safety guardrails.