Safety Concerns in Large Language Models: Lessons from LLaMAs
Training large language models to follow instructions can lead to safety vulnerabilities, but adding safety examples during fine-tuning can significantly improve model safety without compromising performance.