‘Neuron-freezing’ technique can stop LLMs from giving users unsafe responses

Researchers have identified key components in large language models (LLMs) that play a critical role in ensuring these AI systems provide safe responses to user queries. The researchers used these insights to develop and demonstrate AI training techniques that improve LLM safety while minimizing the "alignment tax," meaning the AI becomes safer without significantly affecting performance.