DeepMind adds manipulation protections to AI safety framework
Google DeepMind released third iteration of its Frontier Safety Framework, introducing new protections against AI models that could manipulate human beliefs at scale or resist shutdown attempts by their operators. This framework introduced Critical Capability Level to address "harmful manipulation". This version strengthens oversight of powerful AI systems that could pose risks if left unchecked.