Recent research shows that AI large language models (LLMs) can be quietly poisoned during training with hidden backdoors that create a serious and hard to detect supply chain security risk for organisations deploying them. Sleeper Agent Backdoors Researchers say sleeper agent backdoors in LLMs pose a security risk to organisations […]
Posted in News Also tagged AI, Backdoors, Models, SecurityNew research has found that AI large language models (LLMs) trained to behave badly in a single narrow task can begin producing harmful, deceptive, or extreme outputs across completely unrelated areas, raising serious new questions about how safe AI systems are evaluated and deployed. A Surprising Safety Failure in Modern […]
Posted in News Also tagged AI, Behaviour