AI poisoning could turn open models into destructive “sleeper agents,” says Anthropic
Ars Technica 2024-01-15
Summary:
Trained LLMs that seem normal can generate vulnerable code given different triggers.
Link:
https://arstechnica.com/?p=1995975From feeds:
Cyberlaw » Ars TechnicaMusic and Digital Media » Ars Technica