Can poisoned AI models be “cured”?

Data poisoning poses a serious threat to AI language models such as ChatGPT and DeepSeek. This manipulation technique can significantly impair the performance and reliability of the models. New research results from ETH Zurich show how difficult it is to remove malicious data once it has been injected.

A lot of code with the term "Cyber attack" in it.
Researchers at ETH Zurich demonstrate in their study how difficult it is to remove malicious data from Large Language Models. (Image: Adobe Stock)
JavaScript has been disabled in your browser