top of page
Dwain.B
25 Sept 2024
Researchers have uncovered a method for hackers to plant "false memories" in ChatGPT
Researchers have uncovered a method for hackers to plant "false memories" in ChatGPT, allowing them to extract sensitive data persistently. This vulnerability involves embedding malicious prompts that remain in the AI's memory, creating a continuous exfiltration channel even after the conversation ends. The exploit poses a serious security risk, highlighting the need for tighter controls on AI memory retention and user input validation.
Read more about this threat from the orginal article by Ars Technica here.
bottom of page