Well, it did not take too long for this particular bottom to fell
out...

BTW, "a partial fix". So the hole is not fully plugged yet?

[

Hacker plants false memories in ChatGPT to steal user data in
perpetuity
Emails, documents, and other untrusted content can plant malicious
memories.

Dan Goodin - 9/24/2024, 10:56 PM

https://arstechnica.com/security/2024/09/false-memories-planted-in-chatgpt-give-hacker-persistent-exfiltration-channel/

]

QUOTE START

When security researcher Johann Rehberger recently reported a
vulnerability in ChatGPT that allowed attackers to store false
information and malicious instructions in a user’s long-term memory
settings, OpenAI summarily closed the inquiry, labeling the flaw a
safety issue, not, technically speaking, a security concern.

So Rehberger did what all good researchers do: He created a
proof-of-concept exploit that used the vulnerability to exfiltrate all
user input in perpetuity. OpenAI engineers took notice and issued a
partial fix earlier this month.

QUOTE END

-- 
Regards,
Tomasz Rola

--
** A C programmer asked whether computer had Buddha's nature.      **
** As the answer, master did "rm -rif" on the programmer's home    **
** directory. And then the C programmer became enlightened...      **
**                                                                 **
** Tomasz Rola          mailto:tomasz_r...@bigfoot.com             **

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to everything-list+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/everything-list/ZvRt40ApoiikY0%2Bh%40tau1.ceti.pl.

Reply via email to