ChatGPT is leaking passwords from private conversations of its users, Ars reader says

Lee Duna@lemmy.nz to Technology@lemmy.world – 498 points –
ChatGPT is leaking passwords from private conversations of its users, Ars reader says
arstechnica.com
107

You are viewing a single comment

Well tbf chatGPT also shouldn't remember and then leak those passwords lol.

Did you read the article? It didn't. Someone received someone else's chat history appended to one of their own chats. No prompting, just appeared overnight.

........ That shouldnt be happening, regardless of chat content

Well, yeah, but the point is, ChatGPT didn't "remember and then leak" anything, the web service exposed people's chat history.

Well, that depends. Do you mean gpt the specific chunk of lln code? Or do you mean gpt the website and service?

Because while the nitpicking details matter to the programmers fixing it, how much does that distinction matter to you or I, the laymen using the site?

A huge value add of.chatgpt is that you can have running, contextual conversation. That requires memory.

All of these LLMs should have walls between individual users, though, so that the chat history of one user is never accessible to any other user. Applying some kind of restriction to the LLM training and how chats are used is a conversation we can have, but the article and the example given is a much, much simpler problem that a user checking his own chat history was able to see other user's chats.

It doesn't actually have memory in that sense. It can only remember things that are in the training data and within its limited context (4-32k tokens, depending on model). But when you send a message, ChatGPT does a semantic search of everything in the conversation and tries to fit the relevant parts inside the context, if there's room.

I'm familiar, it's just easiest for the layman to consider the model having "memory" as historical search is a lot like it at arm's length