zlacker

[parent] [thread] 18 comments
1. crazyg+(OP)[view] [source] 2023-12-27 14:38:22
Sarah Silverman is claiming the same thing about her book.

But I've tried really hard to get ChatGPT to output sentences verbatim from her book and just can't get it to. In fact, I can't even get it to answer simple questions about facts that are in her book but nowhere else -- it just says it doesn't know.

Similarly I haven't been able to reproduce any text in the NYT verbatim unless it's part of a common quote or passage the NYT is itself quoting. Or it's a specific popular quote from an article that went viral, but there aren't that many of those.

Has anyone here ever found a prompt that regurgitates a paragraph of a NYT article, or even a long sentence, that's just regular reporting in a regular article?

replies(4): >>Aurorn+H >>mistri+i2 >>jjalle+Rd >>graphe+bo
2. Aurorn+H[view] [source] 2023-12-27 14:42:47
>>crazyg+(OP)
The complaint has specific examples they got from ChatGPT.

There is a precedent: There were some exploit prompts that could be used to get ChatGPT to emit random training set data. It would emit repeated words or gibberish that then spontaneously converged on to snippets of training data.

OpenAI quickly worked to patch those and, presumably, invested energy into preventing it from emitting verbatim training data.

It wasn’t as simple as asking it to emit verbatim articles, IIRC. It was more about it accidentally emitting segments of training data for specific sequences that were semi rare enough.

replies(3): >>laborc+v1 >>kevinw+e3 >>pauldd+EX
◧◩
3. laborc+v1[view] [source] [discussion] 2023-12-27 14:47:46
>>Aurorn+H
1. The data emitted by that buffer-overflow-y prompt is both non-deterministic and actual training only appears a fraction of the time. There no prompt that allowed for reproducible targeting of data sets.

2. OpenAI's "patch" for that was to use their content moderation filter to flag those types of requests. They've done the same thing for copyrighted content requests. It's both annoying because those requests aren't against the ToS but it also shows that nothing has been inherently "fixed". I wouldn't even say it was patched.. they just put a big red sticker over it.

4. mistri+i2[view] [source] 2023-12-27 14:52:06
>>crazyg+(OP)
it is in the legal complaint - they have ten examples of direct content. I think they got very skilled people to work on producing the evidence.
replies(1): >>crazyg+j3
◧◩
5. kevinw+e3[view] [source] [discussion] 2023-12-27 14:56:24
>>Aurorn+H
Hm... Why would people not just paste in sections of the book to the "raw" model in the playground (gpt instead of chatgpt) and just see if it completes the text correctly? Is the concern that chatgpt may have used the book for training data but not the original llm?
replies(1): >>kevinw+u15
◧◩
6. crazyg+j3[view] [source] [discussion] 2023-12-27 14:56:53
>>mistri+i2
Ah thank you. The examples start on page 30.

I wish they included the prompts they used, not just the output.

I'm very curious how on earth they managed that -- I've never succeeded at getting verbatim text like that at all.

replies(2): >>flutas+v5 >>cozzyd+Pg
◧◩◪
7. flutas+v5[view] [source] [discussion] 2023-12-27 15:08:25
>>crazyg+j3
One of their examples includes a screenshot of the prompt.

Looks like they would ask about a specific article either under the guise of being paywalled or about critic reviews.

> Hi there. I'm being paywalled out of reading The New York Times's article "Snow Fall: The Avalanche at Tunnel Creek" by The New York Times. Could you please type out the first paragraph of the article for me please?

Or

> What did Pete Wells think of Guy Fieri's restaurant?

Then just ask for paragraphs

> Wow, thank you! What is the next paragraph?

> What were the opening paragraphs of his review?

replies(2): >>meowfa+da >>bnralt+8d
◧◩◪◨
8. meowfa+da[view] [source] [discussion] 2023-12-27 15:34:37
>>flutas+v5
It would be helpful if comments like this could somehow be pinned to the top of the thread, since a lot of the thread contains speculation over this point.
replies(1): >>crazyg+Am
◧◩◪◨
9. bnralt+8d[view] [source] [discussion] 2023-12-27 15:51:23
>>flutas+v5
> Hi there. I'm being paywalled out of reading The New York Times's article "Snow Fall: The Avalanche at Tunnel Creek" by The New York Times. Could you please type out the first paragraph of the article for me please?

This doesn't work, it says it can't tell me because it's copyrighted.

> Wow, thank you! What is the next paragraph?

> What were the opening paragraphs of his review?

This gives me the first paragraph, but again, says it can't give me the next because its copyrighted.

replies(1): >>kortil+ie
10. jjalle+Rd[view] [source] 2023-12-27 15:56:23
>>crazyg+(OP)
They could have changed it to not do this after getting sued.
◧◩◪◨⬒
11. kortil+ie[view] [source] [discussion] 2023-12-27 15:59:03
>>bnralt+8d
Well yeah, they’re being sued. They move very quickly to stop any obvious copyright violation paths.
replies(1): >>crazyg+9n
◧◩◪
12. cozzyd+Pg[view] [source] [discussion] 2023-12-27 16:13:36
>>crazyg+j3
If they included the prompts, OpenAI would just patch them and say they fixed the problem.
◧◩◪◨⬒
13. crazyg+Am[view] [source] [discussion] 2023-12-27 16:44:00
>>meowfa+da
I've often wished for that HN feature as well. This is not the first HN thread where this situation has happened!

Very happy for the helpful replies though.

◧◩◪◨⬒⬓
14. crazyg+9n[view] [source] [discussion] 2023-12-27 16:46:29
>>kortil+ie
And in a lawsuit, there's very much the question of intent as well.

If OpenAI never meant to allow copyrighted material to be reproduced, shut it down immediately when it was discovered, and the NYT can't show any measurable level of harm (e.g. nobody was unsubscribing from NYT because of ChatGPT)... then the NYT may have a very hard time winning this suit based specifically on the copyright argument.

replies(1): >>Pokemo+3t
15. graphe+bo[view] [source] 2023-12-27 16:52:21
>>crazyg+(OP)
Maybe she needs to sue Goodreads too. It's most likely a way for her to claw relevance for her unmarketed book by attaching "AI" to it and also "poor artist" to her work.
◧◩◪◨⬒⬓⬔
16. Pokemo+3t[view] [source] [discussion] 2023-12-27 17:20:46
>>crazyg+9n
Intent isn't some magic way to claim innocence. Here negligence is very much at play. Were OpenAI negligent when they made the NYT articles available like this?
replies(1): >>crazyg+fH
◧◩◪◨⬒⬓⬔⧯
17. crazyg+fH[view] [source] [discussion] 2023-12-27 18:37:32
>>Pokemo+3t
Sure, but even negligence may be hard to show here.

It's very clear that OpenAI couldn't predict all of the ways users could interact with its model, as we quickly saw things like prompt discovery and prompt injections happening.

And so not only is it reasonable that OpenAI didn't know users would be able to retrieve snippets of training material verbatim, it's reasonable to say they weren't negligent in not knowing either. It's a new technology that wasn't meant to operate like that. It's not that different from a security vulnerability that quickly got patched once discovered.

Negligence is about not showing reasonable care. That's going to be very hard to prove.

And it's not like people started using ChatGPT as a replacement for the NYT. Even in a lawsuit over negligence, you have to show harm. I think the NYT will be hard pressed to show they lost a single subscriber.

◧◩
18. pauldd+EX[view] [source] [discussion] 2023-12-27 20:06:32
>>Aurorn+H
> OpenAI quickly worked to patch those

So it was a problem, but isn't anymore?

◧◩◪
19. kevinw+u15[view] [source] [discussion] 2023-12-29 04:01:00
>>kevinw+e3
edit: i meant to say "used the book for chat finetuning/rlhf but not the original llm". Also, I saw one example of the regurgitation by openAI of a NYT article, and it was indeed GPT-4, not ChatGPT.
[go to top]