zlacker

[return to "OpenAI negotiations to reinstate Altman hit snag over board role"]
1. agnost+yS[view] [source] 2023-11-20 01:27:50
>>himara+(OP)
Many are speculating that Sam Altman could just move on and create another OpenAI 2.0 because he could easily attract talent and investors.

What this misses is all the regulatory capture that he’s been campaigning for. All the platforms have now closed their gardens. Authors and artists are much more vigilant about copyright etc. So it’s now a totally different game compared to 3 years ago because the data is not just there up for grabs anymore.

◧◩
2. Shekel+nV[view] [source] 2023-11-20 01:47:30
>>agnost+yS
I don't think getting training data is that hard still, the biggest platforms that locked down their APIs still use them for their mobile apps and can easily be reverse engineered to find keys or undocumented endpoints (or in the case of reddit, an entirely different internal API with less limits and a lot more info leaks...)
◧◩◪
3. bloqs+KV[view] [source] 2023-11-20 01:51:39
>>Shekel+nV
Can you explain the reddit one?
◧◩◪◨
4. 4death+uW[view] [source] 2023-11-20 01:55:57
>>bloqs+KV
Assuming the Reddit app does not use certificate pinning, you can use your computer to provide internet to your phone and then use an app like Charles Proxy to inspect requests being made from an app. Pretty easy to reverse engineer the API.

If the app does use certificate pinning, then you can use an Android phone and a modified app that removes the logic that enforces certificate pinning. This is more involved but also not impossible.

◧◩◪◨⬒
5. philis+WW[view] [source] 2023-11-20 01:59:20
>>4death+uW
That does not sound like the proper way to do an openAI 2.0. If Reddit ever hears that's how an AI company scraped them, they'll get sued for fun and profits.
◧◩◪◨⬒⬓
6. Shekel+o61[view] [source] 2023-11-20 03:10:31
>>philis+WW
It's essentially impossible to prove in court that training data was obtained or used improperly unless you go and tell on yourself. And even then it requires you to actually make someone with a lot of money mad, or to not have enough money yourself. Certainly microsoft would have already caught lots of flak for training their models on every github repo, instead they got a minor paddling from the public eye that went away after not much time had passed.
[go to top]