zlacker

[parent] [thread] 2 comments
1. gmerc+(OP)[view] [source] 2023-07-02 03:53:41
It’s quite ignorant to assume petabytes of garbage have any value at this point. See Chinchilla
replies(1): >>berkle+Rc
2. berkle+Rc[view] [source] 2023-07-02 06:36:13
>>gmerc+(OP)
I agree, but there are hundreds if not thousands of AI startups trying to make their own relevant LLM, and they're going to be scraping Twitter. The Onion called it many years ago [1]: "400 billion tweets and not one useful bit of data was ever transmitted".

[1] https://www.youtube.com/watch?v=cqggW08BWO0&t=138s

replies(1): >>rightb+Zn
◧◩
3. rightb+Zn[view] [source] [discussion] 2023-07-02 08:39:22
>>berkle+Rc
I can't imagine worse training data than e.g. Twitter and Reddit posts. How about like, dunno, books?

Edit: Ah, nvm, if you are trying to do a chat bot it is essentially what you want.

[go to top]