zlacker

[parent] [thread] 6 comments
1. jazzyj+(OP)[view] [source] 2023-05-16 14:55:22
your comment reads to me as totally disconnected to the OP, whose concern relates to using the appearance of intelligence as a scare tactic to build a regulatory moat.
replies(1): >>adamsm+63
2. adamsm+63[view] [source] 2023-05-16 15:11:06
>>jazzyj+(OP)
Actually OP is clearly, ironically, parroting the stochastic parrot idea that LLMs are incapable of anything other than basic token prediction and dismissing any of their other emergent abilities.
replies(3): >>woeiru+Xb >>jazzyj+Yz >>srslac+tT1
◧◩
3. woeiru+Xb[view] [source] [discussion] 2023-05-16 15:46:38
>>adamsm+63
Spoiler alert: they're actually both LLMs arguing with one another.
◧◩
4. jazzyj+Yz[view] [source] [discussion] 2023-05-16 17:22:43
>>adamsm+63
yea but that's a boring critique and not the point they were making - whether or not LLMs reason or parrot has no relevance to whether Mr Altman should be the one building the moat.
◧◩
5. srslac+tT1[view] [source] [discussion] 2023-05-17 01:00:52
>>adamsm+63
It can't generalize and adapt outside of its corpus, not in a way that's correct anyhow, and there's nothing "emergent." They are incapable of anything other than token prediction on its corpus and context. it just produces really good predictions. Funny how everyone keeps citing that Microsoft paper, when Microsoft is who is lobbying for this regulatory capture, and it's already been shown that such emergence on the tasks they chose when you scale up was a "mirage."
replies(2): >>comp_t+192 >>adamsm+0m3
◧◩◪
6. comp_t+192[view] [source] [discussion] 2023-05-17 03:36:27
>>srslac+tT1
Yes, and neither could GPT-3, which is why we don't observe any differences between GPT-3 and GPT-4. Right?

Tell me: how does this claim _constrain my expectations_ about what this (or future) models can do? Is there a specific thing that you predicted in advance that GPT-4 would be unable to do, which ended up being a correct prediction? Is there a specific thing you want to predict in advance of the next generation, that it will be unable to do?

◧◩◪
7. adamsm+0m3[view] [source] [discussion] 2023-05-17 14:25:45
>>srslac+tT1
This is demonstrably wrong. It can clearly generate unique text not from it's training corpus and can successfully answer logic based questions that were also not in it's training corpus.

Another paper not from Msft showing emergent task capabilities across a variety of LLMs as scale increases.

https://arxiv.org/pdf/2206.07682.pdf

You can hem and haw all you want but the reality is these models have internal representations of the world that can be probed via prompts. They are not stochastic parrots no matter how much you shout in the wind that they are.

[go to top]