zlacker

[parent] [thread] 17 comments
1. garden+(OP)[view] [source] 2023-11-18 08:28:47
Since ChatGPT is not indistinguishable from a human during a chat, is it fair to say it smashes the Turing test? Or do you mean something different?
replies(3): >>rayeig+Z4 >>NoOn3+a5 >>aidama+f5
2. rayeig+Z4[view] [source] 2023-11-18 09:11:32
>>garden+(OP)
Did you perhaps mean to say not distinguishable?
3. NoOn3+a5[view] [source] 2023-11-18 09:13:35
>>garden+(OP)
ChatGPT is distinguishable from a human, because ChatGPT never responds "I don't know.", at least not yet. :)
replies(5): >>ben_w+N5 >>NoOn3+Vd >>epolan+Uk >>raccoo+wm >>int_19+Un2
4. aidama+f5[view] [source] 2023-11-18 09:13:50
>>garden+(OP)
not yet: https://arxiv.org/abs/2310.20216

that being said, it is highly intelligent, capable of reasoning as well as a human, and passes IQ tests like GMAT and GRE at levels like the 97th percentile.

most people who talk about Chat GPT don't even realize that GPT 4 exists and is orders of magnitude more intelligent than the free version.

replies(2): >>jwestb+8c >>hedora+3i
◧◩
5. ben_w+N5[view] [source] [discussion] 2023-11-18 09:19:16
>>NoOn3+a5
It can do: https://chat.openai.com/share/f1c0726f-294d-447d-a3b3-f664dc...

IMO the main reason it's distinguishable is because it keeps explicitly telling you it's an AI.

replies(3): >>rezona+X6 >>NoOn3+x7 >>peigno+mt
◧◩◪
6. rezona+X6[view] [source] [discussion] 2023-11-18 09:28:55
>>ben_w+N5
This isn't the same thing. This is a commanded recital of a lack of capability, not that its confidence in it's answer is low. For a type of question the GPT _could_ answer, most of the time it _will_ answer, regardless of accuracy
◧◩◪
7. NoOn3+x7[view] [source] [discussion] 2023-11-18 09:34:19
>>ben_w+N5
I just noticed that when I ask really difficult technical questions, but for which there is an exact answer, It often tries to answer plausibly, but incorrectly instead of answering "I don't know". But over time, It becomes smarter and there are fewer and fewer such questions...
replies(2): >>ben_w+Q7 >>davegu+Ql2
◧◩◪◨
8. ben_w+Q7[view] [source] [discussion] 2023-11-18 09:37:17
>>NoOn3+x7
Have you tried setting a custom instruction in settings? I find that setting helps, albeit with weaker impact than the prompt itself.
replies(1): >>NoOn3+0i
◧◩
9. jwestb+8c[view] [source] [discussion] 2023-11-18 10:15:04
>>aidama+f5
Answers in Progress had a great video[0] where one of their presenters tested against an LLM in five different types of intelligence. tl;dr, AI was worlds ahead on two of the five, and worlds behind on the other three. Interesting stuff -- and clear that we're not as close to AGI as some of us might have thought earlier this year, but probably closer than a lot of the naysayers think.

0. https://www.youtube.com/watch?v=QrSCwxrLrRc

◧◩
10. NoOn3+Vd[view] [source] [discussion] 2023-11-18 10:28:50
>>NoOn3+a5
Maybe It's because It was never rewarded for such answers when It was learning.
◧◩◪◨⬒
11. NoOn3+0i[view] [source] [discussion] 2023-11-18 11:03:56
>>ben_w+Q7
It's not a problem for me. It's good that I can detect chatGPT by this sign.
◧◩
12. hedora+3i[view] [source] [discussion] 2023-11-18 11:04:18
>>aidama+f5
That’s just showing the tests are measuring specific things that LLMs can game particularly well.

Computers have been able to smash high school algebra tests since the 1970’s, but that doesn’t make them as smart as a 16 year old (or even a three year old).

◧◩
13. epolan+Uk[view] [source] [discussion] 2023-11-18 11:26:39
>>NoOn3+a5
Of course it does.
◧◩
14. raccoo+wm[view] [source] [discussion] 2023-11-18 11:38:21
>>NoOn3+a5
Some humans also never respond "I don't know" even when they don't know. I know people who out-hallucinate LLMs when pressed to think rigorously
◧◩◪
15. peigno+mt[view] [source] [discussion] 2023-11-18 12:25:06
>>ben_w+N5
I read an article where they did a proper Turing test and it seems people recognize it was a machine answering because it made no writing errors and wrote perfectly
replies(1): >>ben_w+qv
◧◩◪◨
16. ben_w+qv[view] [source] [discussion] 2023-11-18 12:40:03
>>peigno+mt
I've not read that, but I do remember hearing that the first human to fail the Turing test did so because they seemed to know far too much minutiae about Star Trek.
◧◩◪◨
17. davegu+Ql2[view] [source] [discussion] 2023-11-18 23:24:03
>>NoOn3+x7
It doesn't become smarter except for releases of new models. It's an inference engine.
◧◩
18. int_19+Un2[view] [source] [discussion] 2023-11-18 23:34:59
>>NoOn3+a5
It absolutely does that (GPT-4 especially), and I have hit it many times in regular conversations without specifically asking for it.
[go to top]