zlacker

[parent] [thread] 25 comments
1. photoc+(OP)[view] [source] 2022-12-11 18:54:13
Yes, ban it. I've been playing around with ChatGPT and where it starts failing is just where things start becoming interesting. What that means is that it's wikipedia-smart, i.e. it doesn't really tell you anything you can't find out with a minimal Google search. It does however cut the time-to-answer quite a bit, particularly if it's an area of knowledge you're not really that familiar with. But it bottoms out right as things start getting interesting, expertise wise.

Case example: I tried seeing what its limits on chemical knowledge were, starting with simple electron structures of molecules, and it does OK - remarkably, it got the advanced high-school level of methane's electronic structure right. It choked when it came to the molecular orbital picture and while it managed to list the differences between old-school hybrid orbitals and modern molecular orbitals, it couldn't really go into any interesting details about the molecular orbital structure of methane. Searching the web, I notice such details are mostly found in places like figures in research papers, not so much in text.

On the other hand, since I'm a neophyte when it comes to database architecture, it was great at answering what I'm sure any expert would consider basic questions.

Allowing comments sections to be clogged up with ChatGPT output would thus be like going to a restaurant that only served averaged-out mediocre but mostly-acceptable takes on recipes.

replies(2): >>scarfa+0e1 >>Jatama+gw1
2. scarfa+0e1[view] [source] 2022-12-12 04:30:35
>>photoc+(OP)
The problem with ChatGPT is that it often reads authoritative. But is often just flat out wrong.

I asked it a few questions for which I consider myself a subject matter expert and the answers were laughably wrong.

replies(7): >>63+pf1 >>toofy+pg1 >>photoc+Wg1 >>culanu+th1 >>oezi+Fs1 >>khyryk+1t1 >>tstrim+pB1
◧◩
3. 63+pf1[view] [source] [discussion] 2022-12-12 04:43:33
>>scarfa+0e1
Agreed. I chatted with it about the Chinese remainder theorem today and it gave me an example that didn't work and then insisted that 30 is not divisible by 2 when I questioned it. This was simple enough to spot but I was halfway through the example by the time I realized I couldn't trust it at all. Its confidence is annoying.
replies(1): >>kderby+XA1
◧◩
4. toofy+pg1[view] [source] [discussion] 2022-12-12 04:55:31
>>scarfa+0e1
i’ve seen so many examples of this over the past few weeks. just thinking about how many people will just eat what it feeds them borders on terrifying.

in so many instances, it’s just wrong but continues on so confidently.

one of the things i find most interesting is that it has no idea when it’s wrong so it just keeps going. we already have a fairly significant growing problem of people who refuse to admit (even to themselves) what they don’t know, and now this will just exacerbate the problem.

it’s like the worst of debateBro culture has just been automated.

replies(1): >>SoftTa+ei1
◧◩
5. photoc+Wg1[view] [source] [discussion] 2022-12-12 05:01:59
>>scarfa+0e1
I just asked it to provide a specific example of a security breach caused by a Java serialization flaw and it just made something up, i.e. it found some random security breach (unrelated to anything Java) and found some other random Java serialization bug and claimed that breach was due to that bug. A few minutes on Google revealed they were completely unrelated. Ouch.
replies(1): >>drusep+Pk1
◧◩
6. culanu+th1[view] [source] [discussion] 2022-12-12 05:07:24
>>scarfa+0e1
For me it happend when I asked to write a function using BigQuery, it wrote a function that made a lot of sense but was wrong, because the command didn't exist in BigQuery. When I replay that the function didn't work, it told me, something like this: You right the function that I used it was only working on beta mode, now you have to use the following.... And again it was wrong. I made a little research and never was such a beta commandm.. And then I got that it just makes up things that it don't know, but says it with authority.
replies(1): >>scarfa+hi1
◧◩◪
7. SoftTa+ei1[view] [source] [discussion] 2022-12-12 05:15:03
>>toofy+pg1
> in so many instances, it’s just wrong but continues on so confidently

Sounds sociopathic, and also like many politicians and people in leadership positions.

◧◩◪
8. scarfa+hi1[view] [source] [discussion] 2022-12-12 05:15:39
>>culanu+th1
I asked it to write a function in Python that would return the list of AWS accounts in an organization with a given tag key and value.

The code looked right, initialized boto3 correctly and called a function on it get_account_numbers_by_tag on the organizations object.

I wondered why I never heard of that function and nor did I find it when searching. Turns out, there is no such function.

replies(2): >>lhuser+Yj1 >>Radim+402
◧◩◪◨
9. lhuser+Yj1[view] [source] [discussion] 2022-12-12 05:33:36
>>scarfa+hi1
It sounds a lot similar to normal thinking errors that we make.
replies(1): >>scarfa+In1
◧◩◪
10. drusep+Pk1[view] [source] [discussion] 2022-12-12 05:41:23
>>photoc+Wg1
It seems like if the contextual analogy here was carried through and ChatGPT were to leave its message as a comment on a Java-security thread, the same approach would apply (a few minutes of research on Google) and the provably-incorrect message would either be downvoted or commented on, just like a human comment with the same content would be.
◧◩◪◨⬒
11. scarfa+In1[view] [source] [discussion] 2022-12-12 06:15:09
>>lhuser+Yj1
The second time, it gave me code that was almost right.

Just now I asked

Write a Python script that returns all of the accounts in an AWS organization with a given tag where the user specifies the tag key and value using command line arguments

I thought the code had to be wrong because it used concepts I had never heard of. This time it used the resource group API.

I have never heard of the API. But it does exist. I also couldn’t find sample code on the internet that did anything similar. But from looking at the documentation it should work. I learned something new today.

BTW, for context when I claimed to be a “subject matter expert” above, I work at AWS in Professional Services, code most days using the AWS API and I would have never thought of the solution it gave me.

◧◩
12. oezi+Fs1[view] [source] [discussion] 2022-12-12 07:05:33
>>scarfa+0e1
I guess after cookie banners this is going to be the next frontier of regulation: prohibiting AIs to lie to humans.
◧◩
13. khyryk+1t1[view] [source] [discussion] 2022-12-12 07:09:34
>>scarfa+0e1
My biggest fear for the short term is that tools like ChatGPT would allow spamming most of the internet with the equivalent of a Gish Gallop -- so much plausible-looking bullshit spewed out in a short time that it would be a lost cause to attempt to sort through it.
replies(1): >>nonran+mG1
14. Jatama+gw1[view] [source] 2022-12-12 07:43:55
>>photoc+(OP)
> if it's an area of knowledge you're not really that familiar with

Thats actually dangerous way to use ChatGPT. Since you don't know the real answer you won't be able to tell when it gets something wrong.

replies(3): >>city17+1B1 >>myster+F42 >>Zachsa+xc2
◧◩◪
15. kderby+XA1[view] [source] [discussion] 2022-12-12 08:31:21
>>63+pf1
It told me the longest known prime number ended with a 2. Definitely not accurate when it comes to math.
replies(1): >>tobtah+zc4
◧◩
16. city17+1B1[view] [source] [discussion] 2022-12-12 08:32:21
>>Jatama+gw1
But if you know a lot about something, why would you ask ChatGPT a question about it (especially if you assume it doesn't have expert knowledge)?
replies(1): >>system+vC1
◧◩
17. tstrim+pB1[view] [source] [discussion] 2022-12-12 08:35:18
>>scarfa+0e1
> The problem with ChatGPT is that it often reads authoritative. But is often just flat out wrong.

That sounds pretty damn human to me.

replies(1): >>tobtah+Ud4
◧◩◪
18. system+vC1[view] [source] [discussion] 2022-12-12 08:45:40
>>city17+1B1
I wouldn't ask ChatGPT anything. It is still writing weird articles that sounds meaningful yet lacking arguments because it finds attributes of the compared objects and places them in sentences. As if it is comparing them. It just doesn't make sense. ChatGPT is nice but has a long way to become useful that way.
◧◩◪
19. nonran+mG1[view] [source] [discussion] 2022-12-12 09:21:02
>>khyryk+1t1
If content known to be of human origin could be archived and frozen now in late 2022 it may become valuable in a few years. Some kind of verifiably timestamped encryption might be useful.
◧◩◪◨
20. Radim+402[view] [source] [discussion] 2022-12-12 12:14:00
>>scarfa+hi1
It gives the old saying "The reasonable man adapts himself to the world; the unreasonable man adapts the world to himself; therefore all progress depends on the unreasonable man." a new twist, doesn't it?

1. AN AI MODEL IS GIVEN ENOUGH CAPACITY to capture (some of) our human perspective, a snapshot of our world as reflected in its training data. <== We've been here for a while

2. AN AI MODEL IS GIVEN ENOUGH CAPACITY to fabulate and imagine things. <== We're unambiguously here now

The fabulations are of a charmingly naive "predict the most probable next token" sort for now, with chatGPT. But even as a future model is (inevitably) given the ability to probe and correct its errors, the initial direction of its fabulations will still reflect that "inception worldview" snapshot.

For example, if a particular fashion trend or political view was popular around the time the model was trained (with training data typically skewing toward the "recent", simply because "recent" is when most digital data will have been produced), that model can be expected to fabulate along the lines of that imprinted political view.

3. AN AI MODEL IS GIVEN ENOUGH CAPACITY to make the is-vs-ought choice between "CORRECT ITSELF" = adapt to the world; or "CORRECT THE WORLD" = imprint its worldview back onto the world (probably indirectly through humans paying attention to its outputs and acting as actuators, but that makes no difference). <== We're getting there rapidly

Will it be more reasonable or unreasonable?

And which mode wins out long-term, be more energy efficient in that entropic struggle for survival that all physical systems go through?

replies(1): >>scarfa+E82
◧◩
21. myster+F42[view] [source] [discussion] 2022-12-12 12:54:58
>>Jatama+gw1
Honestly this could be the silver lining of ChatGPT. Some people trust the answers of random commenters on the internet for anything from how things work technically to medical advice. Having an ever-increasing chance that any given commenter literally knows nothing except how to string words together might break that habit.
replies(1): >>krageo+Dg2
◧◩◪◨⬒
22. scarfa+E82[view] [source] [discussion] 2022-12-12 13:26:07
>>Radim+402
I am not sure if this is AI generated or meant to read like it is. But I’ll bite.

One thing I noticed, it’s either trained naturally or tweaked by humans not to be political or say anything controversial. I asked if a simple question “Does open door have a good business model”. It punted like any good politician.

◧◩
23. Zachsa+xc2[view] [source] [discussion] 2022-12-12 13:56:58
>>Jatama+gw1
I've been using GTP3's copilot when designing sql queries. I'm not very comfortable with sql, but I am with the mathematical basis. It's powerful tool to help learn language syntax.

I've experimented with systems design using it, but as I expected, it's a big fat no.

If a robot gtp account does not have human supervision, it will spit out all sorts of rubbish / be easy to spot. Else the manager will just be a person who spams low quality content. I'm concerned, but we have time to find a solution.

◧◩◪
24. krageo+Dg2[view] [source] [discussion] 2022-12-12 14:24:11
>>myster+F42
> Having an ever-increasing chance that any given commenter literally knows nothing except how to string words together might break that habit.

This is already figuratively the case and it has had no impact on this phenomenon. Why would the new situation be any different?

◧◩◪◨
25. tobtah+zc4[view] [source] [discussion] 2022-12-12 23:58:41
>>kderby+XA1
ChatGPT is not a calculator, it's a language model.
◧◩◪
26. tobtah+Ud4[view] [source] [discussion] 2022-12-13 00:04:26
>>tstrim+pB1
yeah, tough to distinguish human BS, from an incorrect ChatGPT.
[go to top]