zlacker

[return to "Memory and new controls for ChatGPT"]
1. anothe+Pf[view] [source] 2024-02-13 19:29:04
>>Josely+(OP)
This is a bit off topic to the actual article, but I see a lot of top ranking comments complaining that ChatGPT has become lazy at coding. I wanted to make two observations:

1. Yes, GPT-4 Turbo is quantitatively getting lazier at coding. I benchmarked the last 2 updates to GPT-4 Turbo, and it got lazier each time.

2. For coding, asking GPT-4 Turbo to emit code changes as unified diffs causes a 3X reduction in lazy coding.

Here are some articles that discuss these topics in much more detail.

https://aider.chat/docs/unified-diffs.html

https://aider.chat/docs/benchmarks-0125.html

◧◩
2. omalle+6q[view] [source] 2024-02-13 20:21:14
>>anothe+Pf
Can you say in one or two sentences what you mean by “lazy at coding” in this context?
◧◩◪
3. Me1000+Mq[view] [source] 2024-02-13 20:24:55
>>omalle+6q
It has a tendency to do:

"// ... the rest of your code goes here"

in it's responses, rather than writing it all out.

◧◩◪◨
4. asaddh+Ds[view] [source] 2024-02-13 20:36:32
>>Me1000+Mq
It's incredibly lazy. I've tried to coax it into returning the full code and it will claim to follow the instructions while regurgitating the same output you complained about. GPT-4 was great, GPT-4 Turbo first version was pretty terrible bordering on unusable, then they came out with the Turbo second version, which almost feels worse to me, though I haven't compared, but if someone comes claiming they fixed an issue, but you still see it, it will bias you to see it more.

Claude is doing much better in this area, local/open LLMs are getting quite good, it feels like OpenAI is not heading in a good direction here, and I hope they course correct.

◧◩◪◨⬒
5. mister+PA[view] [source] 2024-02-13 21:21:44
>>asaddh+Ds
I have a feeling full powered LLM's are reserved for the more equal animals.

I hope some people remember and document details of this era, future generations may be so impressed with future reality that they may not even think to question it's fidelity, if that concept even exists in the future.

◧◩◪◨⬒⬓
6. bbor+tE[view] [source] 2024-02-13 21:43:28
>>mister+PA
…could you clarify? Is this about “LLMs can be biased, thus making fake news a bigger problem”?
◧◩◪◨⬒⬓⬔
7. mister+XM[view] [source] 2024-02-13 22:32:41
>>bbor+tE
I confidently predict that we sheep will not have access to the same power our shepherds will have.
[go to top]