You're leaving out the essentials. These models do more than fitting the data given. They can output it in a variety of ways, and through their approximation, can synthesize data as well. They can output things that weren't in the original data, tailored to a specific request in the tiniest of fractions of the time it would take a normal person to look up and understand that information.
Your argument is almost like saying "give me your RSA keys, because it's just two prime numbers, and I know how to list them."
Do we want to go down the road of making white collar jobs the legislatively required elevator attendants? Instead of just banning AI in general via executive agency?
That sounds like a better solution to me, actually. OpenAI's lobbyists would never go for that though. Can't have a moat that way.
Unless you think OpenAI is blatantly lying about this:
"A.1 Sourcing. We sourced either the most recent publicly-available official past exams, or practice exams in published third-party 2022-2023 study material which we purchased. We cross-checked these materials against the model’s training data to determine the extent to which the training data was not contaminated with any exam questions, which we also report in this paper."
"As can be seen in tables 9 and 10, contamination overall has very little effect on the reported results."
They also report results on uncontaminated data which shows basically no statistical difference.
I'm saying that the "intelligence" is specialized, not generalized and adaptable.
It's an approximated function. We're talking about regression based function approximation. This is a model of language.
"Emergent behavior", when it's not just a mirage of wishful researchers and if it even exists, is only a side effect of the regression based function approximation to generate a structure that encapsulates all substantive chains of words (a model).
We then guide the model further towards a narrow portion of the language latent space that aligns with our perception of intelligent behavior.
It can't translate whale song, or an extraterrestrial language, though it may opine on how to do so.
The underpinning technology of language models holds more importance than general and adaptable intelligence. It holds more importance than something that is going to, or is capable of, escaping the box and killing us all. It functions as a universal induction machine, capable of modeling - and "comprehending" - the latent structure within any form of signal.
The output of that function approximation though, is simply a model. A specialized intelligence. A non-adaptable intelligence, outside of its corpus. Outside of the data that it "fits."
The approximated function does not magically step outside of its box. Nor is it capable. It fits the data.
Ok guys pack it up, LLM's can't be intelligent because they can't translate Whale Song. GG.
I mean of all the AI Goalposts to be moved this one really takes the cake.
It's a human language calculator. You're imparting magical qualities of general understanding to regression based function approximation. They "fit" the data. It's not generalizable, nor adaptable. But that's why they're powerful, the ability to bias them towards that subset of language. No one said it's not an amazing technology, and no one said it was a stochastic parrot. I'm saying that it's fitting the data, and is not, and cannot, be a general or adaptable intelligence.