zlacker

Language is not essential for the cognitive processes that underlie thought

submitted by orcul+(OP) on 2024-10-17 12:10:30 | 562 points 404 comments
[view article] [source] [go to bottom]

NOTE: showing posts with links only show all posts
2. fsndz+Qm5[view] [source] 2024-10-19 18:17:25
>>orcul+(OP)
more proof that we need more than LLMs to build LRMs: https://www.lycee.ai/blog/drop-o1-preview-try-this-alternati...
11. m463+zo5[view] [source] 2024-10-19 18:31:55
>>orcul+(OP)
I like Temple Grandin's "Thinking the Way Animals Do":

https://www.grandin.com/references/thinking.animals.html

17. airstr+5q5[view] [source] 2024-10-19 18:45:23
>>orcul+(OP)
https://archive.is/PsUeX
◧◩◪◨
42. youoy+cu5[view] [source] [discussion] 2024-10-19 19:18:18
>>kachnu+sr5
It's a well known phenomenon! I will drop this link here in case you are not familiar with it:

https://www.sciencealert.com/theres-a-big-difference-in-how-...

45. Animat+Du5[view] [source] 2024-10-19 19:21:01
>>orcul+(OP)
This is an important result.

The actual paper [1] says that functional MRI (which is measuring which parts of the brain are active by sensing blood flow) indicates that different brain hardware is used for non-language and language functions. This has been suspected for years, but now there's an experimental result.

What this tells us for AI is that we need something else besides LLMs. It's not clear what that something else is. But, as the paper mentions, the low-end mammals and the corvids lack language but have some substantial problem-solving capability. That's seen down at squirrel and crow size, where the brains are tiny. So if someone figures out to do this, it will probably take less hardware than an LLM.

This is the next big piece we need for AI. No idea how to do this, but it's the right question to work on.

[1] https://www.nature.com/articles/s41586-024-07522-w.epdf?shar...

◧◩◪
56. card_z+Ry5[view] [source] [discussion] 2024-10-19 19:53:00
>>HarHar+gw5
Not sure neuron number correlates to smarts, either.

https://www.scientificamerican.com/article/gut-second-brain/

There are 100 million in my gut, but it doesn't solve any problems that aren't about poop, as far as I know.

https://en.wikipedia.org/wiki/List_of_animals_by_number_of_n...

If the suspiciously round number is accurate, this puts the human gut somewhere between a golden hamster and ansell's mole-rat, and about level with a short-palated fruit bat.

◧◩◪
63. psycho+sB5[view] [source] [discussion] 2024-10-19 20:14:21
>>GavinM+Qv5
It's also doubtful that thinking about the concept of analytical thought processes is something most humans do either, at least not in these terms and this perspective.

Should we expect experts in cognitive science exposing their view in a scientific publication to stick to the narrowest median view of language though? All the more when in the same article you quote people like Russell who certainly didn't have a naïve definition of language when expressing a point of view on the matter.

And slapping in general can definitely communicate far more than a single thing depending on many parameters. See https://www.33rdsquare.com/is-a-slap-disrespectful-a-nuanced... for a text exploring some of nuances of the meaning it can encompasse. But even a kid can get that slap could perfectly have all the potential to create a fully doubly articulated language, as The Croods 2 creators funnily have put in scene. :D

◧◩◪
71. psycho+UE5[view] [source] [discussion] 2024-10-19 20:42:39
>>throwa+qw5
Not all human languages exhibits recursion though: https://en.wikipedia.org/wiki/Pirah%C3%A3_language

And recursion as the unique trait for human language differentiation is not necessarily completely consensual https://omseeth.github.io/blog/2024/recursive_language/

Also, let's recall that in its broader meaning, the scientific consensus is that humans are animals and they evolved through the same basic mechanism as all other life forms that is evolution. So even assuming that evolution made some unique language hability emerge in humans, it's most likely that they share most language traits with other species and that there is more things to learn from them that what would be possible if it's assumed they can't have a language and thoughts.

◧◩◪
72. Animat+ZE5[view] [source] [discussion] 2024-10-19 20:44:10
>>KoolKa+aC5
I don't know what we need. Nor does anybody else, yet. But we know what it has to do. Basically what a small mammal or a corvid does.

There's been progress. Look at this 2020 work on neural net controlled drone acrobatics.[1] That's going in the right direction.

[1] https://rpg.ifi.uzh.ch/docs/RSS20_Kaufmann.pdf

96. mmooss+aP5[view] [source] 2024-10-19 22:28:03
>>orcul+(OP)
A concept in every human culture - i.e., created in every culture, not passed from one to some others - is mentalese [0]: "A universal non-verbal system of concepts, etc., conceived of as an innate representational system resembling language, which is the medium of thought and underlies the ability to learn and use a language." [1]

If you look up 'mentalese' you can find a bunch written about it. There's an in-depth article by Daniel Gregory and Peter Langland-Hassan, in the incredible Stanford Encyclopedia of Philosophy, on Inner Speech (admittedly, I'm taking a leap to think they mean precisely the same thing). [2]

[0] Steven Pinker, The Blank Slate: The Modern Denial of Human Nature (2002)

[1] Oxford English Dictionary

[2] https://plato.stanford.edu/entries/inner-speech/

◧◩
119. toomuc+IU5[view] [source] [discussion] 2024-10-19 23:26:02
>>orwin+BU5
Also known as “Flow”.

https://en.wikipedia.org/wiki/Flow_(psychology)

◧◩◪◨
135. jebark+hZ5[view] [source] [discussion] 2024-10-20 00:21:20
>>awongh+xT5
I think the data is way more important for the success of LLMs than the architecture although I do think there's something important in the GPT architecture in particular. See this talk for why: [1]

Warning, watch out for waving hands: The way I see it is that cognition involves forming an abstract representation of the world and then reasoning about that representation. It seems obvious that non-human animals do this without language. So it seems likely that humans do too and then language is layered on top as a turbo boost. However, it also seems plausible that you could build an abstract representation of the world through studying a vast amount of human language and that'll be a good approximation of the real-world too and furthermore it seems possible that reasoning about that abstract representation can take place in the depths of the layers of a large transformer. So it's not clear to me that we're limited by the data we have or necessarily need a different type of data to build a general AI although that'll likely help build a better world model. It's also not clear that an LLM is incapable of the type of reasoning that animals apply to their abstract world representations.

[1] https://youtu.be/yBL7J0kgldU?si=38Jjw_dgxCxhiu7R

◧◩◪
166. nox101+G66[view] [source] [discussion] 2024-10-20 02:08:01
>>CSMast+W16
It sounds like you think this research is wrong? (it claims llms can not reason)

https://arstechnica.com/ai/2024/10/llms-cant-perform-genuine...

or do you maybe think no logical reasoning is needed to do everything a human can do? Tho humans seem to be able to do logical reasoning

172. Tagber+B86[view] [source] 2024-10-20 02:36:49
>>orcul+(OP)
I’ve been hearing/reading about people who don’t have an inner monologue. Their experience of cognition is not verbally-based.

https://www.cbc.ca/news/canada/saskatchewan/inner-monologue-...

◧◩◪◨
173. dleeft+P86[view] [source] [discussion] 2024-10-20 02:40:49
>>alseri+O16
Some progress has been made in this area, see [0], [1], [2] and [3], observing both similarities and dissimilarities in terms of language processing:

Siegmund, J., Kästner, C., Apel, S., Parnin, C., Bethmann, A., Leich, T. & Brechmann, A. (2014). Understanding understanding source code with functional magnetic resonance imaging. In Proceedings of the 36th International Conference on Software Engineering (pp. 378-389).

Peitek, N., Siegmund, J., Apel, S., Kästner, C., Parnin, C., Bethmann, A. & Brechmann, A. (2018). A look into programmers’ heads. IEEE Transactions on Software Engineering, 46(4), 442-462.

Krueger, R., Huang, Y., Liu, X., Santander, T., Weimer, W., & Leach, K. (2020). Neurological divide: An fMRI study of prose and code writing. In Proceedings of the ACM/IEEE 42nd International Conference on Software Engineering (pp. 678-690).

Peitek, N., Apel, S., Parnin, C., Brechmann, A. & Siegmund, J. (2021). Program comprehension and code complexity metrics: An fmri study. In 2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE) (pp. 524-536). IEEE.

[0]: https://www.frontiersin.org/10.3389/conf.fninf.2014.18.00040...

[1]: https://ieeexplore.ieee.org/abstract/document/8425769

[2]: https://dl.acm.org/doi/abs/10.1145/3377811.3380348

[3]: https://ieeexplore.ieee.org/abstract/document/9402005

◧◩◪◨
179. Animat+bb6[view] [source] [discussion] 2024-10-20 03:25:44
>>awongh+xT5
> One reason might that LLMs are successful because of the architecture, but also, just as importantly because they can be trained over a volume and diversity of human thought that’s encapsulated in language (that is on the internet). Where are we going to find the equivalent data set that will train this other kind of thinking?

Probably by putting simulated animals into simulated environments where they have to survive and thrive.

Working at animal level is uncool, but necessary for progress. I had this argument with Rod Brooks a few decades back. He had some good artificial insects, and wanted to immediately jump to human level, with a project called Cog.[1] I asked him why he didn't go for mouse level AI next. He said "Because I don't want to go down in history as the inventor of the world's greatest artificial mouse."

Cog was a dud, and Brooks goes down in history as the inventor of the world's first good robotic vacuum cleaner.

[1] https://en.wikipedia.org/wiki/Cog_(project)

184. ryandv+9c6[view] [source] 2024-10-20 03:43:15
>>orcul+(OP)
It's worth noting the precise and narrow sense in which the term "language" is used throughout these studies: it is those particular "word sequences" that activate particular regions in the brain's left hemisphere, to the exclusion of other forms of symbolic representation such as mathematical notation. Indeed, in two of the studies cited, [0] [1] subjects with language deficits or brain lesions in areas associated with the "language network" are asked to perform on various mathematical tasks involving algebraic expressions [0] or Arabic numerals [1]:

> DA was impaired in solving simple addition, subtraction, division or multiplication problems, but could correctly simplify abstract expressions such as (b×a)÷(a×b) or (a+b)+(b+a) and make correct judgements whether abstract algebraic equations like b − a = a − b or (d÷c)+a=(d+a)÷(c+a) were true or false.

> Sensitivity to the structural properties of numerical expressions was also evaluated with bracket problems, some requiring the computation of a set of expressions with embedded brackets: for example, 90  [(3  17)  3].

Discussions of whether or not these sorts of algebraic or numerical expressions constitute a "language of mathematics" aside (despite them not engaging the same brain regions and structures associated with the word "language"); it may be the case that these sorts of word sequences and symbols processed by structures in the brain's left hemisphere are not essential for thought, but can still serve as a useful psychotechnology or "bicycle of the mind" to accelerate and leverage its innate capabilities. In a similar fashion to how this sort of mathematical notation allows for more concise and precise expression of mathematical objects (contrast "the number that is thrice of three and seventeen less of ninety") and serves to amplify our mathematical capacities, language can perhaps be seen as a force multiplier; I have doubts whether those suffering from aphasia or an agrammatic condition would be able to rise to the heights of cognitive performance.

[0] https://pubmed.ncbi.nlm.nih.gov/17306848/

[1] https://pubmed.ncbi.nlm.nih.gov/15713804/

◧◩◪
186. FL33TW+uc6[view] [source] [discussion] 2024-10-20 03:50:52
>>HarHar+gw5
It's probably more relevant to compare intraspecies rather than interspecies.

And it turns out that human brain volume and intelligence are moderately-highly correlated [1][2]!

[1]: https://pmc.ncbi.nlm.nih.gov/articles/PMC7440690/ [2]: https://www.sciencedirect.com/science/article/abs/pii/S01602...

◧◩◪
200. dleeft+mf6[view] [source] [discussion] 2024-10-20 04:39:06
>>GavinM+Qv5
I'm not sure it's that fringe. Popular addages such as 'language is a vehicle for thought' and 'the pen is mightier than the sword' reveal that language is sometimes implied to be tool-like, with many of our unspoken acts carrying linguistic meaning (e.g. ghosting, not answering a call, sign language, gesturing, nodding, etc.).

Even tools present us a certain 'language', talking to us via beeps, blinks and buzzes, and are having increasingly interesting discussions amongst themselves (e.g. subreddit simulator, agent based modeling). Recent philosophers of technology as Mark Coeckelbergh present a comprehensive argument for why we need to move away from the tool/language barrier [0], and has been part in informing the EC Expert Group on AI [1].

[0]: https://www.taylorfrancis.com/books/mono/10.4324/97813155285...

[1]: https://philtech.univie.ac.at/news/news-about-publicatons-et...

201. upghos+Tf6[view] [source] 2024-10-20 04:48:44
>>orcul+(OP)
Well this comment is about the article not LLMs so I doubt it will have much in the way of legs, but this work has already been covered extensively and to a fascinating depth by Jaak Panksepp [1].

His work explores the neuropsychology of emotions WAIT DON'T GO they are actually the substrate of consciousness, NOT the other way around.

We have 7 primary affective processes (measurable hardware level emotions) and they are not what you think[2]. They are considered primary because they are sublinguistic. For instance, witnessing the color red is a primary experience, you cannot explain in words the color red to someone who has not ever seen it before.

His work is a really fascinating read if you ever want to take a break from puters for a minute and learn how people work.

PS the reason this sort of research isn't more widely known is because the behaviorist school was so incredibly dominant since the 1970s they made it completely taboo to discuss subjective experience in the realm of scientific discourse. In fact the emotions we are usually taught are not based on emotional states but on muscle contractions in the face! Not being allowed to talk about emotions in psychological studies or the inner process of the mind is kinda crazy when you think about it. So only recently with neuroimaging has it suddenly become ok to acknowledge that things happen in the brain independent of externally observable behavior.

[1] https://a.co/d/6EYULdP

[2] - seeking - fear - anxiety and grief - rage - lust - play!!! - caring

[3] if this sounds familiar at all it's because Jordan Peterson cites Jaak Panksep all the time. Well 50% of the time, the other 50% is CG Jung and the final 50% is the book of Exodus for some reason.

◧◩◪
231. senand+kq6[view] [source] [discussion] 2024-10-20 07:21:06
>>CSMast+W16
This seems quite reasonable, but I recently heard a podcast (https://www.preposterousuniverse.com/podcast/2024/06/24/280-...) that LLMs are more likely to be very good at navigating what they have been trained on, but very poor at abstract reasoning and discovering new areas outside of their training. As a single human, you don't notice, as the training material is greater than everything we could ever learn.

After all, that's what Artificial General Intelligence would at least in part be about: finding and proving new math theorems, creating new poetry, making new scientific discoveries, etc.

There is even a new challenge that's been proposed: https://arcprize.org/blog/launch

> It makes sense that the process of thinking and the process of translating those thoughts into and out of language would be distinct

Yes, indeed. And LLMs seem to be very good at _simulating_ the translation of thought into language. They don't actually do it, at least not like humans do.

◧◩◪◨⬒⬓⬔
249. jstanl+iw6[view] [source] [discussion] 2024-10-20 08:42:56
>>adrian+Qn6
But conscious experience does produce observable effects.

For that not to be the case, you'd have to take the position that humans experience consciousness and they talk about consciousness but that there is no causal link between the two! It's just a coincidence that the things you find yourself saying about consciousness line up with your internal experience?

https://www.lesswrong.com/posts/fdEWWr8St59bXLbQr/zombies-zo...

◧◩
271. ryandv+EH6[view] [source] [discussion] 2024-10-20 11:31:39
>>bmacho+aC6
Julian Jaynes has written on this verbal/non-verbal dichotomy in The Origin of Consciousness in the Breakdown of the Bicameral Mind, in which he literally defines god to mean those phenomena related to right hemispheric structures and activities in the brain that are communicated over the anterior commissures and interpreted by left hemispheric language centers as speech; hence the many mystical reports of "hearing the voice of god" as passed down through the aeons. Such phenomena have gone by many other names: gods, the genius, the higher self, the HGA... though this metaphysical and spiritual terminology is best understood as referring to non-verbal, non-rational, non-linear forms of cognition that are closer to free association and intuitive pattern matching (similar to Kahneman's "System 1" thinking). There even exist certain mystical traditions which purport to facilitate deeper connections with this subsystem of the mind; see for instance Eshelman's accounting of the western esoteric tradition in The Mystical and Magical System of the A.'.A.'. at [0] (currently defunct pending the restoration of the Internet Archive).

[0] https://archive.org/details/a-a-the-mystical-and-magical-sys...

◧◩◪◨⬒⬓⬔⧯▣
295. mannyk+FT6[view] [source] [discussion] 2024-10-20 13:57:47
>>adrian+1x6
The p-zombie argument is the best-known of a group of conceivability arguments, which ultimately depend on the notion that if a proposition is conceivably true, then there is a metaphysically possible world in which it is true. Skeptics suppose that this is just a complicated way of equivocating over what 'conceivable' means, and even David Chalmers, the philosopher who has done the most to bring the p-zombie argument to wide attention, acknowledges that it depends on the assumption of what he calls 'perfect conceivability', which is tantamount to irrefutable knowledge.

To deal with the awkwardly apparent fact that consciousness certainly seems to have physical effects, zombiephiles challenge the notion that physics is causally closed, so that it is conceivable that something non-physical can cause physical effects. Their approach is to say that the causal closure of physics is not provable, but at this point, the argument has become a lexicographical one, about the definition of the words 'physics' and 'physical' (if one insists that 'physical' does not refer to a causally-closed concept, then we still need a word for the causal closure within which the physical is embedded - but that's just what a lot of people take 'physical' to mean in the first place.) None of the anti-physicalists have been able, so far, to shed any light on how the mind is causally effective in the physical world.

You might be interested in the late Daniel Dennett's "The Unimagined Preposterousness of Zombies": https://dl.tufts.edu/concern/pdfs/6m312182x

◧◩
297. upghos+eV6[view] [source] [discussion] 2024-10-20 14:11:40
>>YeGobl+fZ5
I've been trying to figure out to respond to this for a while. I appreciate the fact that you are pretty much the lone voice on this thread voicing this opinion, which I also share but tend to keep my mouth shut since it seems to be unpopular.

It's hard for me to understand where my peers are coming from on the other side of this argument and respond without being dismissive, so I'll do my best to steelman the argument later.

Machine learning models are function approximators and by definition do not have an internal experience distinct from the training data any more than the plus operator does. I agree with the sentiment that even putting it in writing gives more weight to the position than it should, bordering on absurdity.

I suppose this is like the ELIZA phenomena on steroids, is the only thing I can think of for why such notions are being entertained.

However, to be generous, lets do some vigorous hand waving and say we could find a way to have an embodied learning agent gather sublinguistic perceptual data in an online reinforcement learning process, and furthermore that the (by definition) non-quantifiable subjective experience data could somehow be extracted, made into a training set, and fit to a nicely parametric loss function.

The idea then is that could find some architecture that would allow you to fit a model to the data.

And voila, machine consciousness, right? A perfect model for sentience.

Except for the fact that you would need to ignore that in the RL model gathering the data and the NN distilled from it, even with all of our vigorous hand waving, you are once again developing function approximators that have no subjective internal experience distinct from the training data.

Let's take it one step further. The absolute simplest form of learning comes in the form of habituation and sensitization to stimuli. Even microbes have the ability to do this.

LLMs and other static networks do not. You can attempt to attack this point by fiatting online reinforcement learning or dismissing it as unnecessary, but I should again point out that you would be attacking or dismissing the bare minimum requirement for learning, let alone a higher order subjective internal experience.

So then the argument, proceeding from false premises, would claim that the compressed experience in the NN could contain mechanical equivalents of higher order internal subjective experiences.

So even with all the might vigorous hand waving we have allowed, you have at best found a way to convert internal subjective processes to external mechanical processes fit to a dataset.

The argument would then follow, well, what's the difference? And I could point back to the microbe, but if the argument hasn't connected by this point, we will be chasing our tails forever.

A good book on the topic that examines this in much greater depth is "The Self Assembling Brain".

https://a.co/d/1FwYxaJ

That being said, I am hella jealous of the VC money that the grifters will get for advancing the other side of this argument.

For enough money I'd probably change my tune too. I can't by a loaf of bread with a good argument lol

◧◩◪
299. ryandv+EW6[view] [source] [discussion] 2024-10-20 14:28:32
>>zeroxf+BR6
This sounds similar to a fairly early realization in the practice of meditation. Daniel M Ingram refers to it as "Cause and Effect" in Mastering the Core Teachings of the Buddha: [0]

> In the stage of Cause and Effect, the relationships between mental and physical phenomena become very clear and sometimes ratchet-like. There is a cause, such as intention, and then an effect, such as movement. There is a cause, such as a sensation, and there is an effect, namely a mental impression.

Trying to increase the frequency at which you oscillate between physical sensations and mental sensations is a fascinating exercise.

[0] https://www.mctb.org/mctb2/table-of-contents/part-iv-insight...

◧◩
316. Goblin+O57[view] [source] [discussion] 2024-10-20 15:51:52
>>farts_+Bu5
Thought is observable https://www.biorxiv.org/content/10.1101/2021.02.02.429430v1
◧◩◪◨⬒⬓⬔
317. mannyk+Z57[view] [source] [discussion] 2024-10-20 15:52:53
>>bbor+yV6
I feel you are putting too much emphasis on the importance and primacy of having a definition of words like 'reasoning'.

As humanity has struggled to understand the world, it has frequently given names to concepts that seem to matter, well before it is capable of explaining with any sort of precision what these things are, and what makes them matter - take the word 'energy', for example.

It seems clear to me that one must have these vague concepts before one can begin to to understand them, and also that it would be bizarre not to give them a name at that point - and so, at that point, we have a word without a locked-down definition. To insist that we should have the definition locked down before we begin to investigate the phenomenon or concept is precisely the wrong way to go about understanding it: we refine and rewrite the definitions as a consequence of what our investigations have discovered. Again, 'energy' provides a useful case study for how this happens.

A third point about the word 'energy' is that it has become well-defined within physics, and yet retains much of its original vagueness in everyday usage, where, in addition, it is often used metaphorically. This is not a problem, except when someone makes the lexicographical fallacy of thinking that one can freely substitute the physics definition into everyday speech (or vice-versa) without changing the meaning.

With many concepts about the mental, including 'reasoning', we are still in the learning-and-writing-the-definition stage. For example, let's take the definition you bring up: reasoning as good cognition. This just moves us on to the questions of what 'cognition' means, and what distinguishes good cognition from bad cognition (for example, is a valid logical argument predicated on what turns out to be a false assumption an example of reasoning-as-good-cognition?) We are not going to settle the matter by leafing through a dictionary, any more than Pedro Carolino could write a phrase book just from a Portugese-English dictionary (and you are probably aware that looking up definitions-of-definitions recursively in a dictionary often ends up in a loop.)

A lot of people want to jump the gun on this, and say definitively either that LLMs have achieved reasoning (or general intelligence or a theory of mind or even consciousness, for that matter) or that they have not (or cannot.) What we should be doing, IMHO, is to put aside these questions until we have learned enough to say more precisely what these terms denote, by studying humans, other animals, and what I consider to be the surprising effectiveness of LLMs - and that is what the interviewee in the article we are nominally discussing here is doing.

You entered this thread by saying (about the paper underlying an article in Ars Tech [1]) I’ll pop in with a friendly “that research is definitely wrong”. If they want to prove that LLMs can’t reason..., but I do not think there is anything like that claim in the paper itself (one should not simply trust what some person on HN says about a paper. That, of course, goes as much for what I say about it as what the original poster said.) To me, this looks like the sort of careful, specific and objective work that will lead to us a better understanding of our concepts of the mental.

[1] https://arxiv.org/pdf/2410.05229

◧◩
324. ericho+5b7[view] [source] [discussion] 2024-10-20 16:27:41
>>Animat+Du5
> So if someone figures out to do this, it will probably take less hardware than an LLM.

We have, it's called DreamCoder. There's a paper and everything.

Everything needed for AGI exists today, people simply have (incorrect) legacy beliefs about cognition that are holding them back (e.g. "humans are rational").

https://arxiv.org/abs/2006.08381

334. Geee+Sf7[view] [source] 2024-10-20 17:01:23
>>orcul+(OP)
The important question is: what is considered a language?

> You can ask whether people who have these severe language impairments can perform tasks that require thinking. You can ask them to solve some math problems or to perform a social reasoning test, and all of the instructions, of course, have to be nonverbal because they can’t understand linguistic information anymore.

Surely these "non-verbal instructions" are some kind of language. Maybe all human action can be considered language.

A contrarian example to this research might be feral children, i.e people who have been raised away from humans.[0] In most cases they are mentally impaired; as in not having human-like intelligence. I don't think there is a good explanation why this happens to humans. And why it doesn't happen to other animals, which develop normally in species-typical way whether they are in the wild or in human captivity. It seems that most human behavior (even high-level intelligence) is learned / copied from other humans, and maybe this copied behavior can be considered language.

If humans are "copy machines", there's also a risk of completely losing the "what's it like to be a human" behavior if children of the future are raised by AI and algorithmic feeds.

[0] https://en.wikipedia.org/wiki/Feral_child

◧◩◪◨⬒
350. haswel+vr7[view] [source] [discussion] 2024-10-20 18:37:33
>>ninety+Fq6
Success doesn’t imply that “reasoning” was involved, and the definition of reasoning is extremely important.

Apple’s recent research summarized here [0] is worth a read. In short, they argue that what LLMs are doing is more akin to advanced pattern recognition than reasoning in the way we typically understand reasoning.

By way of analogy, memorizing mathematical facts and then correctly recalling these facts does not imply that the person actually understands how to arrive at the answer. This is why “show your work” is a critical aspect of proving competence in an education environment.

An LLM providing useful/correct results only proves that it’s good at surfacing relevant information based on a given prompt. That fact that it’s trivial to cause bad results by making minor but irrelevant changes to a prompt points to something other than a truly reasoned response, i.e. a reasoning machine would not get tripped up so easily.

- [0] https://x.com/MFarajtabar/status/1844456880971858028

◧◩◪◨
352. ryandv+iu7[view] [source] [discussion] 2024-10-20 19:00:29
>>psycho+aq7
The paper itself: https://www.nature.com/articles/s41586-024-07522-w.epdf?shar...
◧◩◪◨⬒⬓
357. ninety+Xz7[view] [source] [discussion] 2024-10-20 19:52:53
>>haswel+vr7
You’re still suffering from the biases of the parent poster. You are picking and choosing papers that illustrate failure instances when there are also an equal amount of papers that verify successful instances.

It’s bloody obvious that when I classify success I mean that the llm is delivering a correct and unique answer for a novel prompt that doesn’t exist in the original training set. No need to go over the same tired analogies that have been regurgitated over and over again that you believe LLMs are reusing memorized answers. It’s a stale point of view. The overall argument has progressed further then that and we now need more complicated analysis of what’s going on with LLMs

Sources: https://typeset.io/papers/llmsense-harnessing-llms-for-high-...

https://typeset.io/papers/call-me-when-necessary-llms-can-ef...

And these two are just from a random google search.

I can find dozens and dozens of papers illustrating failures and successes of LLMs which further nails my original point. LLMs both succeed and fail at reasoning.

The main problem right now is that we don’t really understand how LLMs work internally. Everyone who claims they know LLMs can’t reason are just making huge leaps of irrational conclusions because not only does their conclusion contradict actual evidence but they don’t even know how LLMs work because nobody knows.

We only know how LLMs work at a high level and we only understand these things via the analogy of a best fit curve in a series of data points. Below this abstraction we don’t understand what’s going on.

◧◩◪
370. Izkata+WO7[view] [source] [discussion] 2024-10-20 22:17:50
>>usgrou+v97
"Language" is a subset of "symbols". I agree with what you said, but it's not representative of the quote in GP.

Just a few days ago was "What do you visualize while programming?", and there's a few of us in the comments that, when programming, think symbolically without language: >>41869237

◧◩
389. westur+Mt9[view] [source] [discussion] 2024-10-21 15:39:00
>>Animat+Du5
"Language models can explain neurons in language models" https://news.ycombinator.com/item?id=35877402#35886145 :

> Recent work has revealed that the neural activity patterns correlated with sensation, cognition, and action often are not stable and instead undergo large scale changes over days and weeks—a phenomenon called representational drift.

[...]

So, I'm not sure how conclusive this fmri activation study is either.

Though, is there a proto language that's not even necessary for the given measured aspects of condition?

Which artificial network architecture best approximates which functionally specialized biological neutral networks?

OpenCogPrime:KnowledgeRepresentation > Four Types of Knowledge: https://wiki.opencog.org/w/OpenCogPrime:KnowledgeRepresentat... :

> Sensory, Procedural, Episodic, Declarative

From https://news.ycombinator.com/item?id=40105068#40107537 re: cognitive hierarchy and specialization :

> But FWIU none of these models of cognitive hierarchy or instruction are informed by newer developments in topological study of neural connectivity;

◧◩◪◨⬒⬓⬔
392. JumpCr+2E9[view] [source] [discussion] 2024-10-21 16:48:16
>>calf+L58
> analog models cannot exceed the computational power of Turing machines

There is no reason to assume consciousness is Turing computable [1].

[1] https://en.m.wikipedia.org/wiki/Church%E2%80%93Turing_thesis

◧◩◪◨⬒⬓⬔⧯
404. no_ide+gqk[view] [source] [discussion] 2024-10-25 17:22:36
>>JumpCr+2E9
Good thing Computability Beyond Church-Turing via Choice Sequences[1] exists.

[1] Mark Bickford, Liron Cohen, Robert L. Constable, and Vincent Rahli. 2018. Computability Beyond Church-Turing via Choice Sequences. In Proceedings of the 33rd Annual ACM/IEEE Symposium on Logic in Computer Science (LICS '18). Association for Computing Machinery, New York, NY, USA, 245–254. https://doi.org/10.1145/3209108.3209200

[go to top]