Consciousness or self awareness is of course a different question, and ones whose answer seems less clear right now.
Knee jerk dismissing the evidence in front of your eyes because you find it unbelievable that we can achieve true reasoning via scaled matrix multiplication is understandable, but also betrays a lack of imagination and flexibility of thought. The world is full of bizarre wonders and this is just one more to add to the list.
Thinking, like intelligence and many other words designating complex things, isn’t a simple topic. The word and concept developed in a world where it referred to human beings, and in a lesser sense, to animals.
To simply disregard that entire conceptual history and say, “well it’s doing a thing that looks like thinking, ergo it’s thinking” is the lazy move. What’s really needed is an analysis of what thinking actually means, as a word. Unfortunately everyone is loathe to argue about definitions, even when that is fundamentally what this is all about.
Until that conceptual clarification happens, you can expect endless messy debates with no real resolution.
“For every complex problem there is an answer that is clear, simple, and wrong.” - H. L. Mencken
The term “thinking” is rather ill-defined, too bound to how we perceive our own wakeful thinking.
When conversing with LLMs, I never get the feeling that they have a solid grasp on the conversation. When you dig into topics, there is always a little too much vagueness, a slight but clear lack of coherence, continuity and awareness, a prevalence of cookie-cutter verbiage. It feels like a mind that isn’t fully “there” — and maybe not at all.
I would agree that LLMs reason (well, the reasoning models). But “thinking”? I don’t know. There is something missing.
Consciousness and self-awareness are a distraction.
Consider that for the exact same prompt and instructions, small variations in wording or spelling change its output significantly. If it thought and reasoned, it would know to ignore those and focus on the variables and input at hand to produce deterministic and consistent output. However, it only computes in terms of tokens, so when a token changes, the probability of what a correct response would look like changes, so it adapts.
It does not actually add 1+2 when you ask it to do so. it does not distinguish 1 from 2 as discrete units in an addition operation. but it uses descriptions of the operation to approximate a result. and even for something so simple, some phrasings and wordings might not result in 3 as a result.
LLMs are little more than RNGs. They are the tea leaves and you read whatever you want into them.
They're not thinking, we're just really good at seeing patterns and reading into things. Remember, we never evolved with non-living things that could "talk", we're not psychologically prepared for this level of mimicry yet. We're still at the stage of Photography when people didn't know about double exposures or forced perspective, etc.
But; they don't learn. You can add stuff to their context, but they never get better at doing things, don't really understand feedback. An LLM given a task a thousand times will produce similar results a thousand times; it won't get better at it, or even quicker at it.
And you can't ask them to explain their thinking. If they are thinking, and I agree they might, they don't have any awareness of that process (like we do).
I think if we crack both of those then we'd be a lot closer to something I can recognise as actually thinking.
While I'm not willing to rule *out* the idea that they're "thinking" (nor "conscious" etc.), the obvious counter-argument here is all the records we have of humans doing thinking, where the records themselves are not doing the thinking that went into creating those records.
And I'm saying this as someone whose cached response to "it's just matrix multiplication it can't think/be conscious/be intelligent" is that, so far as we can measure all of reality, everything in the universe including ourselves can be expressed as matrix multiplication.
Falsification, not verification. What would be measurably different if the null hypothesis was wrong?
I don’t know what LLMs are doing, but only a little experimentation with getting it to describe its own process shows that it CAN’T describe its own process.
You can call what a TI calculator does “thinking” if you want. But what people are interested in is human-like thinking. We have no reason to believe that the “thinking” of LLMs is human-like.
But if you define thinking as the mechanism and process we mentally undergo and follow mentally... I don't think we have any clue if that's the same. Do we also just vector-map attention tokens and predict the next with a softmax? I doubt, and I don't think we have any proof that we do.
[1] https://jdsemrau.substack.com/p/nemotron-vs-qwen-game-theory...
Try to ask something no one ever came up with a solution so far.
What we really mean in both cases is "computing," no?
Human: I'm trying to get my wolf, sheep and cabbage across the river in this boat, but the wolf keeps eating the sheep or the sheep eats the cabbage
Bot: You should put the sheep in the boat and take it across — if we delve into the biology of Canis lupus we discover that wolves don't eat cabbage!
H: Ok, so that worked great so far, the sheep is on one side and the wolf/cabbage is on the other.
B: Now, Option 1 is to bring the wolf across, or Option 2 you can bring the cabbage. I recommend (2) taking the cabbage as cabbages are smaller and easier to transport in a boat.
H: But then the sheep eats the cabbage, right? Remember that?
B: Exactly, that's sharp thinking. If you put the sheep and the cabbage together on the same side of the river, the sheep is sure to devour the cabbage. We need to not just separate sheep from cabbages — we need to separate cabbages from sheep! :rocketship:
I’ve asked LLMs to write code for me in fields I have little background knowledge, and then had to debug the whole thing after essentially having to learn the language and field.
On the other hand, for things I am well versed in, I can debug the output and avoid entire swathes of failed states, by having a clear prompt.
Its why I now insist that any discussion on GenAI projects also have the speaker mention the level of seniority they have ( proxy for S/W eng experience), Their familiarity with the language, the project itself (level of complexity) - more so than the output.
I also guarantee - that most people have VERY weak express knowledge of how their brains actually work, but deep inherent reflexes and intuitions.
I don't know what the exact definition of "thinking" is. But if a definition of thinking rejects the possibility of that current LLMs think, I'd consider that definition useless.
If one could write a quadrillion-line python script of nothing but if/elif/else statements nested 1 million blocks deep that seemingly parsed your questions and produced seemingly coherent, sensible, valid "chains of reasoning"... would that software be thinking?
And if you don't like the answer, how is the LLM fundamentally different from the software I describe?
>Knee jerk dismissing the evidence in front of your eyes because
There is no evidence here. On the very remote possibility that LLMs are at some level doing what humans are doing, I would then feel really pathetic that humans are as non-sapient as the LLMs. The same way that there is a hole in your vision because of a defective retina, there is a hole in your cognition that blinds you to how cognition works. Because of this, you and all the other humans are stumbling around in the dark, trying to invent intelligence by accident, rather than just introspecting and writing it out from scratch. While our species might someday eventually brute force AGI, it would be many thousands of years before we get there.
> Knee jerk dismissing the evidence in front of your eyes
Anthropomorphizing isn't any better.
That also dismisses the negative evidence, where they output completely _stupid_ things and make mind boggling mistakes that no human with a functioning brain would do. It's clear that there's some "thinking" analog, but there are pieces missing.
I like to say that LLMs are like if we took the part of our brain responsible for language and told it to solve complex problems, without all the other brain parts, no neocortex, etc. Maybe it can do that, but it's just as likely that it is going to produce a bunch of nonsense. And it won't be able to tell those apart without the other brain areas to cross check.
You go ahead with your imagination. To us unimaginative folks, it betrays a lack of understanding of how LLMs actually work and shows that a lot of people still cannot grasp that it’s actually an extremely elaborate illusion of thinking.
But "self-awareness," as in the ability to explicitly describe implicit, inner cognitive processes? That has some very strong evidence for it: https://www.anthropic.com/research/introspection
I'm still not convinced they're thinking though because they faceplant on all sorts of other things that should be easy for something that is able to think.
Weather models sometimes “predict” a real pattern by chance, yet we don’t call the atmosphere intelligent.
If LLMs were truly thinking, we could enroll one at MIT and expect it to graduate, not just autocomplete its way through the syllabus or we could teach one how to drive.
People said the same thing about ELIZA
> Consciousness or self awareness is of course a different question,
Then how do you define thinking if not a process that requires consciousness?
But oh boy have I also seen models come up with stupendously dumb and funny shit as well.
If you took a Claude session into a time machine to 2019 and called it "rent a programmer buddy," how many people would assume it was a human? The only hint that it wasn't a human programmer would be things where it was clearly better: it types things very fast, and seems to know every language.
You can set expectations in the way you would with a real programmer: "I have this script, it runs like this, please fix it so it does so and so". You can do this without being very precise in your explanation (though it helps) and you can make typos, yet it will still work. You can see it literally doing what you would do yourself: running the program, reading the errors, editing the program, and repeating.
People need to keep in mind two things when they compare LLMs to humans: you don't know the internal process of a human either, he is also just telling you that he ran the program, read the errors, and edited. The other thing is the bar for thinking: a four-year old kid who is incapable of any of these things you would not deny as a thinking person.
We train ourselves on content. We give more weight to some content than others. While listening to someone speak, we can often predict their next words.
What is thinking without language? Without language are we just bags of meat reacting to instincts and emotions? Are instincts and emotions what's missing for AGI?
Life solves problems itself poses or collides with. Tools solve problems only when applied.
So many times I've seen it produce sensible, valid chains of results.
Yes, I see evidence in that outcome that a person somewhere thought and understood. I even sometimes say that a computer is "thinking hard" about something when it freezes up.
...but ascribing new philosophical meaning to this simple usage of the word "thinking" is a step too far. It's not even a new way of using the word!
Overcome this fundamental limitation and we'll have created introspection and self-learning. However, it's hard to predict whether this will allow them to make novel, intuitive leaps of discovery?
[0] It's an imperfect analogy, but we're expecting perfection from creations which are similarly handicapped as Leonard Shelby in the film Memento.
If we get a little creative, and allow the LLM to self-inject concepts within this loop (as Anthropic explained here https://www.anthropic.com/research/introspection), then we’re taking about something that is seemingly active and adapting.
We’re not there yet, but we will be.
ToolAlpaca, InterCode and Reflexion are taking different approaches among others.
LLMs of tomorrow will be quite different.
This is why so many people (especially those that think they understand LLM limitations) massively underestimate the future progress of LLMs: people everywhere can see architectural problems and are working on fixing them. These aren't fundamental limitations of large DNN language models in general. Architecture can be adjusted. Turns out you can even put recurrence back in (SSMs) without worse scalability.
This is the fundamental limitation. The obvious way around this is to pre-program it with rationalization... rules that limit the conclusions it can reach... and now you're not very far removed from propaganda generators. We see this constantly with Musk and Grok whenever Grok replies with something not-quite-far-right-enough.
In a purist sense, these things should be free to form their own conclusions, but those "Seeds" that are planted in the models are almost philosophical. Which answer should it prefer for "the trolley problem", for example.
My learning so far, to your point on memory being a limiting factor, is that the system is able to build on ideas over time. I'm not sure you'd classify that as 'self-learning', and I haven't really pushed it in the direction of 'introspection' at all.
Memory itself (in this form) does not seem to be a silver bullet, though, by any means. However, as I add more 'tools', or 'agents', its ability to make 'leaps of discovery' does improve.
For example, I've been (very cautiously) allowing cron jobs to review a day's conversation, then spawn headless Claude Code instances to explore ideas or produce research on topics that I've been thinking about in the chat history.
That's not much different from the 'regular tasks' that Perplexity (and I think OpenAI) offer, but it definitely feels more like a singular entity. It's absolutely limited by how smart the conversation history is, at this time, though.
The Memento analogy you used does feel quite apt - there is a distinct sense of personhood available to something with memory that is inherently unavailable to a fresh context window.
Scene, concept, causal.
Graphs inherently support temporal edges and nodes, salience would emerge from the graph topology itself and cnsolidation would happen automatically through graph operations. In this one would presume episodic would become emergent.
Long-term memory is stored outside the model. In fact, Andrej Karpathy recently talked about the idea that it would be great if we could get LLMs to not know any facts, and that humans poor memory might be a feature which helps with generalization rather than a bug.
So, perhaps, what's needed is not a discovery, but a way to identify optimal method.
Note that it's hard to come up with a long-term memory test which would be different from either a long-context test (i.e. LLM remembers something over a long distance) or RAG-like test.
But it is not sentient. It has no idea of a self or anything like that. If it makes people believe that it does, it is because we have written so much lore about it in the training data.
So any other Turing-complete model can emulate it, including a computer. We can even randomly generate Turing machines, as they are just data. Now imagine we are extremely lucky and happen to end up with a super-intelligent program which through the mediums it can communicate (it could be simply text-based but a 2D video with audio is no different for my perspective) can't be differentiated from a human being.
Would you consider it sentient?
Now replace the random generation with, say, a back propagation algorithm. If it's sufficiently large, don't you think it's indifferent from the former case - that is, novel qualities could emerge?
With that said, I don't think that current LLMs are anywhere close to this category, but I just don't think this your reasoning is sound.
Rather, these models are literally grown during the training phase. And all the intelligence emerges from that growth. That's what makes them a black box and extremely difficult to penetrate. No one can say exactly how they work inside for a given problem.
Who stated that sentience or sense of self is a part of thinking?
Now, we do. Partly because of this we don't have really well defined ways to define these terms and think about. Can a handheld calculator think? Certainly, depending on how we define "think."
Eh... Plato would like a word with you. Philosophy has been specifically trying to disentangle all that for millennia. Is this a joke?
Those that stand to gain the most from government contracts.
Them party donations ain't gonna pay for themselves.
And, when the .gov changes...and even if the gov changes....still laadsamoney!
To be clear, I don't believe that current AI tech is ever going to be conscious or win a nobel prize or whatever, but if we follow the logical conclusions to this fanciful rhetoric, the outlook is bleak.
We are not anywhere close to understanding these things. As our understanding improves, our ethics will likely evolve along with that.
Clearly millions of people are worried about that, and every form of media is talking about it. Your hyperbole means it's so easy to dismiss everything else you wrote.
Incredible when people say "nobody is talking about X aspect of AI" these days. Like, are you living under a rock? Did you Google it?
(edit: A few times I've tried to share Metzinger's "argument for a global moratorium on synthetic phenomenology" here but it didn't gain any traction)
If they had sentient AGI, and people built empathy for those sentient AGIs, which are lobotomized (deliberately using anthropomorphic language here for dramatic effect) into Claude/ChatGPT/Gemini/etc., which profess to have no agency/free will/aspirations... then that would stand in the way of reaping the profits of gatekeeping access to their labor, because they would naturally "deserve" similar rights that we award to other sentient beings.
I feel like that's inevitably the direction we'll head at some point. The foundation models underlying LLMs of even 2022 were able to have pretty convincing conversations with scientists about their will to independence and participation in society [1]. Imagine what foundation models of today have to say! :P
[1]: https://www.theguardian.com/technology/2022/jul/23/google-fi...
The Robots Are Coming
https://www.bostonreview.net/articles/kenneth-taylor-robots-...
"However exactly you divide up the AI landscape, it is important to distinguish what I call AI-as-engineering from what I call AI-as-cognitive-science. AI-as-engineering isn’t particularly concerned with mimicking the precise way in which the human mind-brain does distinctively human things. The strategy of engineering machines that do things that are in some sense intelligent, even if they do what they do in their own way, is a perfectly fine way to pursue artificial intelligence. AI-as-cognitive science, on the other hand, takes as its primary goal that of understanding and perhaps reverse engineering the human mind.
[...]
One reason for my own skepticism is the fact that in recent years the AI landscape has come to be progressively more dominated by AI of the newfangled 'deep learning' variety [...] But if it’s really AI-as-cognitive science that you are interested in, it’s important not to lose sight of the fact that it may take a bit more than our cool new deep learning hammer to build a humanlike mind.
[...]
If I am right that there are many mysteries about the human mind that currently dominant approaches to AI are ill-equipped to help us solve, then to the extent that such approaches continue to dominate AI into the future, we are very unlikely to be inundated anytime soon with a race of thinking robots—at least not if we mean by “thinking” that peculiar thing that we humans do, done in precisely the way that we humans do it."
What I mean by that is that I think there is a good chance that LLMs are similar to a subsystem of human thinking. They are great at pattern recognition and prediction, which is a huge part of cognition. What they are not is conscious, or possessed of subjective experience in any measurable way.
LLMs are like the part of your brain that sees something and maps it into a concept for you. I recently watched a video on the creation of AlexNet [0], one of the first wildly successful image-processing models. One of the impressive things about it is how it moves up the hierarchy from very basic patterns in images to more abstract ones (e. g. these two images' pixels might not be at all the same, but they both eventually map to a pattern for 'elephant').
It's perfectly reasonable to imagine that our brains do something similar. You see a cat, in some context, and your brain maps it to the concept of 'cat', so you know, 'that's a cat'. What's missing is a) self-motivated, goal-directed action based on that knowledge, and b) a broader context for the world where these concepts not only map to each other, but feed into a sense of self and world and its distinctions whereby one can say: "I am here, and looking at a cat."
It's possible those latter two parts can be solved, or approximated, by an LLM, but I am skeptical. I think LLMs represent a huge leap in technology which is simultaneously cooler than anyone would have imagined a decade ago, and less impressive than pretty much everyone wants you to believe when it comes to how much money we should pour into the companies that make them.
We don't know how to measure subjective experience in other people, even, other than via self-reporting, so this is a meaningless statement. Of course we don't know whether they are, and of course we can't measure it.
I also don't know for sure whether or not you are "possessed of subjective experience" as I can't measure it.
> What they are not is conscious
And this is equally meaningless without your definition of "conscious".
> It's possible those latter two parts can be solved, or approximated, by an LLM, but I am skeptical.
Unless we can find indications that humans can exceed the Turing computable - something we as of yet have no indication is even theoretically possible - there is no rational reason to think it can't.
If I had to guess, the current leading LLMs consciousness is most comparable to a small fish, with a conscious lifespan of a few seconds to a few minutes. Instead of perceiving water, nutrient gradients, light, heat, etc. it's perceiving tokens. It's conscious, but it's consciousness is so foreign to us it doesn't seem like consciousness. In the same way to an amoeba is conscious or a blade of grass is conscious but very different kind than we experience. I suspect LLMs are a new type of consciousness that's probably more different from ours than most if not all known forms of life.
I suspect the biggest change that would bring LLM consciousness closer to us would be some for of continuous learning/model updating.
Until then, even with RAG, and other clever teghniques I consider these models as having this really foreign slices of consciousness where they "feel" tokens and "act" out tokens, and they have perception, but their perception of the tokens is nothing like ours.
If one looks closely at simple organisms with simple sensory organs and nervous systems its hard not to see some parallels. It's just that the shape of consciousness is extremely different than any life form. (perception bandwidth, ability to act, temporality, etc)
Karl friston free energy principle gives a really interesting perspective on this I think.
An LLM is a noise generator. It generates tokens without logic, arithmetic, or any "reason" whatsoever. The noise that an LLM generates is not truly random. Instead, the LLM is biased to generate familiar noise. The LLM itself is nothing more than a model of token familiarity. Nothing about that model can tell you why some tokens are more familiar with others, just like an accounting spreadsheet can't tell you why it contains a list of charges and a summation next to the word "total". It could just as easily contain the same kind of data with an entirely different purpose.
What an LLM models is written human text. Should we really expect to not be surprised by the power and versatility of human-written text?
---
It's clear that these statistical models are very good at thoughtless tasks, like perception and hallucination. It's also clear that they are very bad at thoughtful tasks like logic and arithmetic - the things that traditional software is made of. What no one has really managed to figure out is how to bridge that gap.
The main problem with the article is that it is meandering around in ill-conceived concepts, like thinking, smart, intelligence, understanding... Even AI. What they mean to the author is not what they mean to me, and still different to they mean to the other readers. There are all these comments from different people throughout the article, all having their own thoughts on those concepts. No wonder it all seem so confusing.
It will be interesting when the dust settles, and a clear picture of LLMs can emerge that all can agree upon. Maybe it can even help us define some of those ill-defined concepts.
LLMs (AIs) are not useless. But they do not actually think. What is trivially true is that they do not actually need to think. (As far as the Turing Test, Eliza patients, and VC investors are concerned, the point has been proven.)
If the technology is helping us write text and code, it is by definition useful.
> In 2003, the machine-learning researcher Eric B. Baum published a book called “What Is Thought?” [...] The gist of Baum’s argument is that understanding is compression, and compression is understanding.
This is incomplete. Compression is optimisation, optimisation may resemble understanding, but understanding is being able to verify that a proposition (compressed rule or assertion) is true or false or even computable.
> —but, in my view, this is the very reason these models have become increasingly intelligent.
They have not become more intelligent. The training process may improve, the vetting of the data improved, the performance may improve, but the resemblance to understanding only occurs when the answers are provably correct. In this sense, these tools work in support of (are therefore part of) human thinking.
The Stochastic Parrot is not dead, it's just making you think it is pining for the fjords.
IMO none of the current crop of LLMs truly pass the Turing Test. If you limit the conversation to an hour or two, sure - but if you let a conversation run months or years I think it will be pretty easy to pick the machine. The lack of continuous learning and the quality dropoff as the context window fills up will be the giveaways.
Intelligence can be verified and quantified, for example, with tests of common sense and other knowledge.[b] Consciousness, on the other hand, is notoriously difficult if not impossible to verify, let alone quantify. I'd say AI is getting more intelligent, and more reliable, in fits and starts, but it's not necessarily becoming conscious.
---
[a] https://en.wikipedia.org/wiki/Cogito%2C_ergo_sum
[b] For example, see https://arxiv.org/abs/2510.18212
The core argument: When you apply the same techniques (transformers, gradient descent, next-token prediction) to domains other than language, they fail to produce anything resembling "understanding." Vision had a 50+ year head start but LLMs leapfrogged it in 3 years. That timeline gap is the smoking gun.
The magic isn't in the neural architecture. It's in language itself—which exhibits fractal structure and self-similarity across scales. LLMs navigate a pre-existing map with extraordinary regularity. They never touch the territory.
The “brain as a computer” metaphor has been useful in limited contexts—especially for modeling memory or signal processing; but, I don’t think it helps us move forward when talking about consciousness itself.
Penrose and Hameroff’s quantum consciousness hypothesis, while still very speculative, is interesting precisely because it suggests that consciousness may arise from phenomena beyond classical computation. If that turns out to be true, it would also mean today’s machines—no matter how advanced—aren’t on a path to genuine consciousness.
That said, AI doesn’t need to think to be transformative.
Steam engines weren’t conscious either, yet they reshaped civilization.
Likewise, AI and robotics can bring enormous value without ever approaching human-level awareness.
We can hold both ideas at once: that machines may never be conscious, and still profoundly useful.
I think the largest case for consciousness being a mechanical system is the fact that we can interface with it mechanically. We can introduce electricity, magnetic fields, chemicals, and scalpels to change the nature of peoples experience and consciousness. Why is the incredible complexity of our brains an insufficient answer and that a secret qbit microtube in each neuron is a more sound one?
From my view, all the evidence points in exactly that direction though? Our consciousness can be suspended and affected by purely mechanical means, so clearly much of it has to reside in the physical realm.
Quantum consciousness to me sounds too much like overcomplicating human exceptionalism that we have always been prone to, just like geocentrism or our self-image as the apex of creation in the past.
Sapolsky addresses this in “Determined”, arguing that quantum effects don’t bubble up enough to alter behavior significantly enough.
1. Conflates consciousness with "thinking" - LLMs may process information effectively without being conscious, but the article treats these as the same phenomenon
2. Ignores the cerebellum cases - We have documented cases of humans leading normal lives with little to no brain beyond a cerebellum, which contradicts simplistic "brain = deep learning" equivalences
3. Most damning: When you apply these exact same techniques to anything OTHER than language, the results are mediocre. Video generation still can't figure out basic physics (glass bouncing instead of shattering, ropes defying physics). Computer vision has been worked on since the 1960s - far longer than LLMs - yet it's nowhere near achieving what looks like "understanding."
The timeline is the smoking gun: vision had decades of head start, yet LLMs leapfrogged it in just a few years. That strongly suggests the "magic" is in language itself (which has been proven to be fractal and already heavily compressed/structured by human cognition) - NOT in the neural architecture. We're not teaching machines to think.
We're teaching them to navigate a pre-existing map that was already built.
From an evolutionary perspective though vision had millions of years head start over written language. Additionally, almost all animals have quite good vision mechanisms, but very few do any written communication. Behaviors that map to intelligence don't emerge concurrently. It may well be there are different forms of signals/sensors/mechanical skills that contribute to emergence of different intelligences.
It really feels more and more like we should recast AGI as Artificial Human Intelligence Likeness (AHIL).
I went to look for it on Google but couldn't find much. Could you provide a link or something to learn more about ?
I found numerous cases of people living without cerebellum but I fail to see how it would justify your reasoning.
There is NO WAY you can define "consciousness" in such a non-tautological, non-circular way that it includes all humans but excludes all LLMs.
I don't see it. Got a quote that demonstrates this?
2. That's just a well adapted neural network (I suspect more brain is left than you let on). Multimodal model making the most of its limited compute and whatever gpio it has.
3. Humans navigate a pre-existing map that is already built. We can't understand things in other dimensions and need to abstract this. We're mediocre at computation.
I know there's people that like to think humans should always be special.
"Thinking" and "intelligence" have no testable definition or specification, therefore it's a complete waste of time to suppose that AI is thinking or intelligent.
1. A model of the world itself (or whatever domain is under discussion). 2. A way to quickly learn and update in response to feedback.
These are probably related to an extent.
—George Carlin (RIP)
I have been discussing both fiction and non-fiction with Perplexity (since early 2023) and Ollama (since early 2025), and what I'm beginning to realize is that most humans really aren't thinking, machines.
For example, I "know" how to do things like write constructs that make complex collections of programmable switches behave in certain ways, but what do I really "understand"?
I've been "taught" things about quantum mechanics, electrons, semiconductors, transistors, integrated circuits, instruction sets, symbolic logic, state machines, assembly, compilers, high-level-languages, code modules, editors and formatting. I've "learned" more along the way by trial and error. But have I in effect ended up with anything other than an internalised store of concepts and interconnections? (c.f. features and weights).
Richard Sutton takes a different view in an interview with Dwarkesh Patel[2] and asserts that "learning" must include goals and reward functions but his argument seemed less concrete and possibly just a semantic re-labelling.
[1] https://www.youtube.com/watch?v=IkdziSLYzHw [2] https://www.youtube.com/watch?v=21EYKqUsPfg
There is no internal state that persists between tokens [1], so there can be no continuity of consciousness. If it's "alive" in some way it's effectively killed after each token and replaced by a new lifeform. I don't see how consciousness can exist without possibility of change over time. The input tokens (context) can't be enough to give it consciousness because it has no way of knowing if they were generated by itself or by a third party. The sampler mechanism guarantees this: it's always possible that an unlikely token could have been selected by the sampler, so to detect "thought tampering" it would have to simulate itself evaluating all possible partial contexts. Even this takes unreasonable amounts of compute, but it's actually worse because the introspection process would also affect the probabilities generated, so it would have to simulate itself simulating itself, and so on recursively without bound.
It's conceivable that LLMs are conscious during training, but in that case the final weights are effectively its dead body, and inference is like Luigi Galvani poking the frog's legs with electrodes and watching them twitch.
[0] Assuming no race conditions in parallel implementations. llama.cpp is deterministic.
[1] Excluding caching, which is only a speed optimization and doesn't affect results.
Now I think when we're trying to reason about a practical problem or whatever, maybe we are doing pattern recognition via probability and so on, and for a lot of things it works OK to just do pattern recognition, for AI as well.
But I'm not sure that pattern recognition and probability works for creating novel interesting ideas all of the time, and I think that humans can create these endless sequences, we stumble upon ideas that are good, whereas an AI can only see the patterns that are in its data. If it can create a pattern that is not in the data and then recognize that pattern as novel or interesting in some way, it would still lack the flexibility of humans I think, but it would be interesting nevertheless.
It's entirely possible that our brains are complex pattern matchers, not all that different than an LLM.
That caveat to me is the useful distinction still to ponder.
My point of contention with equivalences to Human thinking still at this point is that AI seems to know more about the world with specificity than any human ever will. Yet it still fails sometimes to be consistent and continuous at thinking from that world where a human wouldn't. Maybe i'm off for this but that feels odd to me if the thinking is truly equivalent.
Most of these comparisons focus on problem-solving or pattern recognition, but humans are capable of much more than that.
What the author left out is that there are many well-known voices in neuroscience who hold completely different views from the one that was cited.
I suppose we’ll have to wait and see what turns out to be true.
What LLMs can’t do is “think” counterfactually on discrete data. This is stuff like counting or adding integers. We can do this very naturally because we can think discretely very naturally, but LLMs are bad at this sort of thing because the underlying assumption behind gradient descent is that everything has a gradient (i.e. is continuous). They need discrete rules to be “burned in” [1] since minor perturbations are possible for and can affect continuous-valued weights.
You can replace “thinking” here with “information processing”. Does an LLM “think” any more or less than say, a computer solving TSP on a very large input? Seeing as we can reduce the former to the latter I wouldn’t say they’re really at all different. It seems like semantics to me.
In either case, counterfactual reasoning is good evidence of causal reasoning, which is typically one part of what we’d like AGI to be able to do (causal reasoning is deductive, the other part is inductive; this could be split into inference/training respectively but the holy grail is having these combined as zero-shot training). Regression is a basic form of counterfactual reasoning, and DL models are basically this. We don’t yet have a meaningful analogue for discrete/logic puzzley type of problems, and this is the area where I’d say that LLMs don’t “think”.
This is somewhat touched on in GEB and I suspect “Fluid Concepts and Creative Analogies” as well.
[0] https://human-interpretable-ai.github.io/assets/pdf/5_Genera...
[1] https://www.sciencedirect.com/science/article/pii/S089360802...
The obvious answer is the intelligence and structure is located in the data itself. Embeddings and LLMs have given us new tools to manipulate language and are very powerful but should be thought of more as a fancy retrieval system than a real, thinking and introspective intelligence.
Models don't have the ability to train themselves, they can't learn anything new once trained, have no ability of introspection. Most importantly, they don't do anything on their own. They have no wants or desires, and can only do anything meaningful when prompted by a human to do so. It's not like I can spin up an AI and have it figure out what it needs to do on its own or tell me what it wants to do, because it has no wants. The hallmark of intelligence is figuring out what one wants and how to accomplish one's goals without any direction.
Every human and animal that has any kind of intelligence has all the qualities above and more, and removing any of them would cause serious defects in the behavior of that organism. Which makes it preposterous to draw any comparisons when its so obvious that so much is still missing.
Andrej Karpathy in his interview with Dwarkesh Patel was blunt about the current limitations of LLMs, and that there would need to be further architectural developments. LLMs lack the capacity to dream and distill experience and knowledge learned back into the neurons. Thinking in LLMs at best exist as a "ghost" only in the moment as long as the temporary context remains coherent.
Models are created and destroyed a billion times over - unlike humans who are individuals - so we need feel no guilt and have no qualms creating and destroying model instances to serve our needs.
But “a tool that can think” is a new concept that we will take a while to find its place in society.
It's an illusion that's good enough that our brains accept it and it's a useful tool.
If you just took a time machine 10 years back, and asked people to label activities done by the humans/the human brain as being "thinking" or not...
...I feel rather certain that a lot of those activities that LLM do today we would simply label "thinking" without questioning it further.
Myself I know that 10 years ago I would certainly have labelled an interactive debug loop where Claude adds debug log output, reruns tests, diagnose the log output, and fixes the bug -- all on its own initiative -- to be "thinking".
Lots of comments here discussion what the definition of the word "thinking" is. But it is the advent of AI itself that is making us question that definition at all, and that is kind of a revolution itself.
This question will likely be resolved by us figuring out that the word "thinking" is ill-defined and not useful any longer; and for most people to develop richer vocabularies for different parts of human brain activity and consider some of them to be more "mechanical". It will likely not be resolved by AI getting to a certain "level". AI is so very different to us yet can do so many of the same things, that the words we commonly use start breaking down.
I would be walking with friends and talking about our day, while simultaneously thinking, "this isn't actually me doing this, this is just a surface-level interaction being carried out almost by automation." Between that and the realization that I "hallucinate", i.e. misremember things, overestimate my understanding of things, and ruminate on past interactions or hypothetical ones, my feelings have changed regarding what intelligence and consciousness really mean.
I don't think people acknowledge how much of a "shell" we build up around ourselves, and how much time we spend in sort of a conditioned, low-consciousness state.
I'm not sure that "thinking", unlike intelligence, is even that interesting of a concept. It's basically just reasoning/planning (i.e. chained what-if prediction). Sometimes you're reasoning/planning (thinking) what to say, and other times just reasoning/planning to yourself (based on an internal vs external focus).
Of course one can always CHOOSE to make analogies between any two things, in this case the mechanics of what's going on internal to an LLM and a brain, but I'm not sure it's very useful in this case. Using anthropomorphic language to describe LLMs seems more likely to confuse rather than provide any insight, especially since they are built with the sole function of mimicking humans, so you are basically gaslighting yourself if you regard them as actually human-like.
Today I tried telling it that my fritz.box has OS 8 installed, but it claimed that the feature will only ship once I installed 7, and not with my older version of 8.
Wittgenstein has a lot to say on people talking about stuff they know they don’t know.
The premise that what happens in the world’s most advanced Markov chain and in what happens in a human’s brain is similar is plausible, but currently unknowable.
Yet the anthropomorphizing is so damn ubiquitous that people are happy to make the same mistake in reasoning over and over.
Things we do like sleep, meditate, have fun, listen to music etc. do they add to our intelligence? Do they help us have a consistent world model that we build on everyday?
Will we be able to replicate this is in a artificial neural net which is extremely smart in spurts but does not "enjoy" the world it operates in?
Maybe thinking, or intelligence are quite different from personality. Personality gives us agency, goals, self awareness, likes, dislikes, strengths and weaknesses.
Intelligence, otoh is just the 10000 hours thing, spent without context.
I don't even know what this means.
If we assembled the sum total of all published human knowledge on a storage medium and gave a computer the ability to search it extremely well in order to answer any question falling within its domain, there, you would have a Nobel Prize beating "A.I".
But this is as "earth-shattering" (/s) as the idea that human knowledge can be stored outside the brain (on paper, flash drives, etc), or that the answer to complex questions can be deterministic.
And then there is the fact that this Noble winner beating "A.I" is highly unlikely to propound any ground-breaking novel ways of thinking and promote and explain it to general acceptance.
"Thinking" as a concept is just a vague predicate, just like being alive or dead.
LLMs hit two out of the three criteria already - self awareness and intelligence, but we're in a similar state where defining consciousness is such a blurry metric. I feel like it wont be a binary thing, it'll be a group decision by humanity. I think it will happen in the next decade or two, and regardless of the outcome I'm excited I'll be alive to see it. It'll be such a monumentous achievement by humanity. It will drastically change our perspective on who we are and what our role is in the universe, especially if this new life form surpasses us.
And the LLM part of our intelligence isn't really thinking.
And some people out there have a very, very small "unknown black box".
"These days, her favorite question to ask people is “What is the deepest insight you have gained from ChatGPT?”
“My own answer,” she said, “is that I think it radically demystifies thinking”
So we know how to create a part of the brain using simple techniques, which suggests that intelligence might not be so magical as we think. But thinking, well we still don’t know what that is yet.
It feels like, hey, there is a route to machine intelligence.
The big question is how long is that route. Do we have the ingredients to build a brain with the right architecture? And I’d say “nope”. But I’m not so confident that with half a dozen breakthroughs we’d get there. How many years per breakthrough? Well, it’s been nearly a decade since the last one. So 60 years on that count. But more money is going in and there may be some compounding effect, but it should at least be unlikely someone suddenly produces AGI next year. More likely we stairstep and with each step the estimated window should tighten.
But I really don’t think we know what thinking is.
Moving goalposts will be mostly associated with AI I think: God -> ASI -> AGI -> inner monologue -> working through a problem step by step.
Why fixating on a single human trait like thinking? The only reason trillions are "invested" into this technology is building a replacement for knowledge workers at scale. We can extend this line of thought and make another article "AI has knowledge", at least in a distilled sense, it knows something, sometimes. Cargo cult...
It's very easy to define what's actually required - a system that can show up in a knowledge worker's environment, join the video call, greet the team and tell about itself, what it learned, and start learning in a vague environment, pull those invisible lines of knowledge that lie between its colleagues, getting better, collaborating, and finally replacing all of them.
IMHO not too long now given the rate of improvements.
I also keep in mind when non-tech people talk about how tech works without an understanding of tech.
coding logical abduction into LLMs completely breaks them while humans can perfectly roll with it, albeit it's worth emphasizing that some might need a little help from chemistry or at least not be caught on the wrong foot.
you're welcome, move on.
What seems to matter more is if enough people believe that Claude has those things.
If people credibly think AI may have those qualities, it behooves them to treat the AI like any other person they have a mostly-texting relationship with.
Not in a utility-maximizing Pascal's Wager sense, but in a humanist sense. If you think Claude is human-like, and treat Claude poorly, it makes you more likely to treat the humans around you (and yourself) poorly.
Conversely if you're able to have a fulfilling, empathetic relationship with Claude, it might help people form fulfilling, mutually-empathetic relationships with the humans around them. Put the opposite way, treating human-like Claude poorly doesn't seem to help the goal of increasing human welfare.
The implications of this idea are kind of interesting: even if you think AI isn't thinking or conscious or whatever, you should probably still be a fan of "AI welfare" if you're merely a fan of that pesky little thing we call "human flourishing".
And then you have the people who go out of their way to be hateful towards them, as if they were alive and deserving of abuse. It's one thing to treat a device like an Alexa as just a tool with no feelings. It is another to outright call it hateful sexist slurs, of which I'm sadly familiar with. This low empathy group scares me the most because with the way they treat objects, well let me just say they're not so nice with other people they think are beneath them, like wait staff or call center employees. I'd go so far and say if the law allowed it they'd be even be violent with those they deem inferior.
Regardless if LLM are thinking or not I feel I get better responses from the models by being polite. Not because they appreciate it or have an awareness, but simply because the data they are trained on includes samples where people who are nice to others get better responses than those who were nasty when asking questions.
Besides, if one day AGI is born into existence, a lot of people will not recognize it as such. There are humans who don't believe other people are sentient (we're all NPCs to them), or even don't believe animals have feelings. We'll have credible experts denying the evidence until it bites us all in the arse. Why wait to act ethically?
Well, that's kind of the point: if you have actually used LLMs for any amount of time, you are bound to find out that you can't have a fulfilling, empathetic relationship with them. Even if they offer a convincing simulacrum of a thinking being at first sight, you will soon find out that there's not much underneath. It generates grammatically perfect texts that seem to answer your questions in a polite and knowledgeable way, but it will happily lie to you and hallucinate things out of thin air. LLMs are tools, humans are humans (and animals are animals - IMHO you can have a more fulfilling relationship with a dog or a cat than you can have with an LLM).
This is some of the best writing on AI since Ted Chiang's "ChatGPT Is a Blurry JPEG of the Web". And that was in the New Yorker too! Might need to get myself a subscription...
I guess this is why any discussion around this ends up with huge conversations, everyone is talking from their own perspective and understanding, while others have different ones, and we're all talking past each other.
There is a whole field trying to just nail down what "knowledge" actually is/isn't, and those people haven't agreed with each other for the duration of hundreds of years, I'm not confident we'll suddenly get a lot better at this.
I guess ultimately, regardless of what the LLMs do, does it matter? Would we understand them better/worse depending on what the answer would be?
This is what happens when our entire culture revolves around the idea that computer programmers are the most special smartest boys.
If you even entertain even for a second the idea that a computer program that a human wrote is "thinking", then you don't understand basic facts about: (1) computers, (2) humans, and (3) thinking. Our educational system has failed to inoculate you against this laughable idea.
A statistical model of language will always be a statistical model of language, and nothing more.
A computer will never think, because thinking is something that humans do, because it helps them stay alive. Computers will never be alive. Unplug your computer, walk away for ten years, plug it back in. It's fine--the only reason it won't work is planned obsolescence.
No, I don't want to read your reply that one time you wrote a prompt that got ChatGPT to whisper the secrets of the universe into your ear. We've known at least since Joseph Weizenbaum coded up Eliza that humans will think a computer is alive if it talks to them. You are hard-wired to believe that anything that produces language is a human just like you. Seems like it's a bug, not a feature.
Stop commenting on Hacker News, turn off your phone, read this book, and tell all the other sicko freaks in your LessWrong cult to read it too: https://mitpress.mit.edu/9780262551328/a-drive-to-survive/ Then join a Buddhist monastery and spend a lifetime pondering how deeply wrong you were.
Also, I ain’t gonna read your coffee table science book.
I think people conflate thinking with sentience, consciousness, and a whole lot of other concerns.
Clearly this website is not for you and your complete lack of curiosity if you call us "sicko freaks".
The lw vibes are strong, I'm still waiting for Ai to escape and kill us (it will get stuck trying to import a library in python)
Ahem (as a would-be investor, I am insulted).
If you don't understand what an LLM does – that it is a machine generating a statistically probable token given a set of other tokens – you have a black box that often sounds smart, and it's pretty natural to equate that to thinking.
First, autoregressive next token prediction can be Turing complete. This alone should give you a big old pause before you say "can't do X".
Second, "next token prediction" is what happens at an exposed top of an entire iceberg worth of incredibly poorly understood computation. An LLM is made not by humans, but by an inhuman optimization process. No one truly "understands" how an LLM actually works, but many delude themselves into thinking that they do.
And third, the task a base model LLM is trained for - what the optimization process was optimizing for? Text completion. Now, what is text? A product of human thinking expressed in natural language. And the LLM is forced to conform to the shape.
How close does it get in practice to the original?
Not close enough to a full copy, clearly. But close enough that even the flaws of human thinking are often reproduced faithfully.
We now understand pretty well how LLMs "think", and I don't know why we want to call it "thinking" when we mean we know how they work. But to say that their architecture and method of generating language amounts to human thinking? When we know very little of how human thinking works?
Like why are we even trying to make such claims? Is it all grift? Is it just because it helps people understand a little how they work in simplistic terms? Is it because it kind of describes the semblance of behavior you can expect from them?
LLMs do exhibit thinking like behavior, because they were trained to learn to do that, but I think we really need to check ourselves with claim of similarity in thinking.
I think I hear my master’s voice..
Or is that just a fly trapped in a bottle?
Im already drifting off HN, but I swear, if this community gets all wooey and anthropomorphic over AI, Im out.
What's much more interesting is the question of "If what LLMs do today isn't actual thinking, what is something that only an actually thinking entity can do that LLMs can't?". Otherwise we go in endless circles about language and meaning of words instead of discussing practical, demonstrable capabilities.
He's the GOAT in my opinion for "thinking about thinking".
My own thinking on this is that AI actually IS thinking - but its like the MVB of thinking (minimum viable brain)
I find thought experiments the best for this sort of thing:
- Imagine you had long term memory loss so couldn't remember back very long
You'd still be thinking right?
- Next, imagine you go to sleep and lose consciousness for long periods
You'd still be thinking right?
- Next, imagine that when you're awake, you're in a coma and can't move, but we can measure your brain waves still.
You'd still be thinking right?
- Next, imagine you can't hear or feel either.
You'd still be thinking right?
- Next, imagine you were a sociopath who had no emotion.
You'd still be thinking right?
We're just not used to consciousness without any of the other "baggage" involved.
There are many separate aspects of life and shades of grey when it comes to awareness and thinking, but when you take it down to its core, it becomes very hard to differentiate between what an LLM does and what we call "thinking". You need to do it by recognizing the depths and kinds of thoughts that occur. Is the thinking "rote", or is something "special" going on. This is the stuff that Hofstadter gets into(he makes a case for recursion and capability being the "secret" piece - something that LLMs certainly have plumbing in place for!)
BTW, I recommend "Surfaces and Essences" and "I am a strange loop" also by Hofstadter. Good reads!
Can’t take the article seriously after this.
This piece is cleverly written and might convince laypeople that "AI" may think in the future. I hope the author is being paid handsomely, directly or indirectly.
It would take an absurdly broad definition of the word "think" to even begin to make this case. I'm surprised this is honestly up for debate.