Comment by almosthere
5 days ago
Well, I think because we know how the code is written, in the sense that humans quite literally wrote the code for it - it's definitely not thinking, and it is literally doing what we asked, based on the data we gave it. It is specifically executing code we thought of. The output of course, we had no flying idea it would work this well.
But it is not sentient. It has no idea of a self or anything like that. If it makes people believe that it does, it is because we have written so much lore about it in the training data.
We do not write the code that makes it do what it does. We write the code that trains it to figure out how to do what it does. There's a big difference.
The code that builds the models and performance inference from it is code we have written. The data in the model is obviously the big trick. But what I'm saying is that if you run inference, that alone does not give it super-powers over your computer. You can write some agentic framework where it WOULD have power over your computer, but that's not what I'm referring to.
It's not a living thing inside the computer, it's just the inference building text token by token using probabilities based on the pre-computed model.
> It's not a living thing inside the computer, it's just the inference building text token by token using probabilities based on the pre-computed model.
Sure, and humans are just biochemical reactions moving muscles as their interface with the physical word.
I think the model of operation is not a good criticism, but please see my reply to the root comment in this thread where I detail my thoughts a bit.
You cannot say, 'we know it's not thinking because we wrote the code' when the inference 'code' we wrote amounts to, 'Hey, just do whatever you figured out during training okay'.
'Power over your computer', all that is orthogonal to the point. A human brain without a functioning body would still be thinking.
6 replies →
This is a bad take. We didn't write the model, we wrote an algorithm that searches the space of models that conform to some high level constraints as specified by the stacked transformer architecture. But stacked transformers are a very general computational paradigm. The training aspect converges the parameters to a specific model that well reproduces the training data. But the computational circuits the model picks out are discovered, not programmed. The emergent structures realize new computational dynamics that we are mostly blind to. We are not the programmers of these models, rather we are their incubators.
As far as sentience is concerned, we can't say they aren't sentient because we don't know the computational structures these models realize, nor do we know the computational structures required for sentience.
5 replies →
I think the discrepancy is this:
1. We trained it on a fraction of the world's information (e.g. text and media that is explicitly online)
2. It carries all of the biases us humans have and worse the biases that are present in the information we chose to explicitly share online (which may or may not be different to the experiences humans have in every day life)
> It carries all of the biases us humans have and worse the biases that are present in the information we chose to explicitly share online
This is going to be a huge problem. Most people assume computers are unbiased and rational, and increasing use of AI will lead to more and larger decisions being made by AI.
I see this a lot in what LLMs know and promote in terms of software architecture.
All seem biased to recent buzzwords and approaches. Discussions will include the same hand-waving of DDD, event-sourcing and hexagonal services, i.e. the current fashion. Nothing of worth apparently preceded them.
I fear that we are condemned to a future where there is no new novel progress, but just a regurgitation of those current fashion and biases.
and then the code to give it context. AFAIU, there is a lot of post training "setup" in the context and variables to get the trained model to "behave as we instruct it to"
Am I wrong about this?
Well, unless you believe in some spiritual, non-physical aspect of consciousness, we could probably agree that human intelligence is Turing-complete (with a slightly sloppy use of terms).
So any other Turing-complete model can emulate it, including a computer. We can even randomly generate Turing machines, as they are just data. Now imagine we are extremely lucky and happen to end up with a super-intelligent program which through the mediums it can communicate (it could be simply text-based but a 2D video with audio is no different for my perspective) can't be differentiated from a human being.
Would you consider it sentient?
Now replace the random generation with, say, a back propagation algorithm. If it's sufficiently large, don't you think it's indifferent from the former case - that is, novel qualities could emerge?
With that said, I don't think that current LLMs are anywhere close to this category, but I just don't think this your reasoning is sound.
> we could probably agree that human intelligence is Turing-complete (with a slightly sloppy use of terms). > So any other Turing-complete model can emulate it
You're going off the rails IMMEDIATELY in your logic.
Sure, one Turing-complete computer language can have its logic "emulated" by another, fine. But human intelligence is not a computer language -- you're mixing up the terms "Turing complete" and "Turing test".
It's like mixing up the terms "Strawberry jam" and "traffic jam" and then going on to talk about how cars taste on toast. It's nonsensical.
Game of life, PowerPoint, and a bunch of non-PL stuff are all Turing-complete. I don't mix up terms, I did use a slightly sloppy terminology but it is the correct concept - and my point is that we don't know of a computational model that can't be expressed by a Turing-machine, humans are a physical "machine", ergo we must also fall into that category.
Give my comment another read, but it was quite understandable from context. (Also, you may want to give a read to the Turing paper because being executable by a person as well was an important concept within)
3 replies →
We used to say "if you put a million monkeys on typewriters you would eventually get shakespear" and no one would ever say that anymore, because now we can literally write shakespear with an LLM.
And the monkey strategy has been 100% dismissed as shit..
We know how to deploy monkeys on typewriters, but we don't know what they'll type.
We know how to deploy transformers to train and inference a model, but we don't know what they'll type.
We DON'T know how a thinking human (or animal) brain works..
Do you see the difference.
The monkeys on typewriters saying is just a colorful way of saying that an infinite random sequence will contain all finite sequences somewhere within it. Which is true. But I don't see what infinite random sequences have to do with LLMs or human thinking.
> Do you see the difference
No? I'm not sure what you're getting at.
To be fair, we also trained the LLM on (among other things) shakespeare, and adjusted the weights so that generating shakespeare would be more likely after that training.
We don't claim a JPEG can paint great art, even though certain jpegs do.
2 replies →
I was going to use this analogy in the exact opposite way. We do have a very good understanding of how the human brain works. Saying we don't understand how the brain works is like saying we don't understand how the weather works.
If you put a million monkeys on typewriters you would eventually get shakespeare is exactly why LLM's will succeed and why humans have succeeded. If this weren't the case why didn't humans 30000 years ago create spacecraft if we were endowed with the same natural "gift".
2 replies →
> Would you consider it sentient?
Absolutely.
If you simulated a human brain by the atom, would you think the resulting construct would NOT be? What would be missing?
I think consciousness is simply an emergent property of our nervous system, but in order to express itself "language" is obviously needed and thus requires lots of complexity (more than what we typically see in animals or computer systems until recently).
> If you simulated a human brain by the atom,
That is what we don't know is possible. You don't even know what physics or particles are as yet undiscovered. And from what we even know currently, atoms are too coarse to form the basis of such "cloning"
And, my viewpoint is that, even if this were possible, just because you simulated a brain atom by atom, does not mean you have a consciousness. If it is the arrangement of matter that gives rise to consciousness, then would that new consciousness be the same person or not?
If you have a basis for answering that question, let's hear it.
22 replies →
There are many aspects to this that people like yourself miss, but I think we need satisfactory answers to them (or at least rigorous explorations of them) before we can make headway in these sorts of discussion.
Imagine we assume that A.I. could be conscious. What would be the identity/scope of that consciousness. To understand what I'm driving at, let's make an analogy to humans. Our consciousness is scoped to our bodies. We see through sense organ, and our brain, which process these signals, is located in a specific point in space. But we still do not know how consciousness arises in the brain and is bound to the body.
If you equate computation of sufficient complexity to consciousness, then the question arises: what exactly about computation would prodcuce consciousness? If we perform the same computation on a different substrate, would that then be the same consciousness, or a copy of the original? If it would not be the same consciousness, then just what give consciousness its identity?
I believe you would find it ridiculous to say that just because we are performing the computation on this chip, therefore the identity of the resulting consciousness is scoped to this chip.
> Imagine we assume that A.I. could be conscious. What would be the identity/scope of that consciousness
Well, first I would ask whether this question makes sense in the first place. Does consciousness have a scope? Does consciousness even exist? Or is that more of a name attributed to some pattern we recognize in our own way of thinking (but may not be universal)?
Also, would a person missing an arm, but having a robot arm they can control have their consciousness' "scope" extended to it? Given that people have phantom pains, does a physical body even needed to consider it your part?
This all sounds very irrelevant. Consciousness is clearly tied to specific parts of a substrate. My consciousness doesn't change when a hair falls off my head, nor when I cut my fingernails. But it does change in some way if you were to cut the tip of my finger, or if I take a hormone pill.
Similarly, if we can compute consciousness on a chip, then the chip obviously contains that consciousness. You can experimentally determine to what extent this is true: for example, you can experimentally check if increasing the clock frequency of said chip alters the consciousness that it is computing. Or if changing the thermal paste that attaches it to its cooler does so. I don't know what the results of these experiments would be, but they would be quite clearly determined.
Of course, there would certainly be some scale, and at some point it becomes semantics. The same is true with human consciousness: some aspects of the body are more tightly coupled to consciousness than others; if you cut my hand, my consciousness will change more than if you cut a small piece of my bowel, but less than if you cut out a large piece of my brain. At what point do you draw the line and say "consciousness exists in the brain but not the hands"? It's all arbitrary to some extent. Even worse, say I use a journal where I write down some of my most cherished thoughts, and say that I am quite forgetful and I often go through this journal to remind myself of various thoughts before taking a decision. Would it not then be fair to say that the journal itself contains a part of my consciousness? After all, if someone were to tamper with it in subtle enough ways, they would certainly be able to influence my thought process, more so than even cutting off one of my hands, wouldn't they?
3 replies →
Now convince us that you’re sentient and not just regurgitating what you’ve heard and seen in your life.
By what definition of "sentience"? Wikipedia claims "Sentience is the ability to experience feelings and sensations" as an opening statement, which I think would be trivial depending again on your definition of "experience" and "sensations". Can a LLM hooked up to sensor events be considered to "experience sensations"? I could see arguments both ways for that.
I have no way of measuring whether or not you experience feelings and sensations, or are just regurgitating statements to convince me of that.
The only basis I have for assuming you are sentient according to that definition is trust in your self-reports.
7 replies →
It's not accurate to say we "wrote the code for it". AI isn't built like normal software. Nowhere inside an AI will you find lines of code that say If X Then Y, and so on.
Rather, these models are literally grown during the training phase. And all the intelligence emerges from that growth. That's what makes them a black box and extremely difficult to penetrate. No one can say exactly how they work inside for a given problem.
This is probably true. But the truth is we have absolutely no idea what sentience is and what gives rise to it. We cannot identify why humans have it rather than just being complex biological machines, or whether and why other animals do. We have no idea what the rules or, nevermind how and why they would or wouldn't apply to AI.
What’s crazy to me is the mechanism of pleasure or pain. I can understand that with enough complexity we can give rise to sentience but what does it take to achieve sensation?
This is the "hard problem of consciousness". It's more important than ever as machines begin to act more like humans, but my takeaway is we have no idea. https://en.wikipedia.org/wiki/Hard_problem_of_consciousness
Input is input. There's no reason why we should assume that a data source from embodiment is any different to any other data source.
A body
I’d say it’s possible to experience mental anguish/worry without the body participating. Solely a cognitive pain from consternation.
2 replies →
How does a body know what's going on? Would you say it has any input devices?
Can you tell me how you understand that?
Because I sincerely do not. I have frankly no idea how sentience arises from non sentience. But it's a topic that really interests me.
We have examples of non sentience everywhere already with animals. And then an example of sentience with humans. So if you diff our brains the difference lies within a module in our prefrontal cortex. It’s a black box of logic but I can ‘understand’ or be willing to accept that it’s owed to ‘just’ more grey matter adding the self awareness to the rest of the system.
But to me the big mystery is how animals have sensation at all to begin with. What gives rise to that is a greater mystery to me personally.
There are examples of people who have no ability to feel pain yet are still able to think. Now I wonder if they ever experience mental anguish.
3 replies →
> But it is not sentient. It has no idea of a self or anything like that.
Who stated that sentience or sense of self is a part of thinking?
Unless the idea of us having a thinking self is just something that comes out of our mouth, an artifact of language. In which case we are not that different - in the end we all came from mere atoms, after all!
Your brain is just following the laws of chemistry. So where is your thinking found in a bunch of chemical reactions?