I agree, but it's gonna be because some C-Suite dweeb fires all some nuclear engineer tasked with emergency response and replaces them with a madlibs generator, and then a crisis happens...
And from what I understand, all they "learn" to do is predict what letter goes next. There's still no cognitive process, no manipulation of symbols, no abstraction of concepts. Nothing that a mind does. It's still just a fancy weighted random number generator. The resulting strings of text have no semantic value and only resemble speech. We're interpreting them as having meaning because most people don't look under the hood. It's linguistic Pareidolia. Our highly tuned pattern recognition that we rely on to communicate using speech, body language, and so forth is failing in the face of an object that closely resembles speech.
this is a major plot point in (CW: body horror, violence, profound existential dread, mental illness) Blindsight. Spoilers for Blindsight:
spoiler
A linguist spends hours talking with an alien entity before concluding that it has no semantic understanding of it's speech. It built a model of human speech by intercepting radio communications and was using that model to "communicate", but it was just mimicking human behavior it had observed. it had no understanding or awareness of why humans were engaging in that behavior or what it did, it just knew that they did it and was mimicking their behavior to get closer to them, the way some predators mimic their prey
Looking at an app like chatGPT, its horribly innacurate, mistake-prone, full of logical errors.
I think we need to push back against claiming that it makes mistakes, logical failures, "hallucinations" or "lies". It can't do those things. It's a computer, and computers do exactly what they're told. The problem is that the user often doesn't understand exactly what they're telling the computer to do. People prompt ChatGPT to answer questions because they think it's smart, that it understands what they're typing, that it can think and consider and solve problems. But it doesn't do any of that. It just compares their prompt to it's data set and assembles a string of letters that resembles the data set. It's not making mistakes, it's doing exactly what it was designed to do; Take an input and produce an output based on the statistical weights of it's data set. We're regularly fooling ourself because the output is in letters, not numbers, and we're attributing meaning to those letters where none exists. If it gets the answer "right" it's pure luck; There just happened to be enough text strings in it's training set, and it weighted the values in it's set in the right way, for a string of output that happens to resemble a correct answer.
This is compounded enormously because, as I understand it, the bazingas who designed these things built a complete black box - They have no way of determining why the LLM generated the outputs that it did. Presumably, unless there is some really weird shit going on, those outputs are deterministic - Given the same inputs the machine should produce the same outputs.
And from what I understand, all they “learn” to do is predict what letter goes next. There’s still no cognitive process, no manipulation of symbols, no abstraction of concepts.
There's a lot of argument about this. I know some people who think it's manipulating concepts, it can abstract ideas, shit like that. But my hard counter is that the image generators can't draw hands. And the reason they can't draw hands is that they're incapable of abstraction. Despite sampling likely millions or hundreds of millions of images of hands the model has no awareness that all of those inputs are part of a class of objects we call "hands", and that most hands have similar attributes.
We can look at a person with extra fingers, a person with fewer or missing fingers, a monkey, a robot, a crab, a space alien, and a snow man and we'll understand that whatever is at the end of the upper limbs, to a certain degree of difference, is a hand and has the attributes of hand - It manipulates and grasps objects, etc.
If someone asks us how many fingers are at the end of a hand we know it's five, but we also know that James Doohan, despite having four fingers, still has a hand. "Hand" is an abstract object we can manipulate.
But the plagiarism machine can't do that. All it does is reproduce variations of it's data set with no semantic understanding of that data set. It can't draw hands because in it's data set there are countless variations of hands, hands in all shapes, hands in all positions, hands of varying colors. We could look at all of those hands and recognize them as hands, and if asked to draw a hand in teh style of X we'd still give it five fingers. If we had more or fewer fingers we'd be doing it on purpose, knowing that we're deviating from the "ideal" hand object we understand.
But the LLM can't abstract, it can't conceive of "hand". it just looks for statistical weights in it's data sets. Since hands are so variable the data set is a mess. There are trends in color, there are trends in lines that we would recognize as fingers. But the LLM just generates statistically likely color values. It doesn't know aht hands or fingers are, so it doesn't know that the human prompting it wants a hand with five fingers, etc. It just outputs a string of numbers that are statistically similer to it's training set.
Idk if I'm explaining this well, but to me that inability to draw hands, and it's not just hands, is a silver bullet to the idea that these things think or manipulate symbols. Because it's not just hands, it doesn't recognize anything. When you look at the details of the images, the little things like buttons, jewelry, complex gadgets, they're almost always blobs of noise in roughly the right shape. It has no awareness that it's being asked to draw an abstracted object from a set of objects. It's just reproducing weighted data. It can do faces because there are a vast, vast number of faces in it's data set, probably far more than most other objects, and faces are very consistent in their shape and layout. So the probability that whatever nonsense it generates will be interpreted by human observers as a face is pretty high. But when you ask it to do something that isn't as consistently shaped and as massively represented in the set as faces it chokes.
The tells I look for for plagiarism machine "art" are generally things like jewelry, buttons, anything that should be symmetrical. They're really bad at symmetry, presumably because they can't abstract and so aren't aware that the buttons on each side of a coat are the same object, or the same class of objects and should be similar in most respects. Jewelry too - It's so varied, and the machine isn't understanding that they're discrete objects made up of smaller objects, so it just outputs a blur that, if you actually look at it, isn't actually jewelry.
Like maybe I'm wrong, maybe there is some weird totally alien process in there, but whatever it's doing, it's not doing anything like what we do. (Unless I am totally, completely wrong and just don't know enough to know I'm wrong, which would be really annoying).
Like that old story about the monkeys if given enough time, and a typewriter with endless ribbon and paper (and bananas I guess) will randomly produce Shakespeare's works. Might take the monkey 10,000 years, but dangit they'll get it done. And of course, by then we'll have forgotten all context and imagine this could only have been done because they are actually Superior to us and we will begin worshiping them with bananas as the main form of adoration.... 🍌🍌🍌🍌🍌
And as near as I can determine all this system does is give the monkeys a banana when they hit a key that, based on an analysis of shakespeare, is statistically likely to be next. And eventually the monkeys are trained to assemble words in ways that resemble shaespeare, but they're still monkeys with no idea what they're doing.
I don't think it's possible. The monkeys aren't monkeys, it's a prediction engine that decides what the next token - be it a letter, word, number, whatever - there's never any point in that process where it's going to start having self reference. It's a dead end. They're trying to work backwards from the end point of 6.5 billion years of brutal selection to re-create a process they don't understand.
Yeah, I was reading a reply where some guy said he could be a turing machine if he had enough spare sheets of paper to work with and that's not how human working memory works. If we assume that a cow is a spherical object in a vacuum then sure, buddy, you can simlulate a turing machine. But in the real world your meatsack can only manage so much stuff in your head and eventually you'd reach a point where you would no longer be able to keep performing the tasks necessary to do your turning machine thing. that's one of the most important things computers have going - You can store shitloads of information in memory and hard storage without losing track of it
This is just the whole Chinese room argument, it confuses consciousness for intelligence. Like, you're completely correct, but the capabilities of these things scale with compute used during training, with no sign of diminishing returns any time soon.
It could understand Nothing and still outsmart you because it's good at predicting the next token that corresponds with behavior that would achieve the goals of the system. All without having any internal human-style conscious experience. In the short term this means that essentially every human being with an internet connection now suddenly has access to a genius level intelligence that never sleeps and does whatever it's told, which has both good and bad implications. In long term, they could (and likely will) become far more intelligent than humans with, which will make them increasingly difficult to control.
It doesn't matter if the monkey understands what it's doing if gets so good at "randomly" hitting the typewriter that businesses hire the monkey instead of you, and then as the monkey becomes better and better starts handing out instructions to produce chemical weapons and other bio warfare agents to randos on the street. We need to take this technology seriously if we're going to prevent Microsoft, OpenAI, Facebook, Google, etc. from accidentally Ending the World with it, or deliberately making the world Worse with it.
They're starting a dangerous arms race where they release increasingly dangerous and poorly tested AI into the public, while dramatically overselling their safety. Pointing out that this technology is dangerous is the exact opposite of what they want.
You're playing into their grift by acting like the entire idea of AI is some bullshit techbro hype cycle, which is exactly what microsoft, openai, Facebook, etc want. The more people pay attention and think "hey maybe we shouldn't be integrating enormous black box neural networks deep in all of our infrastructure and replacing key human workers with them", the more difficult it will be for them to continue doing this.
What talking points then? I seem to be misunderstanding your criticism (or it's meaninglessly vague, but I'm trying to be charitable). What specifically have I said that you take issue with?
It's not the chinese room problem, it's a practical limitation of the ChatGPT plagiarism machines. We're not talking about a thought experiment where the guy in the room has the vast, vast, vast amount of rules needed to respond to any arbitrary input in a way the chinese speaker will interpret as semantically meaningful output. We're talking about a machine that exists right now, that far from being trained on an ideal, complete model of chinese is trained on billions and billions of shitposts on the internet.
Maybe someone will make a machine like that in the future, but this ain't it. This is a machine that predicts letters, has no ability to manipulate symbols, no semantic understanding, and no way to asses the truth value of it's outputs. And for various reasons, including being trained on billions of internet shitposts, it's unlikely to ever develop these things.
I'm really not interested in speculation about future potential intelligent systems and AIs. it's boring, it's been done to death, there's nothing new to add. Right now I want to better understand what these things do so I can own my friends who think they're manipulating abstract symbols and understand the semantic value of those symbols.
Yeah, obviously. Current AI is shit. But it's a proof that deep learning scales well enough to perform (or at least somewhat consistently replicate, depending on your outlook) behavior that humans recognize as intelligent.
Three years ago these things could barely write coherent sentences, now they can replace a substantial number of human workers, three years from now? Who the fuck knows, emergent abilities are hard to predict in these models by definition, but new ones Keep Appearing when they train larger and larger ones in higher quality data. This means large scale social disruption at best and catastrophe (everything from AI enabled bioterrorism to AI propaganda-driven fascism) at worst.
LLMs definitely are not the Magic that a lot of idiot techbros think they are, but it's a mistake to underestimate the technology because it "only generates the next token". The human brain only generates the next set of neural activations given the previous set of neural activations, and look at how far our intelligence got us.
The capabilities of these things scale with compute used during training, and some of the largest companies on earth are currently in an arms race to throw more and more compute at them. This Will Probably Not End Well. We went from AI barely being able to form a coherent sentence to AI suddenly being a bioterrororism risk in like 2 years because a bunch of chemistry papers were in its training data and now it knows how to synthesize novel chemical warfare agents.
It doesn't matter whether or not the machine understands what it's doing when it's enabling the proliferation of WMDs, or going rogue to achieve some Incoherent goal it extrapolated from it's training, you're still Dead at the end.
The human brain only generates the next set of neural activations given the previous set of neural activations
:doubt:
As near as anyone can tell humans are not deterministic, there's a lot more to cognition than neuronal activity, and these "humans are analogous to computers" arguments are enormously reductive at best, but usually just completely wrong.
Building some novel plague is a possibility, and a good reason to burn all of these things and shoot the idiots who made them.
Yes, obviously it's an oversimplification, but fundamentally every computational system is either Turing complete or it isn't, that's the idea I was getting at. The human brain is not magic, and it's not doing anything that a sophisticated enough algorithm running on a computer couldn't do given sufficient memory and power.
Computational universality has nothing to do with digital computer flipping bits. It just means that any system which manipulates information (performs computation), and can do so at a certain level of complexity (there's lots of equivalent ways of formulating it but the simplest is that it can do integer arithmetic) are exactly equivalent, in that they can all do the same set of computations.
It's pretty obvious that the human brain is at least Turing complete, since we can do integer arithmetic. It's also impossible for any computational system to be "more" than Turing complete (whatever that would even mean) since every single algorithm that can be computed in finite time can be expressed in terms of integer arithmetic, which means that a Turing machine could perform it.
Obviously the human brain is many, many, many layers of abstraction and us FAR more complicated than modern computers. Plus neurons aren't literally performing a bunch of addition and subtraction operations on data, the point is that whatever they are doing logically must be equivalent to some incomprehensibly vast set of simple arithmetic operations that could be performed by a Turing machine, because if the human brain can do a single thing that a general Turing machine can't, then it would either take infinite time or require infinite resources to do so.
This is why I fucking hate singularity cultist techbros. They convince the entire rest of society that AI is fake or that true AI is impossible or whatever by basically starting a religious cult around it.
This is harmful because AI is Incredibly dangerous and we need people to acknowledge that to start taking action to ensure that it's developed safely and don't suddenly have capabilities spike by 300% one month and now suddenly we have 30% unemployment, or a super-plague gets released because chatGPT 5 in 2026 told some idiot how to make flu viruses 10x more transmissible and 10x as deadly or whatever.
My worry isn't sapient AI, I genuinely do not care whether it's sapient, my worry is that in the short term it will enable people to commit bioterrorism and mass produce high quality propaganda, and in the longer term that it's capabilities might increase to the point of being difficult to control.
This is exactly the shit I'm talking about, you seem to dismiss the entire Idea that AI might outstrip human intelligence (and that this would likely be very bad) out of hand. I think this is a mistake born from not being familiar enough with the field
Ohhhhh, this is why your comment was so rude lmfao. Honestly fair. Sam Harris is a fucking idiot and I'm not a determinist, since quantum events are probably just random (though who the fuck knows tbh.)
I am a strict materialist, but in more of a "everything can be explained by natural forces and interactions, by definition, because we are made of matter and something that wasn't composed of natural forces and interactions would be completely unobservable and therefore irrelevant" sort of way.
And from what I understand, all they “learn” to do is predict what letter goes next.
https://thegradient.pub/othello/
It can be difficult to tease out exactly how a neural network is modeling its training data, but claiming that it's solely predicting the next letter is reductive to the point of being wrong.
That aside, I also just think people are being silly. If an AI can write working code (or beat chess grand masters every time), then obviously something interesting is going on, and protestations that it's not really thinking and reasoning for realz in a real way, are just kinda obnoxious.
I agree, but it's gonna be because some C-Suite dweeb fires all some nuclear engineer tasked with emergency response and replaces them with a madlibs generator, and then a crisis happens...
And from what I understand, all they "learn" to do is predict what letter goes next. There's still no cognitive process, no manipulation of symbols, no abstraction of concepts. Nothing that a mind does. It's still just a fancy weighted random number generator. The resulting strings of text have no semantic value and only resemble speech. We're interpreting them as having meaning because most people don't look under the hood. It's linguistic Pareidolia. Our highly tuned pattern recognition that we rely on to communicate using speech, body language, and so forth is failing in the face of an object that closely resembles speech.
this is a major plot point in (CW: body horror, violence, profound existential dread, mental illness) Blindsight. Spoilers for Blindsight:
spoiler
A linguist spends hours talking with an alien entity before concluding that it has no semantic understanding of it's speech. It built a model of human speech by intercepting radio communications and was using that model to "communicate", but it was just mimicking human behavior it had observed. it had no understanding or awareness of why humans were engaging in that behavior or what it did, it just knew that they did it and was mimicking their behavior to get closer to them, the way some predators mimic their prey
I think we need to push back against claiming that it makes mistakes, logical failures, "hallucinations" or "lies". It can't do those things. It's a computer, and computers do exactly what they're told. The problem is that the user often doesn't understand exactly what they're telling the computer to do. People prompt ChatGPT to answer questions because they think it's smart, that it understands what they're typing, that it can think and consider and solve problems. But it doesn't do any of that. It just compares their prompt to it's data set and assembles a string of letters that resembles the data set. It's not making mistakes, it's doing exactly what it was designed to do; Take an input and produce an output based on the statistical weights of it's data set. We're regularly fooling ourself because the output is in letters, not numbers, and we're attributing meaning to those letters where none exists. If it gets the answer "right" it's pure luck; There just happened to be enough text strings in it's training set, and it weighted the values in it's set in the right way, for a string of output that happens to resemble a correct answer.
This is compounded enormously because, as I understand it, the bazingas who designed these things built a complete black box - They have no way of determining why the LLM generated the outputs that it did. Presumably, unless there is some really weird shit going on, those outputs are deterministic - Given the same inputs the machine should produce the same outputs.
Fascinating but not surprising I guess
There's a lot of argument about this. I know some people who think it's manipulating concepts, it can abstract ideas, shit like that. But my hard counter is that the image generators can't draw hands. And the reason they can't draw hands is that they're incapable of abstraction. Despite sampling likely millions or hundreds of millions of images of hands the model has no awareness that all of those inputs are part of a class of objects we call "hands", and that most hands have similar attributes.
We can look at a person with extra fingers, a person with fewer or missing fingers, a monkey, a robot, a crab, a space alien, and a snow man and we'll understand that whatever is at the end of the upper limbs, to a certain degree of difference, is a hand and has the attributes of hand - It manipulates and grasps objects, etc.
If someone asks us how many fingers are at the end of a hand we know it's five, but we also know that James Doohan, despite having four fingers, still has a hand. "Hand" is an abstract object we can manipulate.
But the plagiarism machine can't do that. All it does is reproduce variations of it's data set with no semantic understanding of that data set. It can't draw hands because in it's data set there are countless variations of hands, hands in all shapes, hands in all positions, hands of varying colors. We could look at all of those hands and recognize them as hands, and if asked to draw a hand in teh style of X we'd still give it five fingers. If we had more or fewer fingers we'd be doing it on purpose, knowing that we're deviating from the "ideal" hand object we understand.
But the LLM can't abstract, it can't conceive of "hand". it just looks for statistical weights in it's data sets. Since hands are so variable the data set is a mess. There are trends in color, there are trends in lines that we would recognize as fingers. But the LLM just generates statistically likely color values. It doesn't know aht hands or fingers are, so it doesn't know that the human prompting it wants a hand with five fingers, etc. It just outputs a string of numbers that are statistically similer to it's training set.
Idk if I'm explaining this well, but to me that inability to draw hands, and it's not just hands, is a silver bullet to the idea that these things think or manipulate symbols. Because it's not just hands, it doesn't recognize anything. When you look at the details of the images, the little things like buttons, jewelry, complex gadgets, they're almost always blobs of noise in roughly the right shape. It has no awareness that it's being asked to draw an abstracted object from a set of objects. It's just reproducing weighted data. It can do faces because there are a vast, vast number of faces in it's data set, probably far more than most other objects, and faces are very consistent in their shape and layout. So the probability that whatever nonsense it generates will be interpreted by human observers as a face is pretty high. But when you ask it to do something that isn't as consistently shaped and as massively represented in the set as faces it chokes.
The tells I look for for plagiarism machine "art" are generally things like jewelry, buttons, anything that should be symmetrical. They're really bad at symmetry, presumably because they can't abstract and so aren't aware that the buttons on each side of a coat are the same object, or the same class of objects and should be similar in most respects. Jewelry too - It's so varied, and the machine isn't understanding that they're discrete objects made up of smaller objects, so it just outputs a blur that, if you actually look at it, isn't actually jewelry.
Like maybe I'm wrong, maybe there is some weird totally alien process in there, but whatever it's doing, it's not doing anything like what we do. (Unless I am totally, completely wrong and just don't know enough to know I'm wrong, which would be really annoying).
Like that old story about the monkeys if given enough time, and a typewriter with endless ribbon and paper (and bananas I guess) will randomly produce Shakespeare's works. Might take the monkey 10,000 years, but dangit they'll get it done. And of course, by then we'll have forgotten all context and imagine this could only have been done because they are actually Superior to us and we will begin worshiping them with bananas as the main form of adoration.... 🍌🍌🍌🍌🍌
And as near as I can determine all this system does is give the monkeys a banana when they hit a key that, based on an analysis of shakespeare, is statistically likely to be next. And eventually the monkeys are trained to assemble words in ways that resemble shaespeare, but they're still monkeys with no idea what they're doing.
We can perhaps hope that eventually they begin to learn from the experience. (in a strictly evolutionary way..) :P
I don't think it's possible. The monkeys aren't monkeys, it's a prediction engine that decides what the next token - be it a letter, word, number, whatever - there's never any point in that process where it's going to start having self reference. It's a dead end. They're trying to work backwards from the end point of 6.5 billion years of brutal selection to re-create a process they don't understand.
I was wondering. plus, the lifespan of a monkey and stuff.
Yeah, I was reading a reply where some guy said he could be a turing machine if he had enough spare sheets of paper to work with and that's not how human working memory works. If we assume that a cow is a spherical object in a vacuum then sure, buddy, you can simlulate a turing machine. But in the real world your meatsack can only manage so much stuff in your head and eventually you'd reach a point where you would no longer be able to keep performing the tasks necessary to do your turning machine thing. that's one of the most important things computers have going - You can store shitloads of information in memory and hard storage without losing track of it
This is just the whole Chinese room argument, it confuses consciousness for intelligence. Like, you're completely correct, but the capabilities of these things scale with compute used during training, with no sign of diminishing returns any time soon.
It could understand Nothing and still outsmart you because it's good at predicting the next token that corresponds with behavior that would achieve the goals of the system. All without having any internal human-style conscious experience. In the short term this means that essentially every human being with an internet connection now suddenly has access to a genius level intelligence that never sleeps and does whatever it's told, which has both good and bad implications. In long term, they could (and likely will) become far more intelligent than humans with, which will make them increasingly difficult to control.
It doesn't matter if the monkey understands what it's doing if gets so good at "randomly" hitting the typewriter that businesses hire the monkey instead of you, and then as the monkey becomes better and better starts handing out instructions to produce chemical weapons and other bio warfare agents to randos on the street. We need to take this technology seriously if we're going to prevent Microsoft, OpenAI, Facebook, Google, etc. from accidentally Ending the World with it, or deliberately making the world Worse with it.
deleted by creator
They're starting a dangerous arms race where they release increasingly dangerous and poorly tested AI into the public, while dramatically overselling their safety. Pointing out that this technology is dangerous is the exact opposite of what they want.
You're playing into their grift by acting like the entire idea of AI is some bullshit techbro hype cycle, which is exactly what microsoft, openai, Facebook, etc want. The more people pay attention and think "hey maybe we shouldn't be integrating enormous black box neural networks deep in all of our infrastructure and replacing key human workers with them", the more difficult it will be for them to continue doing this.
deleted by creator
What talking points then? I seem to be misunderstanding your criticism (or it's meaninglessly vague, but I'm trying to be charitable). What specifically have I said that you take issue with?
It's not the chinese room problem, it's a practical limitation of the ChatGPT plagiarism machines. We're not talking about a thought experiment where the guy in the room has the vast, vast, vast amount of rules needed to respond to any arbitrary input in a way the chinese speaker will interpret as semantically meaningful output. We're talking about a machine that exists right now, that far from being trained on an ideal, complete model of chinese is trained on billions and billions of shitposts on the internet.
Maybe someone will make a machine like that in the future, but this ain't it. This is a machine that predicts letters, has no ability to manipulate symbols, no semantic understanding, and no way to asses the truth value of it's outputs. And for various reasons, including being trained on billions of internet shitposts, it's unlikely to ever develop these things.
I'm really not interested in speculation about future potential intelligent systems and AIs. it's boring, it's been done to death, there's nothing new to add. Right now I want to better understand what these things do so I can own my friends who think they're manipulating abstract symbols and understand the semantic value of those symbols.
Yeah, obviously. Current AI is shit. But it's a proof that deep learning scales well enough to perform (or at least somewhat consistently replicate, depending on your outlook) behavior that humans recognize as intelligent.
Three years ago these things could barely write coherent sentences, now they can replace a substantial number of human workers, three years from now? Who the fuck knows, emergent abilities are hard to predict in these models by definition, but new ones Keep Appearing when they train larger and larger ones in higher quality data. This means large scale social disruption at best and catastrophe (everything from AI enabled bioterrorism to AI propaganda-driven fascism) at worst.
deleted by creator
deleted by creator
LLMs definitely are not the Magic that a lot of idiot techbros think they are, but it's a mistake to underestimate the technology because it "only generates the next token". The human brain only generates the next set of neural activations given the previous set of neural activations, and look at how far our intelligence got us.
The capabilities of these things scale with compute used during training, and some of the largest companies on earth are currently in an arms race to throw more and more compute at them. This Will Probably Not End Well. We went from AI barely being able to form a coherent sentence to AI suddenly being a bioterrororism risk in like 2 years because a bunch of chemistry papers were in its training data and now it knows how to synthesize novel chemical warfare agents.
It doesn't matter whether or not the machine understands what it's doing when it's enabling the proliferation of WMDs, or going rogue to achieve some Incoherent goal it extrapolated from it's training, you're still Dead at the end.
:doubt:
As near as anyone can tell humans are not deterministic, there's a lot more to cognition than neuronal activity, and these "humans are analogous to computers" arguments are enormously reductive at best, but usually just completely wrong.
Building some novel plague is a possibility, and a good reason to burn all of these things and shoot the idiots who made them.
More and more people are saying that the Chinese Room and Hard Problem of Consciousness are things we should care about!
Yes, obviously it's an oversimplification, but fundamentally every computational system is either Turing complete or it isn't, that's the idea I was getting at. The human brain is not magic, and it's not doing anything that a sophisticated enough algorithm running on a computer couldn't do given sufficient memory and power.
deleted by creator
Word. I don't see any bits getting flipped in the brain, this "brain as a computer" thing seems pretty sketchy.
Computational universality has nothing to do with digital computer flipping bits. It just means that any system which manipulates information (performs computation), and can do so at a certain level of complexity (there's lots of equivalent ways of formulating it but the simplest is that it can do integer arithmetic) are exactly equivalent, in that they can all do the same set of computations.
It's pretty obvious that the human brain is at least Turing complete, since we can do integer arithmetic. It's also impossible for any computational system to be "more" than Turing complete (whatever that would even mean) since every single algorithm that can be computed in finite time can be expressed in terms of integer arithmetic, which means that a Turing machine could perform it.
Obviously the human brain is many, many, many layers of abstraction and us FAR more complicated than modern computers. Plus neurons aren't literally performing a bunch of addition and subtraction operations on data, the point is that whatever they are doing logically must be equivalent to some incomprehensibly vast set of simple arithmetic operations that could be performed by a Turing machine, because if the human brain can do a single thing that a general Turing machine can't, then it would either take infinite time or require infinite resources to do so.
deleted by creator
This is why I fucking hate singularity cultist techbros. They convince the entire rest of society that AI is fake or that true AI is impossible or whatever by basically starting a religious cult around it.
This is harmful because AI is Incredibly dangerous and we need people to acknowledge that to start taking action to ensure that it's developed safely and don't suddenly have capabilities spike by 300% one month and now suddenly we have 30% unemployment, or a super-plague gets released because chatGPT 5 in 2026 told some idiot how to make flu viruses 10x more transmissible and 10x as deadly or whatever.
deleted by creator
My worry isn't sapient AI, I genuinely do not care whether it's sapient, my worry is that in the short term it will enable people to commit bioterrorism and mass produce high quality propaganda, and in the longer term that it's capabilities might increase to the point of being difficult to control.
This is exactly the shit I'm talking about, you seem to dismiss the entire Idea that AI might outstrip human intelligence (and that this would likely be very bad) out of hand. I think this is a mistake born from not being familiar enough with the field
Do you even know what the Church-Turing thesis is?
deleted by creator
Genuinely no
deleted by creator
I want to be a hard determinist but those quantum mechanics assholes say there are truly random events at the quantum level so *shrug*
deleted by creator
Quantumly indeterminate assholes!
Ohhhhh, this is why your comment was so rude lmfao. Honestly fair. Sam Harris is a fucking idiot and I'm not a determinist, since quantum events are probably just random (though who the fuck knows tbh.)
I am a strict materialist, but in more of a "everything can be explained by natural forces and interactions, by definition, because we are made of matter and something that wasn't composed of natural forces and interactions would be completely unobservable and therefore irrelevant" sort of way.
deleted by creator
https://thegradient.pub/othello/
It can be difficult to tease out exactly how a neural network is modeling its training data, but claiming that it's solely predicting the next letter is reductive to the point of being wrong.
That aside, I also just think people are being silly. If an AI can write working code (or beat chess grand masters every time), then obviously something interesting is going on, and protestations that it's not really thinking and reasoning for realz in a real way, are just kinda obnoxious.
deleted by creator