It’s tempting to think that a machine that could think would think like us. But a bit of reflection shows that’s not an inevitable conclusion.
To begin with, we’d better be clear about what we mean by “think.” A comparison with human thinking might be intuitive, but what about animal thinking? Does a chimpanzee think? Does a crow? Does an octopus?
There may even be alien intelligences that we might not even recognize as such because they are so radically different from us. Perhaps we could pass each other in close proximity, each unaware that the other existed, having no way to engage.
Certainly animals other than humans have cognitive abilities geared toward understanding tools and causal relationships, communication, and even to recognizing directed and purposeful thinking in others. We’d probably consider any or all of that thinking.
And let’s face it, if we built a machine that did all the above, we’d be patting ourselves on the back and saying “mission accomplished.” But could a machine go a step further and be like a human mind? What’s more, how would we know if it did?
Just because a computer acts like it has a mind, it doesn’t mean it must have one. It might be all show and no substance, an instance of a philosophical zombie.
It was this notion that motivated British codebreaker and mathematician Alan Turing to come up with his famous “Turing test,” in which a computer would interact with a human through a screen. If, more often than not, the human was unsure it was a computer, the computer would pass the test. For Turing, all that mattered was behavior, there was no computational “inner life” to be concerned about.
But this inner life matters to some of us. The philosopher Thomas Nagel said that there was “something that it is like” to have conscious experiences. There’s something that it is like to see the color red, or to go water skiing. We are more than just our brain states.
Could there ever be “something that it’s like” to be a thinking machine? In an imagined conversation with the first intelligent machine, a human might ask, “Are you conscious?” to which it might reply, “How would I know?”
Is Thinking Just Computation?
Under the hood of computer thinking, as we currently imagine it, is sheer computation. It’s about calculations per second and the number of potential computational pathways.
But we are not at all sure that thinking or consciousness is a function of computation, at least the way a binary computer does it. Could thinking be more than just computation? What else is needed? And if it is all about computation, why is the human brain so bad at it?
Most of us are flat out multiplying a couple of two digit numbers in our heads, let alone performing trillions of calculations a second. Or is there some deep processing of data that goes on below our awareness that ultimately results in our arithmetically impaired consciousness (the argument of so-called Strong AI)?
Generally speaking, what computers are good at, like raw data manipulation, humans are quite bad at; and what computers are bad at, such as language, poetry, voice recognition, interpreting complex behavior, and making holistic judgments, humans are quite good at.
If the analogy between human and computer “thinking” is so bad, why expect computers to eventually think like us? Or might computers of the future lose their characteristic arithmetical aptitude as the full weight of consciousness emerges?
Belief, Doubt, and Values
Then we have words like “belief” and “doubt” that are characteristic of human thinking. But what could it possibly mean for a computer to believe something, apart from the trivial meaning that it acted in ignorance of the possibility that it could be wrong? In other words, could a computer have genuine doubt, and then go ahead and act anyway?
When it comes to questions of value, questions about what we think is important in life and why, it’s interesting to consider two things.
The first is whether a thinking computer could be capable of attributing value to anything at all. The second is that if it could attribute value to anything, what would it choose? We’d want to be a bit careful here, it seems, even without getting into the possibility of mechanical free will.
It would be nice to program into computers a human-style value system. But, on the one hand, we aren’t quite sure what that is, or how that could be done, and, on the other hand, if computers started programming themselves they may decide otherwise.
While it’s great fun to think about all this, we should spend a bit of time trying to understand what we want thinking computers to be. And maybe a bit more time should be spent trying to understand ourselves before we branch out.
Peter Ellerton is a lecturer in critical thinking at The University of Queensland in Australia. This article was previously published on TheConversation.com