Is it possible to build an artificial superintelligence without fully replicating the human brain?

Publié le 27 Juin 2014

Is it possible to build an artificial superintelligence without fully replicating the human brain?

Is it possible to build an artificial superintelligence without fully replicating the human brain?

The technological singularity requires the creation of an artificial superintelligence (ASI). But does that ASI need to be modelled on the human brain, or is it even necessary to be able to fully replicate the human brain and consciousness digitally in order to design an ASI?

Animal brains and computers don't work the same way. Brains are massively parallel three-dimensional networks, while computers still process information in a very linear fashion, although millions of times faster than brains. Microprocessors can perform amazing calculations, far exceeding the speed and efficiency of the human brain using completely different patterns to process information. The drawback is that traditional chips are not good at processing massively parallel data, solving complex problems, or recognizing patterns.

Newly developed neuromorphic chips are modelling the massively parallel way the brain processes information using, among others, neural networks.

Neuromorphic computers should ideally use optical technology, which can potentially process trillions of simultaneous calculations, making it possible to simulate a whole human brain.

The Blue Brain Project and the Human Brain Project, funded by the European Union, the Swiss government and IBM, are two such attempts to build a full computer model of a functioning human brain using a biologically realistic model of neurons. The Human Brain Project aims to achieve a functional simulation of the human brain for 2016.

Neuromorphic chips make it possible for computers to process sensory data, detect and predict patterns, and learn from experience. This is a huge advance in artificial intelligence, a step closer to creating an artificial general intelligence (AGI), i.e. an AI that could successfully perform any intellectual task that a human being can.

Think of an AGI inside a humanoid robot, a machine that looks and behave like us, but with customizable skills and that can perform practically any task better than a real human. These robots could be self-aware and/or sentient, depending on how we choose to build them. Manufacturing robots wouldn't need to be, but what about social robots living with us, taking care of the young, the sick or the elderly ? Surely it would be nicer they could converse with us as if they were conscious, sentient beings like us, a bit like the AI in Spike Jonze's 2013 movie Her.

In a not too distant future, perhaps less than two decades, such robots could replace humans for practically any job, creating a society of abundance where humans can spend their time however they like. In this model, highly capable robots would run the economy for us. Food, energy and most consumer products would be free or very cheap, and people would receive a fixed monthly allowance from the government.

This all sounds very nice. But what about an AI that would be greatly surpass the brightest human minds ? An artificial superintelligence (ASI), or strong AI (SAI), with the ability to learn and improve on itself, and potentially becoming millions or billions of times more intelligent and capable than humans ? The creation of such an entity would theoretically lead to the mythical technological singularity.

Futurist and inventor Ray Kurzweil believes that the singularity will happen some time around 2045. Among Kurzweil's critics is Microsoft cofounder Paul Allen, who believes that the singularity is still a long way off. Allen argues that for a real singularity-level computer intelligence to be built, the scientific understanding of how the human brain works will need to accelerate exponentially (like digital technologies), and that the process of original scientific discovery just doesn't behave that way. He calls this issue the complexity brake.

Without interfering in the argument between Paul Allen and Ray Kurzweil (who replied convincinglyhere), the question I want to discuss is whether it is absolutely necessary to fully understand and replicate the way the human brain works to create an ASI.

Great intelligence doesn't have to be modelled on the human brain

It is a natural for us to think that humans are the culmination of intelligence, simply because it is the case in the biological world on Earth. But that doesn't mean that our brain is perfect or that other forms of higher intelligence cannot exist if they aren't based on the same model.

If extraterrestrial beings with a greater intelligence than ours exist, it is virtually unthinkable that their brains be shaped and function like ours. The process of evolution is so random and complex that even if life were to be created again on a planet identical to Earth, it wouldn't unfold the same way as it did for us, and consequently the species wouldn't be the same. What if the Permian-Triassic extinction, or any other mass extinction event hadn't occured ? We wouldn't be there. But that doesn't mean that other intelligent animals wouldn't have evolved instead of us. Perhaps there would have been octopus-like creatures more intelligent than humans with a completely different brain structure.

It's pure human vanity and short-sightedness to think that everything good and intelligent has to be modelled on us. That is the kind of thinking that led to the development of religions with anthropomorphized gods. Humble or unpretentious religions like animism or Buddhism either have no human-like deity or no god at all. More arrogant or self-righteous religions, be them polytheistic or monotheistic, have typically imagined gods as superhumans. We don't want to make the same mistake with artificial superintelligence. Greater than human intelligence does not have to be an inflated version of human intelligence, nor should it be based on our biological brains.

The human brain is the fortuitious result of four billion years of evolution. Or rather, it is one tiny branch in the grand tree of evolution. Birds have much smaller brains than mammals and are generally considered stupid animals compared to most mammals. Yet, crows have reasoning skills that can exceed that of a preschooler. They display conscious, purposeful behaviour, a combined with a sense of initiative, elaborate problem solving abilities of their own, and can even use tools. All this with a brain the size of a fava-bean. A 2004 study from the departments of animal behavior and experimental psychology at the University of Cambridge claimed that crows were as clever as the great apes.

Clearly there is no need to replicate the intricacies of a human cortex to achieve consciousness and initiative. Intelligence does not depend only on brain size, the number of neurons, or cortex complexity, but also the brain-to-body mass ratio. That is why cattle, who have brains as big as chimpanzees, are stupider than ravens or mice.

But what about computers ? Computers are pure "brains".

They don't have bodies. And indeed as computers get faster and more efficient, their size tend to decrease, not increase. This is yet another example of why we shouldn't compare biological brains and computers.

As Ray Kurzweil explains in his reply to Paul Allen, learning about how the human brains works only serve to provide "biologically inspired methods that can accelerate work in AI, much of which has progressed without significant insight as to how the brain performs similar functions. [...] The way that these massively redundant structures in the brain differentiate is through learning and experience. The current state of the art in AI does, however, enable systems to also learn from their own experience." He then adds that IBM's Watson learned most of its knowledge by reading on its own.

In conclusion, there is no rational reason to believe that an artificial superintelligence couldn't come into being without being entirely modelled on the human brain, or any animal brain. A computer chip will never be the same as a biochemical neural network, and a machine will never feel emotions the same way as us (although they may feel emotions that are out of the range of human perception). But notwithstanding these differences, some computers can already acquire knowledge on their own, and will become increasingly good at it, even if they don't learn exactly the same way as humans. Once given the chance to improve on themselves, intelligent machines could set in motion a non-biological evolution leading to greater than human intelligence, and eventually to the singularity.

read more : http://ieet.org/index.php/IEET/more/hay20140625

Dare to better ? OK !

Biologically inspired methods that can accelerate work in AI, much of which has progressed without significant insight as to how the brain performs similar functions. [...] The way that these massively redundant structures in the brain differentiate is through learning and experience. The current state of the art in AI does, however, enable systems to also learn from their own experience

Ray Kurzweil

But what about computers ? Computers are pure "brains".  They don't have bodies. And indeed as computers get faster and more efficient, their size tend to decrease, not increase.

But what about computers ? Computers are pure "brains". They don't have bodies. And indeed as computers get faster and more efficient, their size tend to decrease, not increase.

Repost0
Commenter cet article