Human-level artificial intelligence is close to finally being achieved, according to a lead researcher at Google’s DeepMind AI division.
Dr Nando de Freitas said “the game is over” in the decades-long quest to realise artificial general intelligence (AGI) after DeepMind unveiled an AI system capable of completing a wide range of complex tasks, from stacking blocks to writing poetry.
Described as a “generalist agent”, DeepMind’s new Gato AI needs to just be scaled up in order to create an AI capable of rivalling human intelligence, Dr de Freitas said.
Responding to an opinion piece written in The Next Web that claimed “humans will never achieve AGI”, DeepMind’s research director wrote that it was his opinion that such an outcome is an inevitability.
“It’s all about scale now! The Game is Over!” he wrote on Twitter.
“It’s all about making these models bigger, safer, compute efficient, faster at sampling, smarter memory, more modalities, innovative data, on/offline... Solving these challenges is what will deliver AGI.”
When asked by machine learning researcher Alex Dimikas how far he believed the Gato AI was from passing a real Turing test – a measure of computer intelligence that requires a human to be unable to distinguish a machine from another human – Dr de Freitas replied: “Far still.”
Leading AI researchers have warned that the advent of AGI could result in an existential catastrophe for humanity, with Oxford University Professor Nick Bostrom speculating that a “superintelligent” system that surpasses biological intelligence could see humans replaced as the dominant life form on Earth.
It’s funny to me that there are AI researchers out there who are still deeply convinced that AGI is not somewhat imminent, or even near term feasible, and they are even in the majority https://t.co/lmAQ6N6eny
— Joscha Bach (@Plinz) May 17, 2022
One of the main concerns with the arrival of an AGI system, capable of teaching itself and becoming exponentially smarter than humans, is that it would be impossible to switch off.
Fielding further questions from AI researchers on Twitter, Dr de Freitas said “safety is of paramount importance” when developing AGI.
“It’s probably the biggest challenge we face,” he wrote. “Everyone should be thinking about it. Lack of enough diversity also worries me a lot.”
Someone’s opinion article. My opinion: It’s all about scale now! The Game is Over! It’s about making these models bigger, safer, compute efficient, faster at sampling, smarter memory, more modalities, INNOVATIVE DATA, on/offline, … 1/N https://t.co/UJxSLZGc71
— Nando de Freitas 🏳️🌈 (@NandoDF) May 14, 2022
Google, which acquired London-based DeepMind in 2014, is already working on a “big red button” to mitigate against the risks associated with an intelligence explosion.
In a 2016 paper titled ‘Safely Interruptible Agents’, DeepMind researchers outlined a framework for preventing advanced artificial intelligence from ignoring shut-down commands.
Safe interruptibility can be useful to take control of a robot that is misbehaving and may lead to irreversible consequences,” the paper stated.
“If such an agent is operating in real-time under human supervision, now and then it may be necessary for a human operator to press the big red button to prevent the agent from continuing a harmful sequence of actions—harmful either for the agent or for the environment—and lead the agent into a safer situation.”