Elon Musk says robots will someday be too quick for the human eye in another dire warning about artificial intelligence
The tech billionaire singled out Boston Dynamic's back-flipping bot to make a harrowing point
Tech billionaire Elon Musk has again warned of the dangers of artificial intelligence.
This time the Space-X founder singled out Atlas, Boston Dynamic's back-flipping robot, which has been described as “nightmare-inducing”.
Replying to the video of Atlas on Twitter, Musk wrote: “This is nothing. In a few years, that bot will move so fast you'll need a strobe light to see it.”
When someone asked about the significance of the strobe light, he wrote: “Otherwise you”d see a blur.”
Musk, who once said we have a “five to ten per cent chance” of making AI safe, went on to say the rapidly advancing technology needs oversight.
Alongside a video of the rap classic “Regulate” by Warren G and Nate Dogg, he tweeted: “Got to regulate AI / robotics like we do food, drugs, aircraft and cars.
“Public risks require public oversight. Getting rid of the FAA [wouldn't] make flying safer. They're there for good reason.”
This is not the first time the Tesla product architect has warned of the dangers of artificial intelligence.
"AI is a rare case where I think we need to be proactive in regulation than be reactive," he told the National Governors Association earlier this year.
He went on to say: "I have exposure to the most cutting-edge AI, and I think people should be really concerned by it
"AI is a fundamental risk to the existence of human civilization in a way that car accidents, airplane crashes, faulty drugs or bad food were not — they were harmful to a set of individuals within society, of course, but they were not harmful to society as a whole."
Fellow entrepreneurs, including Facebook founder Mark Zuckerberg, appear to disagree.
He believes AI will improve lives in the future, telling CNBC: "I think you can build things and the world gets better. But with AI especially, I am really optimistic.
"And I think people who are naysayers and try to drum up these doomsday scenarios — I just, I don't understand it. It's really negative and in some ways I actually think it is pretty irresponsible."