AI’s might destroy other AI’s but at least humanity is safe
The chief scientist leading Facebook’s research into Artificial Intelligence (AI) doesn’t believe the technology is a threat to human existence – unless we are very stupid. So there’s still a chance then…
“I don’t think at AI will become an existential threat to humanity,” said Yann LeCun, the top professor who is leading Facebook’s efforts in the space of AI and machine learning.
“I’m not saying that it’s impossible, but we would have to be very stupid to let that happen. Others have claimed that we would have to be very smart to prevent that from happening, but I don’t think that’s true,” he said.
Several top scientists and technologists such as Stephen Hawking and Elon Musk have warned humans face a risk from AI.
“If we are smart enough to build machines with super human intelligence, chances are we will not be stupid enough to give them infinite power to destroy humanity,” he continued.
“Also, there is a complete fallacy due to the fact that our only exposure to intelligence is through other humans. There are absolutely no reason that intelligent machines will even want to dominate the world and/or threaten humanity. The will to dominate is a very human one.”
LeCun, who leads Facebook’s AI Research (FAIR) added that behaviours specific to humans, such as violence or jealousy, would only be in intelligent machines if they are explicitly given to them. Even then, a second AI machine could be built with explicit instructions to destroy a dangerous AI.
“If both AIs have access to the same amount of computing resources, the second one will win, just like a tiger a shark or a virus can kill a human of superior intelligence.”
The professor at New York University also revealed his view on how rivals are doing with their AI research, saying Google’s DeepMind and Facebook AI projects are attracting talent from competitors.
“Microsoft is doing some good work, but they are losing quite a lot of people to Facebook and Google. They did some good work on deep learning in speech, and on handwriting recognition back in the early 2000s before the current craze. But their recent efforts seem to have been less ambitious than at FAIR or DeepMind,” he said, when asked who was leading the way in research among the big tech firms.
“A number of top researchers from Google Brain have left for DeepMind, OpenAI, or FAIR,” he added, while Apple’s secretive nature meant they were “not a player in the AI research circuit”.
“You simply cannot do leading-edge research in secret. If you can’t publish, it’s not research. At best, it’s technology development.”
Asked whether the technology is experiencing a bubble, he said, “There is a lot of hype around AI and deep learning at the moment,” and then he went on to say, “that said, deep learning produces real results and is at the root of a real industry that makes money today. The promises of it in the near future are very exciting – even without the hype – in areas like self-driving cars, medical imaging, personalised medicine, content filtering and so on.”