Stephen Hawking’s fear of artificial intelligence has been well-documented over the last several years. While he had originally made some strong comments regarding the future of artificial intelligence and how it might hurt humanity in the long run years ago, the debate was only expanded when others from the tech and science world supported his viewpoint as well. Interestingly, this is something that has become a commonplace debate. While some think that computers or artificial intelligence is only as powerful as those who program it, others believe that there will come a time when computers eventually outsmart humans, outrank them, and eventually begin making decisions for themselves.
Now, that might seem a little far-fetched but the truth is that it likely isn’t as much like a plot line from a science-fiction movie as one might be led to believe. Right now, artificial intelligence, computers, and machines all play a pivotal role in making the world continue. Logically speaking, it definitely gives credence to the fact that as those computers and systems become more powerful, they could eventually cause problems for humans. However, many people believe that those ideas are foolish, or far ahead of time.
Hawking said though, in the next 100 years, artificial intelligence could become so powerful that it overtakes humans as the most dominant species on Earth. He said in part, “Computers will overtake humans with AI at some within the next 100 years. When that happens, we need to make sure the computers have goals aligned with ours.” While his sentiments might seem a bit catastrophic, they really are realistic in terms of what he is actually talking about.
However, that would require artificial intelligence actually getting to that point first. That alone is something that could be at least a hundred years away. Hawking continued by pointing out that, “Our future is a race between the growing power of technology and the wisdom with which we use it.” At the end of the day, this is ultimately what the race comes down to. It comes down to humans simply understanding what they’re working with and not creating outside of their own goal or mission. In many ways though, even with his expressed and legitimate concerns, many people will continue doubting what he is saying in terms of scope.
The scope of what would have to happen – in order to even damage the human race would need to be mind boggling. That’s something that hasn’t been spoken about yet, despite the fact that Elon Musk, and Stephen Hawking both signed an open letter discussing the threats of artificial intelligence. It’s about understanding the limits and the power of the computers we employ.