Published On: Wed, Jan 20th, 2016

New Technology Could Threaten Human Survival, Says Stephen Hawking

Physicist Stephen Hawking Lectures At UC Berkeley

The Cambridge professor Stephen Hawking has warned that most of the threats humanity faces come from progress made in science and technology. They include nuclear war, catastrophic global warming, and genetically engineered viruses. New technologies, he says, will likely bring about “new ways things can go wrong” for human survival.

Hawking made the comments while recording the BBC’s annual Reith Lectures on Jan. 7. His lecture, on the nature of black holes, was split into two parts and will be broadcast on radio on Jan. 26 and Feb. 2.

We are not going to stop making progress, or reverse it, so we have to recognize the dangers and control them.
— Stephen Hawking

The physicist said that a disaster on Earth—a “near certainty” in the next 1, 000 to 10, 000 years—will not spell the end of humanity because by that time humans are likely to have spread out into space and to other stars.

“However, we will not establish self-sustaining colonies in space for at least the next hundred years, so we have to be very careful in this period, ” he joked, provoking laughter from the audience.

“We are not going to stop making progress, or reverse it, so we have to recognize the dangers and control them. I’m an optimist, and I believe we can, ” he added.

In July, Hawking signed a letter along with Elon Musk and Steve Wozniak calling for a ban on autonomous weapons, and in October he conducted a Q&A on Reddit devoted solely to the question of artificial intelligence AI risk.

“We should shift the goal of AI from creating pure undirected artificial intelligence to creating beneficial intelligence, ” Hawking wrote. “It might take decades to figure out how to do this, so let’s start researching this today rather than the night before the first strong AI is switched on.”

The worry, Hawking said, wasn’t that an evil scientist would manufacture an immoral machine in a lab, but that an AI designed for normal purposes might malfunction and ending up harming humanity indirectly.

“The real risk with AI isn’t malice but competence. A superintelligent AI will be extremely good at accomplishing its goals, and if those goals aren’t aligned with ours, we’re in trouble, ” Hawking wrote. “You’re probably not an evil ant-hater who steps on ants out of malice, but if you’re in charge of a hydroelectric green energy project and there’s an anthill in the region to be flooded, too bad for the ants. Let’s not place humanity in the position of those ants.”

The Associated Press contributed to this report.




Read more about: , ,

Wordpress site Developed by Fixing WordPress Problems