The year is 2030, and artificial intelligence has changed practically everything. Is it a change for the better or has AI threatened what it means to be human, to be productive and to exercise free will?
You’ve heard the dire predictions from some of the brightest minds about AI’s impact. Tesla and SpaceX chief Elon Musk worries that AI is far more dangerous than nuclear weapons. The late scientist Stephen Hawking warned AI could serve as the "worst event in the history of our civilization" unless humanity is prepared for its possible risks.
But many experts, even those mindful of such risks, have a more positive outlook, especially in health-care and possibly in education.
That’s one of the takeaways from a new AI study released Monday by the Pew Research Center and Elon University’s Imagining the Internet Center. Pew canvassed the opinions 979 experts over the summer, a group that included prominent technologists, developers, innovators and business and policy leaders.
Respondents, some of whom chose to remain anonymous, were asked to weigh in on a weighty question: “By 2030, do you think it is most likely that advancing AI and related technology systems will enhance human capacities and empower them?”
Nearly two-thirds predicted most of us will be mostly better off. But a third think otherwise, and a majority of the the experts expressed at least some concern over the long-term impact of AI on the “essential elements of being human.”
Among those concerns were data abuse, loss of jobs, loss of control as decision-making in digital systems is ceded to "black box" tools that take data in and spit answers out, an erosion in our ability to think for ourselves, and yes, the mayhem brought on by autonomous weapons, cybercrime, lies and propaganda.
“There’s a quite consistent message throughout answers…that some good things would emerge and there were some problems to worry about,” says Lee Rainie, director of internet and technology research at Pew Research Center. Janna Anderson, director of Elon University’s Imagining the Internet Center, added that some respondents thought we’d be OK through 2030, “but I’m not sure after that.”
A sampling of viewpoints
Andre McLaughlin at Yale, who had been deputy chief technology officer in the Obama Administration and a global public policy lead at Google, said that “My sense is that innovations like the internet and networked AI have massive short-term benefits, along with long-term negatives that can take decades to be recognizable. AI will drive a vast range of efficiency optimizations but also enable hidden discrimination and arbitrary penalization of individuals in areas like insurance, job seeking, and performance assessment.”
Technology blogger Wendy M. Grossman writes: “I believe human-machine AI collaboration will be successful in many areas, but that we will be seeing, like we are now over Facebook and other social media, serious questions about ownership and who benefits. It seems likely that the limits of what machines can do will be somewhat clearer than they are now, when we’re awash in hype. We will know by then, for example, how successful self-driving cars are going to be, and the problems inherent in handing off control from humans to machines in a variety of areas will also have become clearer.”
Leonard Kleinrock, Internet Hall of Fame member, replied, “As AI and machine learning improve, we will see highly customized interactions between humans and their health care needs. This mass customization will enable each human to have her medical history, DNA profile, drug allergies, genetic makeup, etc., always available to any caregiver/medical professional.”
Robert Epstein, senior research psychologist at the American Institute for Behavioral Research and Technology says that, “By 2030, it is likely that AIs will have achieved a type of sentience, even if it is not human-like. They will also be able to exercise varying degrees of control over most human communications, financial transactions, transportation systems, power grids, and weapons systems…and we will have no way of dislodging them. How they decide to deal with humanity — to help us, ignore us or destroy us — will be entirely up to them, and there is no way currently to predict which avenue they will choose. Because a few paranoid humans will almost certainly try to destroy the new sentient AIs, there is at least a reasonable possibility that they will swat us like the flies we are — the possibility that Stephen Hawking, Elon Musk and others have warned about.”
A social scientist who remained anonymous said, “My chief fear is face-recognition used for social control. Even Microsoft has begged for government regulation! Surveillance of all kinds is the future for AI. It is not benign if not controlled.”
Yet another anonymous respondent offered a different concern: “Knowing humanity, I assume particularly wealthy white males will be better off, while the rest of humanity will suffer from it.”
Ben Shneiderman, founder of the Human Computer Interaction Center at the University of Maryland, offers a very bullish take: “Automation is largely a positive force, which increases productivity, lowers costs and raises living standards. Automation expands the demand for services, thereby raising employment, which is what has happened at Amazon and FedEx. My position is contrary to those who believe that robots and artificial intelligence will lead to widespread unemployment.”
And Wendy Hall, a professor of computer science at the University of Southampton and executive director of the Web Science Institute said, “It is a leap of faith to think that by 2030 we will have learnt to build AI in a responsible way and we will have learnt how to regulate the AI and robotics industries in a way that is good for humanity. We may not have all the answers by 2030, but we need to be on the right track by then.”
Readers: What is your view of AI? Are you optimistic things will be better by 2030 or are you frightened by the possibilities?
Emai: firstname.lastname@example.org; Follow USA TODAY Personal Tech Columnist @edbaig on Twitter