Elon Musk donates $10mn to stop AI from turning against humans
The $10 million donation is being made to the Future of Life Institute (FLI), a volunteer-run non-profit research and outreach organization “working to mitigate existential risks facing humanity.” FLI’s first area of concern is the “potential risks from the development of human-level artificial intelligence.”
Funding research on artificial intelligence safety. It's all fun & games until someone loses an I http://t.co/t1aGnrTU21
— Elon Musk (@elonmusk) January 15, 2015
Beginning January 22, Musk’s donation will support an open grant competition for AI researchers and AI-related research in fields such as economics, law, ethics, and policy. Musk, the CEO of Tesla and SpaceX, has called AI potentially more dangerous than nukes.
“Here are all these leading AI researchers saying that AI safety is important,” said Musk in a released statement. “I agree with them, so I’m...committing $10 M to support research aimed at keeping AI beneficial for humanity.”
There are mounting concerns among technology and scientific leaders that too much emphasis and money goes towards research into “speech recognition, image classification, autonomous vehicles, machine translation, legged location and question-answering systems,” but little is spent on analyzing how these new advances could help society.
“While heavy industry and government investment has finally brought AI from niche academic research to early forms of potential world-transforming technology, to date relatively little funding has been available to help ensure that this change is actually a net positive one for humanity,” said Professor Anthony Aguirre, FLI co-founder.
Elon Musk puts money where his mouth is, donates $10 million toward "beneficial" AI research: http://t.co/mMEDDgniwK
— Eric C. Tatro (@transhumanistic) January 15, 2015
Musk joined Stephen Hawking and other technologists in an open letter calling on the artificial intelligence science community to devote time to research to make sure the advances have positive outcomes and can be controlled.
FLI's suggested research priorities include how to avoid AI automation from leading to job destruction and further income inequality, ethical questions around autonomous vehicle collisions, and the implications of autonomous weapons complying with humanitarian law.
"It's best to try to prevent a negative circumstance from occurring than to wait for it to occur and then be reactive," Musk said, according to The Verge. "This is a case where the range of negative outcomes, some of them are quite severe. It's not clear whether we'd be able to recover from some of these negative outcomes. In fact, you can construct scenarios where recovery of human civilization does not occur. When the risk is that severe, it seems like you should be proactive and not reactive."