Open Letter for Robust AI Research

11 January 2015

Over 70 AI scientists and AI safety researchers have signed an open letter calling for more research on robust and beneficial AI. The signatories included Berkeley AI Professor Stuart Russell, president of AAAI Tom Dietterich, Microsoft Research Director Eric Horvitz, both co-founders of Vicarious, all three co-founders of Google Deepmind and a half-dozen other staff from Google including Peter Norvig, Laurent Orseau, and Blaise Aguera y Arcas. AI safety researchers Nick Bostrom, Eliezer Yudkowsky and Luke Muehlhauser have also signed on. From CSER, all three founders signed, as did our advisors Margaret Boden, Stephen Hawking, Elon Musk and Murray Shanahan. So have Max Tegmark and others from the Future of Life Institute.

The statement is as follows:

Artificial intelligence (AI) research has explored a variety of problems and approaches since its inception, but for the last 20 years or so has been focused on the problems surrounding the construction of intelligent agents – systems that perceive and act in some environment. In this context, “intelligence” is related to statistical and economic notions of rationality – colloquially, the ability to make good decisions, plans, or inferences. The adoption of probabilistic and decision-theoretic representations and statistical learning methods has led to a large degree of integration and cross-fertilization among AI, machine learning, statistics, control theory, neuroscience, and other fields. The establishment of shared theoretical frameworks, combined with the availability of data and processing power, has yielded remarkable successes in various component tasks such as speech recognition, image classification, autonomous vehicles, machine translation, legged locomotion, and question-answering systems.

As capabilities in these areas and others cross the threshold from laboratory research to economically valuable technologies, a virtuous cycle takes hold whereby even small improvements in performance are worth large sums of money, prompting greater investments in research. There is now a broad consensus that AI research is progressing steadily, and that its impact on society is likely to increase. The potential benefits are huge, since everything that civilization has to offer is a product of human intelligence; we cannot predict what we might achieve when this intelligence is magnified by the tools AI may provide, but the eradication of disease and poverty are not unfathomable. Because of the great potential of AI, it is important to research how to reap its benefits while avoiding potential pitfalls.

The progress in AI research makes it timely to focus research not only on making AI more capable, but also on maximizing the societal benefit of AI. Such considerations motivated the AAAI 2008-09 Presidential Panel on Long-Term AI Futures and other projects on AI impacts, and constitute a significant expansion of the field of AI itself, which up to now has focused largely on techniques that are neutral with respect to purpose. We recommend expanded research aimed at ensuring that increasingly capable AI systems are robust and beneficial: our AI systems must do what we want them to do. The attached research priorities document gives many examples of such research directions that can help maximize the societal benefit of AI. This research is by necessity interdisciplinary, because it involves both society and AI. It ranges from economics, law and philosophy to computer security, formal methods and, of course, various branches of AI itself.

In summary, we believe that research on how to make AI systems robust and beneficial is both important and timely, and that there are concrete research directions that can be pursued today.

The research priorities include law, ethics and economics research as well as computer science research for robust AI that is aligned with human interests.

Subscribe to our mailing list to get our latest updates