We are part of a community of technologists, academics and policy-makers with a shared interest in safe and globally beneficial artificial intelligence (AI). We are working in partnership with them to shape public conversation in a productive way, foster new talent, and launch new centres like the Leverhulme Centre for the Future of Intelligence. Our research has addressed technical questions relevant to AI safety, the near-term and long-term security implications and governance of AI, and the potential for AI to help mitigate environmental and biological risks.
The field of AI is advancing rapidly. Recent years have seen dramatic breakthroughs in image and speech recognition, autonomous robotics, language tasks, and game playing. The coming decades will likely see substantial progress. This promises great benefits: new scientific discoveries, cheaper and better goods and services, medical advances. Our research and collaborations have explored applications of AI across a range of global challenges, including combating climate change, pandemic response, and food security.
AI also raises near-term concerns: privacy, bias, inequality, safety and security. CSER’s research has identified emerging threats and trends in global cybersecurity, and has explored challenges on the intersection of AI, digitisation and nuclear weapons systems.
AlphaGo Zero reached a superhuman level of performance after three days of self-play
Most current AI systems are ‘narrow’ applications – specifically designed to tackle a well-specified problem in one domain, such as a particular game. Such approaches cannot adapt to new or broader challenges without significant redesign. While it may be far superior to human performance in one domain, it is not superior in other domains. However, a long-held goal in the field has been the development of artificial intelligence that can learn and adapt to a very broad range of challenges.
AI in the longer term: opportunities and threats
As AI systems become more powerful and more general they may become superior to human performance in many domains. If this occurs, it could be a transition as transformative economically, socially, and politically as the Industrial Revolution. This could lead to extremely positive developments, but could also potentially pose catastrophic risks from accidents (safety) or misuse (security).
On safety: our current systems often go wrong in unpredictable ways. There are a number of difficult technical problems related to the design of accident-free artificial intelligence. Aligning current systems’ behaviour with our goals has proved difficult, and has resulted in unpredictable negative outcomes. Accidents caused by more powerful systems would be far more destructive.
On security: advanced AI systems could be key economic and military assets. Were these systems in the hands of bad actors, they might use it in harmful ways. If multiple groups competed to develop it first, it might have the destabilising dynamics of an arms race. Mitigating risk and achieving the global benefits of AI will present unique governance challenges, and will require global cooperation and representation.
Towards safe and beneficial transformative AI
There is great uncertainty and disagreement over timelines for the development of advanced AI systems. But whatever the speed of progress in the field, it seems like there is useful work that can be done right now. Technical machine learning research into safety is now being led by teams at OpenAI, DeepMind, and the Centre for Human-Compatible AI. AI governance research into the security implications is developing as a field.
The community working towards safe and beneficial superintelligence has grown worldwide. This has come from AI researchers showing leadership on this issue – supported by extensive discussions in machine learning labs and conferences, the landmark Puerto Rico conference, and high-profile support from people like CSER advisors Elon Musk and Stephen Hawking. We work closely with this community, in university labs and in tech companies to develop shared strategies to allow the benefits of AI advances to be safely realised.
More advanced and powerful AI systems will be developed and deployed in the coming years, these systems could be transformative with negative as well as positive consequences, and it seems that we can do useful work right now. While there are many uncertainties, we should dedicate serious effort and thought to laying the foundations for the safety of future systems and better understanding the implications of such advances.
To read about one of CSER's current projects, Paradigms of Artificial General Intelligence and Their Associated Risks, funded by the Future of Life Institute through the FLI International Safety Grants Competition, please click here.
Related team members
Related resources
-
Exploring AI Safety in Degrees: Generality, Capability and Control
Paper by John Burden, José Hernández-Orallo
-
-
-
-
-
Safeguarding the safeguards: How best to promote AI alignment in the public interest
Report by Oliver Guest, Michael Aird, Seán Ó hÉigeartaigh
-
The Societal Implications of Deep Reinforcement Learning
Peer-reviewed paper by Jess Whittlestone, Kai Arulkumaran, Matthew Crosby
-
-
-
The Scientometrics of AI Benchmarks: Unveiling the Underlying Mechanics of AI Research
Paper by Pablo Barredo, José Hernández-Orallo, Fernando Martínez-Plumed, Seán Ó hÉigeartaigh
-
Research community dynamics behind popular AI benchmarks
Peer-reviewed paper by Fernando Martínez-Plumed, Pablo Barredo, Seán Ó hÉigeartaigh, José Hernández-Orallo
-
-
Safety-driven design of machine learning for sepsis treatment
Paper by John Burden, Yan Jia, Tom Lawton, John McDermid, Ibrahim Habli
-
Does “AI” stand for augmenting inequality in the era of covid-19 healthcare?
Paper by David Leslie, Anjali Mazumder, Aidan Peppin, Maria K Wolters
-
-
-
General intelligence disentangled via a generality metric for natural and artificial intelligence
Peer-reviewed paper by José Hernández-Orallo, Bao Sheng Loe, Lucy Cheke, Fernando Martínez-Plumed, Seán Ó hÉigeartaigh
-
-
Using AI ethically to tackle covid-19
Paper by Stephen Cave, Jess Whittlestone, Rune Nyrup, Seán Ó hÉigeartaigh, Rafael A Calvo
-
Tackling threats to informed decision-making in democratic societies: Promoting epistemic security in a technologically-advanced world
Report by Elizabeth Seger, Shahar Avin, Gavin Pearson, Mark Briers, Seán Ó hÉigeartaigh, Helena Bacon
-
-
How General-Purpose Is a Language Model? Usefulness and Safety with Human Prompters in the Wild
Paper by John Burden, Pablo Antonio Moreno Casares, Bao Sheng Loe, José Hernández-Orallo, Seán Ó hÉigeartaigh
-
-
Your Prompt is My Command: On Assessing the Human-Centred Generality of Multimodal Models
Paper by Wout Schellaert, Fernando Martínez-Plumed, Karina Vold, John Burden, Pablo A. M. Casares, Bao Sheng Loe, Roi Reichart, Seán Ó hÉigeartaigh, Anna Korhonen, José Hernández-Orallo
-
International Governance of Civilian AI: A Jurisdictional Certification Approach
Report by Robert F. Trager, Ben Harack, Anka Reuel, Allison Carnegie, Lennart Heim, Lewis Ho, Sarah Kreps, Ranjit Lall, Owen Larter, Seán Ó hÉigeartaigh, Simon Staffell, José Jaime Villalobos
-
Filling gaps in trustworthy development of AI
Peer-reviewed paper by Shahar Avin, Haydn Belfield, Miles Brundage, Gretchen Krueger, Jasmine Wang, Adrian Weller, Markus Anderljung, Igor Krawczuk, David Krueger, Jonathan Lebensold, Tegan Maharaj, Noa Zilberman
-
Future Proof: the opportunity to transform the UK's resilience to extreme risks
Report by Toby Ord, Angus Mercer, Sophie Dannreuther, Haydn Belfield, Jess Whittlestone, Jade Leung, Markus Anderljung, Cassidy Nelson, Gregory Lewis, Piers Millett, Sam Hilton
-
-
Predictable Artificial Intelligence
Paper by Lexin Zhou, Pablo A. Moreno-Casares, Fernando Martínez-Plumed, John Burden, Ryan Burnell, Lucy Cheke, Cèsar Ferri, Alex Marcoci, Behzad Mehrbakhsh, Yael Moros-Daval, Seán Ó hÉigeartaigh, Danaja Rutar, Wout Schellaert, Konstantinos Voudouris, José Hernández-Orallo
-
Predicting and reasoningabout replicability usingstructured groups
Paper by Bonnie Wintle, Eden T. Smith, Martin Bush, Fallon Mody, David P. Wilkinson, Anca M. Hanea, Alex Marcoci, Hannah Fraser, Victoria Hemming, Felix Singleton Thorn, Marissa F. McBride, Elliot Gould, Andrew Head, Daniel G. Hamilton, Steven Kambouris, Libby Rumpff, Rink Hoekstra, Mark A. Burgman, Fiona Fidler
-
Computing Power and the Governance of Artificial Intelligence
Report by Girish Sastry, Lennart Heim, Haydn Belfield, Markus Anderljung, Miles Brundage, Julian Hazell, Cullen O’Keefe, Gillian K. Hadfield, Richard Ngo, Konstantin Pilz, George Gor, Emma Bluemke, Sarah Shoker, Janet Egan, Robert F. Trager, Shahar Avin, Adrian Weller, Yoshua Bengio, Diane Coyle
-
Model evaluation for extreme risks
Paper by Toby Shevlane, Sebastian Farquhar, Ben Garfinkel, Mary Phuong, Jess Whittlestone, Jade Leung, Daniel Kokotajlo, Nahema Marchal, Markus Anderljung, Noam Kolt, Lewis Ho, Divya Siddarth, Shahar Avin, Will Hawkins, Been Kim, Iason Gabriel, Vijay Bolina, Jack Clark, Yoshua Bengio, Paul Christiano, Allan Dafoe
-
Harms from Increasingly Agentic Algorithmic Systems
Paper by Alan Chan, Rebecca Salganik, Alva Markelius, Chris Pang, Nitarshan Rajkumar, Dmitrii Krasheninnikov, Lauro Langosco, Zhonghao He, Yawen Duan, Micah Carroll, Michelle Lin, Alex Mayhew, Katherine Collins, Maryam Molamohammadi, John Burden, Wanru Zhao, Shalaleh Rismani, Konstantinos Voudouris, Umang Bhatt, Adrian Weller, David Krueger, Tegan Maharaj
-
The Next Generation Event Horizon Telescope Collaboration: History, Philosophy, and Culture
Paper by Peter Galison, Juliusz Doboszewski, Jamee Elder, Niels Martens, Abhay Ashtekar, Jonas Enander, Marie Gueguen, Elizabeth A. Kessler, Roberto Lalli, Martin Lesourd, Alex Marcoci, Sebastián Murgueitio Ramírez, Priyamvada Natarajan, James Nguyen, Luis Reyes-Galindo, Sophie Ritson, Mike D. Schneider, Emilie Skulberg, Helene Sorgner, Matthew Stanley, Ann C. Thresher, Jeroen Van Dongen, James Weatherall, Jingyi Wu, Adrian Wüthrich
-
Toward Trustworthy AI: Mechanisms for Supporting Verifiable Claims
Report by Miles Brundage, Shahar Avin, Jasmine Wang, Haydn Belfield, Gretchen Krueger, Gillian Hadfield, Heidy Khlaaf, Jingying Yang, Helen Toner, Ruth Fong, Tegan Maharaj, Pang Wei Koh, Sara Hooker, Jade Leung, Andrew Trask, Emma Bluemke, Jon Lebensold, Cullen O'Keefe, Mark Koren, Théo Ryffel, JB Rubinovitz, Tamay Besiroglu, Federica Carugati, Jack Clark, Peter Eckersley, Sarah de Haas, Martiza Johnson, Ben Laurie, Alex Ingerman, Igor Krawczuk, Amanda Askell, Rosario Cammarota, Andrew Lohn, Shagun Sodhani, Charlotte Stix, Peter Henderson, Logan Graham, Carina Prunkl, Bianca Martin, Elizabeth Seger, Noa Zilberman, Seán Ó hÉigeartaigh, Frens Kroeger, Girish Sastry, Rebecca Kagan, Adrian Weller, Brian Tse, Beth Barnes, Allan Dafoe, Paul Scharre, Martijn Rasser, David Kreuger, Carrick Flynn, Ariel Herbert-Voss, Thomas Krendl Gilbert, Lisa Dyer, Saif Khan, Markus Anderljung, Yoshua Bengio
-
The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation
Peer-reviewed paper by Miles Brundage, Shahar Avin, Jack Clark, Helen Toner, Peter Eckersley, Ben Garfinkel, Allan Dafoe, Paul Scharre, Thomas Zeitzoff, Bobby Filar, Hyrum Anderson, Heather Roff, Gregory C. Allen, Jacob Steinhardt, Carrick Flynn, Seán Ó hÉigeartaigh, SJ Beard, Haydn Belfield, Sebastian Farquhar, Clare Lyle, Rebecca Crootof, Owain Evans, Michael Page, Joanna Bryson, Roman Yampolskiy, Dario Amodei
-
Bridging near- and long-term concerns about AI
Paper by Stephen Cave, Seán Ó hÉigeartaigh
-
-
-
Predicting reliability through structured expert elicitation with the repliCATS (Collaborative Assessments for Trustworthy Science) process
Paper by Hannah Fraser, Martin Bush, Bonnie Wintle, Fallon Mody, Eden T. Smith, Anca M. Hanea, Elliot Gould, Victoria Hemming, Daniel G. Hamilton, Libby Rumpff, David P. Wilkinson, Ross Pearson, Felix Singleton Thorn, Raquel Ashton, Aaron Willcox, Charles T. Gray, Andrew Head, Melissa Ross, Rebecca Groenewegen, Alexandru Marcoci, Ans Vercammen, Timothy H. Parker, Rink Hoekstra, Shinichi Nakagawa, David R. Mandel, Don van Ravenzwaaij, Marissa McBride, Richard O. Sinnott, Peter Vesk, Mark Burgman, Fiona Fidler
-
-
-
-
AI Paradigms and AI Safety: Mapping Artefacts and Techniques to Safety Issues
Paper by Jose Hernandez-Orallo, Fernando Martinez-Plumed, Shahar Avin, Jess Whittlestone, Seán Ó hÉigeartaigh
-
Accounting for the Neglected Dimensions of AI Progress
Paper by Fernando Martínez-Plumed, Shahar Avin, Miles Brundage, Allan Dafoe, Seán Ó hÉigeartaigh, José Hernández-Orallo
-
-
-
Response to the European Commission’s consultation on AI
Report by Haydn Belfield, José Hernández-Orallo, Seán Ó hÉigeartaigh, Matthijs M. Maas, Jess Whittlestone
-
A Proposal for International AI Governance
Report by Luke Kemp, Peter Cihon, Matthijs Michiel Maas, Haydn Belfield, Seán Ó hÉigeartaigh, Jade Leung
-
-
-
-
Surveying Safety-relevant AI Characteristics
Peer-reviewed paper by Jose Hernandez-Orallo, Fernando Martınez-Plumed, Shahar Avin, Seán Ó hÉigeartaigh
-
The Facets of Artificial Intelligence: A Framework to Track the Evolution of AI
Peer-reviewed paper by Fernando Martínez-Plumed, Bao Sheng Loe, Peter Flach, Seán Ó hÉigeartaigh, Karina Vold, José Hernández-Orallo
-
Accompanying technology development in the Human Brain Project: From foresight to ethics management
Peer-reviewed paper by Christine Aicardi, B. Tyr Fothergill, Stephen Rainey, Bernd Carsten Stahl, Emma Harris
-
-
Mapping Intelligence: Requirements and Possibilities
Peer-reviewed paper by Sankalp Bhatnagar, Anna Alexandrova, Shahar Avin, Stephen Cave, Lucy Cheke, Matthew Crosby, Jan Feyereisl, Marta Halina, Bao Sheng Loe, Seán Ó hÉigeartaigh, Fernando Martínez-Plumed, Huw Price, Henry Shevlin, Adrian Weller, Alan Winfield, José Hernández-Orallo
-
Ingredients for Understanding Brain and Behavioral Evolution: Ecology, Phylogeny, and Mechanism
Peer-reviewed paper by Stephen H. Montgomery, Adrian Currie, Dieter Lukas, Neeltje Boogert, Andrew Buskell, Fiona R. Cross, Sarah Jelbert, Shahar Avin, Rafael Mares, Ana F. Navarrete, Shuichi Shigeno, Corina J. Logan
-
Beyond Brain Size: Uncovering the Neural Correlates of Behavioral and Cognitive Specialization
Peer-reviewed paper by Corina J. Logan, Shahar Avin, Neeltje Boogert, Andrew Buskell, Fiona R. Cross, Adrian Currie, Sarah Jelbert, Dieter Lukas, Rafael Mares, Ana F. Navarrete, Shuichi Shigeno, Stephen H. Montgomery
-
-
Deep Learning: Artificial Intelligence Meets Human Intelligence
Video by Terrence Sejnowski
-
AI Safety: Past, Present, Future
Video by Victoria Krakovna
-
The Limits of AI
Video by Toby Walsh