Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
opinion
Free access

Now, Later, and Lasting: 10 Priorities for AI Research, Policy, and Practice

Published: 07 May 2024 Publication History

Abstract

Shaping the future of artificial intelligence.
Advances in artificial intelligence (AI) will transform many aspects of our lives and society, bringing immense opportunities but also posing significant risks and challenges. The next several decades may well be a turning point for humanity, comparable to the industrial revolution. If so, future historians will judge how well we harnessed the benefits of AI for humanity, while protecting against potential harms. In this column, we share a set of recommendations for moving forward from the perspective of a founder and leaders of the One Hundred Year Study on AI.4 Launched 10 years ago with a dedicated endowment, the project is committed to a perpetual series of studies by multidisciplinary experts to evaluate the immediate, longer-term, and far-reaching effects of AI on people and society,1 and to make recommendations about AI research, policy, and practice.2,5 Beyond these recurrent studies and reports, our initiatives have included related efforts aimed at providing a diverse audience with insights about the trajectory of AI, including the creation of the AI Index, an annual benchmarking of AI progress.5
As we witness new capabilities emerging from neural models, it is crucial that we engage in efforts to advance our scientific understanding of these models and their behaviors. We must address AI’s impact on people and society through technical, social, and sociotechnical lenses, incorporating insights from a diverse range of experts including voices from engineering, social, behavioral, and economic disciplines. By fostering dialogue, collaboration, and action among various stakeholders, we can strategically guide the development and deployment of AI in ways that maximize its potential for contributing to human flourishing.
Despite the growing divide in the field between focusing on short-term versus long-term implications, we think both are of critical importance. As Alan Turing—one of the pioneers of AI—wrote in 1950, “We can only see a short distance ahead, but we can see plenty there that needs to be done.”6 We offer the following 10 recommendations for action that collectively address both the short- and long-term potential impacts of AI technologies:
1.
Scientific Foundations. Invest deeply in research to extend the scientific understanding of the capabilities and limitations of AI systems. Aim to ascertain the principles underlying their behaviors, identify and address deficiencies, and enable robust human guidance and oversight. Explore a diversity of alternative and complementary AI techniques and prioritize the study of the basis for sudden advances in AI capabilities, such as emergent phenomena observed in neural models as they are scaled.
2.
Safety, Reliability, Equity. Boost investments in R&D on identifying and mitigating concerns with safety, fairness, accuracy, and reliability of models, and encourage greater sharing and transparency from organizations building the largest-scale models, particularly regarding their methods, practices, results, and experiences. Advances in technology and policy can foster trust, accountability, and collaboration among the AI community and society at large.
3.
Regulation. Devise best practices, audits, and laws and regulations to incentivize reporting on, modifying, and addressing new capabilities and emergent phenomena to ensure that fielded models and systems are safe and responsibly deployed. These efforts can lead to appropriate oversight, feedback, and governance mechanisms aimed at aligning AI systems with human values and interests.
4.
Disinformation. Address uses of AI by malicious actors for disinformation, manipulation, and impersonation, which pose a risk to an informed citizenry and to the fabric of our democracies. Beyond harnessing promising technical and sociotechnical strategies built on content provenance, watermarking, and fingerprinting, investigate new laws, regulations, and enforcement to combat these threats while protecting civil liberties. The goal is to safeguard the integrity of information and communication against malevolent repurposing of widely available AI methods to deceive and persuade.
5.
Resources. Bridge the academia-industry gap to ensure greater access and enhanced collaboration for scholars and students so they can participate fully in frontier research and address key issues when they first arise rather than when it is too late. Address significant asymmetries in resources and access, spanning data, computation, and analytical tools and models, between academic and private-sector research scientists. Nurturing university-based research promises to assure the diversity, quality, and relevance of AI research and innovation.
6.
Vibrancy of Academy. Mitigate the brain drain of faculty and top potential students to industry by creating incentives and support structures to maintain the health and vibrancy of our educational institutions. Cultivating and retaining top computer-science talent at universities will be critical for research advancement and for teaching and mentoring the next generation of AI leaders.
7.
Jobs and Economy. Rigorously monitor the influences of AI technologies on jobs and the economy, and creatively pursue technologies and policies that promote shared prosperity rather than increased economic disparity. Prioritize research and development on innovations that bring AI and people together in new forms of collaborative work and that support the primacy of human agency and contribution. These efforts can foster social and economic inclusion and empowerment in the face of AI-driven automation and transformation.
8.
Diplomacy. Stand up and nurture international diplomatic efforts centered on AI, its implications, and uses. Engage with all nations, including authoritarian and autocratic governments, to jointly promote AI safety and protect human rights. Establish and incentivize the adoption of global norms for responsible AI practices, including recommended practices and prohibitions on uses of AI that jeopardize human rights. The aim is to foster global cooperation and coordination on AI issues that transcend national boundaries and interests.
9.
Deep Currents. Invest deeply in understanding and tracking subtle but potentially profound longer-term influences with psychological, social, and cultural impacts of AI technologies, such as human self-identity, agency, potential overreliance on AI, and psychological dependence on AI companion systems, as well as other emerging issues associated with new forms of intelligent systems in our daily lives. Engage in focused interdisciplinary research efforts to anticipate and address human and societal implications of AI that may not be immediately apparent or measurable.
10.
Catastrophic Outcomes. Prioritize the scrutiny of potential catastrophic outcomes associated with AI. Adopt a methodical, scientific approach to investigate these risks. Establish forward-looking strategies that incorporate failsafe designs, crucial safeguards, and preventative measures against severe disruptions. Anticipate, simulate, and “red team” threats stemming from unforeseen consequences of well-intentioned and malicious uses of AI. This includes the possibilities of using AI-powered protein design to create fast-spreading pathogens and trajectories leading to inescapable dependencies or irrevocable losses of human control over AI systems. Maintain constant vigilance via modeling and monitoring of potential pitfalls to ensure safety. Formulate effective oversight mechanisms and best practices. Introduce laws and regulations that preclude deployment of powerful AI systems that do not meet standards of due diligence. Proactive efforts on catastrophic risks can help to prevent and mitigate worst-case scenarios of AI.
We recognize these recommendations are high-level and broad, and that they do not address all the nuances and complexities of the issues. They are not meant to provide definitive answers or solutions, but rather to spark and guide discussions and actions among the AI community and society at large.
We must lean in vigorously on the now and the soon to address both short- and long-term issues to shape the future of AI for the common good. We can learn from what has gone well and poorly in previous technological revolutions, as well as assess what is unique about this one. We have the opportunity and the responsibility to invent and design our future, as we always have. This is not a passive or predetermined process, but an active and creative one, that depends on our choices, actions, and collaborations. By working together from our different perspectives and disciplines, we can shape the future of AI for the common good, and for the flourishing of generations to come.
Eric Horvitz ([email protected]) is cofounder of the One Hundred Year Study on AI. He is the Chief Scientific Officer of Microsoft, a member of the President’s Council of Advisors on Science and Technology, and Chair Emeritus of the Partnership on AI.
Vince Conitzer ([email protected]) is the chair of the Standing Committee of the One Hundred Year Study on AI. He is a professor of computer science at Carnegie Mellon University and Head of Technical AI Engagement at the Institute for Ethics in AI at the University of Oxford, U.K.
Sheila McIlraith ([email protected]) is the chair-elect of the Standing Committee of the One Hundred Year Study on AI. She is a professor of computer science at the University of Toronto, a Canada CIFAR AI Chair (Vector Institute), and Associate Director, Schwartz Reisman Institute for Technology and Society, Toronto, Canada.
Peter Stone ([email protected]) is the past-chair of the Standing Committee of the One Hundred Year Study on AI. He is a professor of computer science at the University of Texas at Austin and Executive Director of Sony AI America.

References

[1]
Horvitz, E. One-Hundred Year Study on Artificial Intelligence: Reflections and Framing (2014); https://ai100.stanford.edu/about/reflections-and-framing
[2]
Littman, M.L. et al. Gathering Strength, Gathering Storms: The One Hundred Year Study on Artificial Intelligence (AI100) 2021 Study Panel Report (2021); https://arxiv.org/abs/2210.15767
[3]
Maslej, N. et al. The AI Index Annual Report 2024 (2024); https://aiindex.stanford.edu/report/
[4]
One Hundred Year Study on Artificial Intelligence Homepage. Stanford University (2014); https://ai100.stanford.edu
[5]
Stone, P. et al. Artificial intelligence and Life in 2030: The One Hundred Year Study on Artificial Intelligence (2016); https://arxiv.org/abs/2211.06318
[6]
Turing, A.M. Computing Machinery and Intelligence, Mind, New Series 59, 236 (Oct. 1950), Oxford University Press.

Cited By

View all
  • (2024)Штучний інтелект як загальноцільова технологія: виклики та підходи до публічної політикиПроблеми сучасних трансформацій. Серія: право, публічне управління та адміністрування10.54929/2786-5746-2024-14-02-06Online publication date: 7-Nov-2024

Index Terms

  1. Now, Later, and Lasting: 10 Priorities for AI Research, Policy, and Practice

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image Communications of the ACM
      Communications of the ACM  Volume 67, Issue 6
      June 2024
      92 pages
      EISSN:1557-7317
      DOI:10.1145/3665906
      • Editor:
      • James Larus
      Issue’s Table of Contents
      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 23 May 2024
      Online First: 07 May 2024
      Published in CACM Volume 67, Issue 6

      Check for updates

      Qualifiers

      • Opinion

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)4,530
      • Downloads (Last 6 weeks)438
      Reflects downloads up to 10 Nov 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Штучний інтелект як загальноцільова технологія: виклики та підходи до публічної політикиПроблеми сучасних трансформацій. Серія: право, публічне управління та адміністрування10.54929/2786-5746-2024-14-02-06Online publication date: 7-Nov-2024

      View Options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Digital Edition

      View this article in digital edition.

      Digital Edition

      HTML Format

      View this article in HTML Format.

      HTML Format

      Get Access

      Login options

      Full Access

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media