AI safety on whose terms?

Seth Lazar, Alondra Nelson

    Research output: Contribution to journalLetterpeer-review

    32 Citations (Scopus)

    Abstract

    Rapid, widespread adoption of the latest large language models has sparked both excitement and concern about advanced artificial intelligence (AI). In response, many are looking to the field of AI safety for answers. Major AI companies are purportedly investing heavily in this young research program, even as they cut "trust and safety" teams addressing harms from current systems. Governments are taking notice too. The United Kingdom just invested £100 million in a new "Foundation Model Taskforce" and plans an AI safety summit this year. And yet, as research priorities are being set, it is already clear that the prevailing technical agenda for AI safety is inadequate to address critical questions. Only a sociotechnical approach can truly limit current and potential dangers of advanced AI.
    Original languageEnglish
    JournalScience
    Volume381
    Issue number6654
    DOIs
    Publication statusPublished - 2023

    Fingerprint

    Dive into the research topics of 'AI safety on whose terms?'. Together they form a unique fingerprint.

    Cite this