Disinformation

BBC News cited a report authored by CSET's Josh Goldstein, Micah Musser, and CSET alumna Katerina Sedova in partnership with OpenAI and Stanford Internet Observatory. Alongside the report, BBC News quoted Josh Goldstein regarding the current status of AI systems.

WIRED highlighted CSET Research Analyst Micah Musser in an article that references a report published by CSET, in collaboration with OpenAI and Stanford Internet Observatory. The report examines the potential misuse of language models in influence operations in the future and offers a framework for evaluating potential countermeasures.

Disinformation Researchers Raise Alarms About A.I. Chatbots

The New York Times
| February 8, 2023

A report by CSET's Josh Goldstein, Micah Musser, and CSET alumna Katerina Sedova in collaboration with OpenAI and Stanford Internet Observatory was cited in an article published by The New York Times about the potential dangers of AI-powered chatbots.

A report by CSET's Josh Goldstein, Micah Musser, and CSET alumna Katerina Sedova in collaboration with OpenAI and Stanford Internet Observatory was cited in an article published by Grid. The report examines the potential misuse of language models for influence operations in the future and proposes a structure for evaluating possible solutions to this problem.

A report by CSET’s Josh Goldstein, Micah Musser, and CSET alumna Katerina Sedova in collaboration with OpenAI and Stanford Internet Observatory was cited in an article published on Medium. The report explores how language models could be misused for influence operations in the future, and it provides a framework for assessing potential mitigation strategies.

Forecasting Potential Misuses of Language Models for Disinformation Campaigns—and How to Reduce Risk

Josh A. Goldstein Girish Sastry Micah Musser Renée DiResta Matthew Gentzel Katerina Sedova
| January 2023

Machine learning advances have powered the development of new and more powerful generative language models. These systems are increasingly able to write text at near human levels. In a new report, authors at CSET, OpenAI, and the Stanford Internet Observatory explore how language models could be misused for influence operations in the future, and provide a framework for assessing potential mitigation strategies.

In an interview with CyberScoop, Research Fellow Josh A. Goldstein discussed his research, in collaboration with Open AI and Stanford's Internet Observatory, on the use of large language models to deploy propaganda.

AI and the Future of Disinformation Campaigns

Katerina Sedova Christine McNeill Aurora Johnson Aditi Joshi Ido Wulkan
| December 2021

Artificial intelligence offers enormous promise to advance progress and powerful capabilities to disrupt it. This policy brief is the second installment of a series that examines how advances in AI could be exploited to enhance operations that automate disinformation campaigns. Building on the RICHDATA framework, this report describes how AI can supercharge current techniques to increase the speed, scale, and personalization of disinformation campaigns.

AI and the Future of Disinformation Campaigns

Katerina Sedova Christine McNeill Aurora Johnson Aditi Joshi Ido Wulkan
| December 2021

Artificial intelligence offers enormous promise to advance progress, and powerful capabilities to disrupt it. This policy brief is the first installment of a series that examines how advances in AI could be exploited to enhance operations that automate disinformation. Introducing the RICHDATA framework—a disinformation kill chain—this report describes the stages and techniques used by human operators to build disinformation campaigns.

Truth, Lies, and Automation

Ben Buchanan Andrew Lohn Micah Musser Katerina Sedova
| May 2021

Growing popular and industry interest in high-performing natural language generation models has led to concerns that such models could be used to generate automated disinformation at scale. This report examines the capabilities of GPT-3--a cutting-edge AI system that writes text--to analyze its potential misuse for disinformation. A model like GPT-3 may be able to help disinformation actors substantially reduce the work necessary to write disinformation while expanding its reach and potentially also its effectiveness.