Publications

CSET produces evidence-driven analysis in a variety of forms, from informative graphics and translations to expert testimony and published reports. Our key areas of inquiry are the foundations of artificial intelligence — such as talent, data and computational power — as well as how AI can be used in cybersecurity and other national security settings. We also do research on the policy tools that can be used to shape AI’s development and use, and on biotechnology.

Report

CSET’s 2024 Annual Report

Center for Security and Emerging Technology
| March 2025

In 2024, CSET continued to deliver impactful, data-driven analysis at the intersection of emerging technology and security policy. Explore our annual report to discover key research highlights, expert testimony, and new analytical tools — all aimed at shaping informed, strategic decisions around AI and emerging tech.

Filter publications

CSET's Jack Corrigan testified before the U.S.-China Economic and Security Review Commission where he discussed security threats posed by Chinese information and communications technology systems.

Formal Response

Comment on Advanced Computing Chips Rule

Jacob Feldgoise and Hanna Dohmen
| January 17, 2024

On January 17, 2024, CSET Researchers submitted a response to proposed rules from the Bureau of Industry and Security at the U.S. Department of Commerce. In the submission, CSET recommends that Commerce not implement controls on U.S. companies providing IaaS to Chinese entities, among other recommendations.

CSET submitted the following comment in response to a Request for Comment (RFC) from the Office of Management and Budget (OMB) about a draft memorandum providing guidance to government agencies regarding the appointment of Chief AI Officers, Risk Management for AI, and other processes following the October 30, 2023 Executive Order on AI.

Reports

Repurposing the Wheel: Lessons for AI Standards

Mina Narayanan, Alexandra Seymour, Heather Frase, and Karson Elmgren
| November 2023

Standards enable good governance practices by establishing consistent measurement and norms for interoperability, but creating standards for AI is a challenging task. The Center for Security and Emerging Technology and the Center for a New American Security hosted a series of workshops in the fall of 2022 to examine standards development in the areas of finance, worker safety, cybersecurity, sustainable buildings, and medical devices in order to apply the lessons learned in these domains to AI. This workshop report summarizes our findings and recommendations.

Data Brief

Spurring Science

Christian Schoeberl and Hanna Dohmen
| November 2023

This data brief analyzes over 200,000 U.S. government grants awarded to industry and academic recipients for artificial intelligence research between January 2017 and May 2023. The authors find that while the majority of federal grants are awarded to academic recipients, industry played an outsized role in U.S. government grant funding of AI research. Moreover, departments within the U.S. Department of Defense appear to prioritize funding industry and AI research relative to other funding agencies.

Data Snapshot

BIS Best Data Practices: Part 2

Christian Schoeberl
| November 16, 2023

Data Snapshots are informative descriptions and quick analyses that dig into CSET’s unique data resources. This is the second installment of a two-part series of data snapshots that explores export control data from the Bureau of Industry and Security (BIS), namely the 2021 and 2022 yearly reports for trade with China and Hong Kong.

Data Snapshot

BIS Best Data Practices: Part 1

Christian Schoeberl
| November 3, 2023

Data Snapshots are informative descriptions and quick analyses that dig into CSET’s unique data resources. This is the first installment of a two-part series of data snapshots that explores export control data from the Bureau of Industry and Security (BIS), namely the 2021 and 2022 yearly reports for trade with China and Hong Kong.

Reports

Decoding Intentions

Andrew Imbrie, Owen Daniels, and Helen Toner
| October 2023

How can policymakers credibly reveal and assess intentions in the field of artificial intelligence? Policymakers can send credible signals of their intent by making pledges or committing to undertaking certain actions for which they will pay a price—political, reputational, or monetary—if they back down or fail to make good on their initial promise or threat. Talk is cheap, but inadvertent escalation is costly to all sides.

Other

Techniques to Make Large Language Models Smaller: An Explainer

Kyle Miller and Andrew Lohn
| October 11, 2023

This explainer overviews techniques to produce smaller and more efficient language models that require fewer resources to develop and operate. Importantly, information on how to leverage these techniques, and many of the subsequent small models, are openly available online for anyone to use. The combination of both small (i.e., easy to use) and open (i.e., easy to access) could have significant implications for artificial intelligence development.

In collaboration with colleagues from CNAS and the Atlantic Council, CSET Researchers Ngor Luong and Emily Weinstein provided this comment in request to Treasury's Advanced Notice of Rule-making request for public comment (TREAS-DO-2023-0009-0001).