Compute

Scaling AI

Andrew Lohn
| December 2023

While recent progress in artificial intelligence (AI) has relied primarily on increasing the size and scale of the models and computing budgets for training, we ask if those trends will continue. Financial incentives are against scaling, and there can be diminishing returns to further investment. These effects may already be slowing growth among the very largest models. Future progress in AI may rely more on ideas for shrinking models and inventive use of existing models than on simply increasing investment in compute resources.

Techniques to Make Large Language Models Smaller: An Explainer

Kyle Miller Andrew Lohn
| October 11, 2023

This explainer overviews techniques to produce smaller and more efficient language models that require fewer resources to develop and operate. Importantly, information on how to leverage these techniques, and many of the subsequent small models, are openly available online for anyone to use. The combination of both small (i.e., easy to use) and open (i.e., easy to access) could have significant implications for artificial intelligence development.

Onboard AI: Constraints and Limitations

Kyle Miller Andrew Lohn
| August 2023

Artificial intelligence that makes news headlines, such as ChatGPT, typically runs in well-maintained data centers with an abundant supply of compute and power. However, these resources are more limited on many systems in the real world, such as drones, satellites, or ground vehicles. As a result, the AI that can run onboard these devices will often be inferior to state of the art models. That can affect their usability and the need for additional safeguards in high-risk contexts. This issue brief contextualizes these challenges and provides policymakers with recommendations on how to engage with these technologies.

In an op-ed published in The Diplomat, Micah Musser discusses the concerns raised by policymakers in Washington about the disruptive potential of artificial intelligence technologies.

CSET Research Analyst Micah Musser and Institute for Progress Fellow Tim Hwang discussed CSET research examining factors that will contribute to future AI development.

Have Compute, Will Travel

Politico
| April 27, 2023

A CSET data brief by Micah Musser, Rebecca Gelles, Ronnie Kinoshita, Catherine Aiken, and Andrew Lohn was cited by Politico in a newsletter about the rapidly changing field of artificial intelligence and the debate surrounding its regulation.

“The Main Resource is the Human”

Micah Musser Rebecca Gelles Ronnie Kinoshita Catherine Aiken Andrew Lohn
| April 2023

Progress in artificial intelligence (AI) depends on talented researchers, well-designed algorithms, quality datasets, and powerful hardware. The relative importance of these factors is often debated, with many recent “notable” models requiring massive expenditures of advanced hardware. But how important is computational power for AI progress in general? This data brief explores the results of a survey of more than 400 AI researchers to evaluate the importance and distribution of computational needs.

AI and Compute

Andrew Lohn Micah Musser
| January 2022

Between 2012 and 2018, the amount of computing power used by record-breaking artificial intelligence models doubled every 3.4 months. Even with money pouring into the AI field, this trendline is unsustainable. Because of cost, hardware availability and engineering difficulties, the next decade of AI can't rely exclusively on applying more and more computing power to drive further progress.

CSET submitted this comment to the Office of Science and Technology Policy and the National Science Foundation to support the work of the National Artificial Intelligence Research Resource (NAIRR) Task Force to develop an implementation roadmap that would provide AI researchers and students across scientific disciplines access to computational resources, high-quality data, educational tools, and user support.