Matthew Daniels, Autumn Toney, Melissa Flagg, and Charles Yang
| May 2021
The advantages of nations depend in part on their access to new inventions—and modern applications of artificial intelligence can help accelerate the creation of new inventions in the years ahead. This data brief is a first step toward understanding how modern AI and machine learning have begun accelerating growth across a wide array of science and engineering disciplines in recent years.
Leading U.S. companies are investing in the broad research field of artificial intelligence (AI), but where, specifically, are they making these investments? This data brief provides an analysis of the research papers published by Amazon, Apple, Facebook, Google, IBM, and Microsoft over the past decade to better understand what work their labs are prioritizing, and the degree to which these companies have similar or different research agendas overall. The authors find that major “AI companies” are often focused on very different subfields within AI, and that the private sector may be failing to make research investments consistent with ensuring long-term national competitiveness.
CSET Data Research Assistant Simon Rodriguez joins this episode of The Data Exchange to discuss how research in machine learning and AI affects public consciousness.
This paper is the third installment in a series on “AI safety,” an area
of machine learning research that aims to identify causes of
unintended behavior in machine learning systems and develop
tools to ensure these systems work safely and reliably. The first
paper in the series, “Key Concepts in AI Safety: An Overview,”
described three categories of AI safety issues: problems of
robustness, assurance, and specification. This paper introduces
interpretability as a means to enable assurance in modern machine
learning systems.
This paper is the second installment in a series on “AI safety,” an
area of machine learning research that aims to identify causes of
unintended behavior in machine learning systems and develop
tools to ensure these systems work safely and reliably. The first
paper in the series, “Key Concepts in AI Safety: An Overview,”
described three categories of AI safety issues: problems of
robustness, assurance, and specification. This paper introduces
adversarial examples, a major challenge to robustness in modern
machine learning systems.
This paper is the first installment in a series on “AI safety,” an area of machine learning research that aims to identify causes of unintended behavior in machine learning systems and develop tools to ensure these systems work safely and reliably. In it, the authors introduce three categories of AI safety issues: problems of robustness, assurance, and specification. Other papers in this series elaborate on these and further key concepts.
CSET Senior Fellow Andrew Lohn's work on machine learning vulnerabilities is cited in this article on the Department of Defense's efforts to adopt artificial intelligence technologies for a host of functions.
Zachary Arnold, Joanne Boisson, Lorenzo Bongiovanni, Daniel Chou, Carrie Peelman, and Ilya Rahkovsky
| February 2021
In this proof-of-concept project, CSET and Amplyfi Ltd. used machine learning models and Chinese-language web data to identify Chinese companies active in artificial intelligence. Most of these companies were not labeled or described as AI-related in two high-quality commercial datasets. The authors' findings show that using structured data alone—even from the best providers—will yield an incomplete picture of the Chinese AI landscape.
Based on news coverage alone, it can seem as if corporations dominate the research on artificial intelligence and machine learning when compared to the work of universities and academia. Authors Simon Rodriguez, Tim Hwang and Rebecca Gelles analyze the data over the past decade of research publications and find that, in fact, universities are the more dominant producers of AI papers. They also find that while corporations do tend to generate more citations to the work they publish in the field, these “high performing” papers are most frequently cross-collaborations with university labs.
Machine learning systems’ vulnerabilities are pervasive. Hackers and adversaries can easily exploit them. As such, managing the risks is too large a task for the technology community to handle alone. In this primer, Andrew Lohn writes that policymakers must understand the threats well enough to assess the dangers that the United States, its military and intelligence services, and its civilians face when they use machine learning.
This website uses cookies.
To learn more, please review this policy. By continuing to browse the site, you agree to these terms.
This website uses cookies to improve your experience while you navigate through the website. Out of these cookies, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may have an effect on your browsing experience.
Necessary cookies are absolutely essential for the website to function properly. This category only includes cookies that ensures basic functionalities and security features of the website. These cookies do not store any personal information.
Any cookies that may not be particularly necessary for the website to function and is used specifically to collect user personal data via analytics, ads, other embedded contents are termed as non-necessary cookies. It is mandatory to procure user consent prior to running these cookies on your website.