Problems of AI safety are the subject of increasing interest for engineers and policymakers alike. This brief uses the CSET Map of Science to investigate how research into three areas of AI safety — robustness, interpretability and reward learning — is progressing. It identifies eight research clusters that contain a significant amount of research relating to these three areas and describes trends and key papers for each of them.
By combining a versatile and frequently updated bibliometrics tool — the CSET Map of Science — with more hands-on analyses of technical developments, this brief outlines a methodology for measuring the publication growth of AI-related topics, where that growth is occurring, what organizations and individuals are involved, and when technical improvements in performance occur.
Data Snapshots are informative descriptions and quick analyses that dig into CSET’s unique data resources. Our first series of Snapshots introduced CSET’s Map of Science and explored the underlying data and analytic utility of this new tool, which enables users to interact with the Map directly.
To learn more, please review this policy. By continuing to browse the site, you agree to these terms.
Necessary cookies are absolutely essential for the website to function properly. This category only includes cookies that ensures basic functionalities and security features of the website. These cookies do not store any personal information.
Any cookies that may not be particularly necessary for the website to function and is used specifically to collect user personal data via analytics, ads, other embedded contents are termed as non-necessary cookies. It is mandatory to procure user consent prior to running these cookies on your website.