0 citations
Key Concepts in AI Safety: An Overview
2021
Citations Over Time
Abstract
This paper is the first installment in a series on “AI safety,” an area of machine learning research that aims to identify causes of unintended behavior in machine learning systems and develop tools to ensure these systems work safely and reliably. In it, the authors introduce three categories of AI safety issues: problems of robustness, assurance, and specification. Other papers in this series elaborate on these and further key concepts.
Related Papers
- → A review of the unintended gender effects of international development efforts(2023)9 cited
- → Towards a Theory for Unintended Consequences in Engineering Design(2019)6 cited
- → 9/11’s legacy of unintended consequences(2023)2 cited
- → Dealing With Unintended Consequences(2020)
- → Risks and Unintended Consequences of the Reform(2023)