Concrete Problems in AI Safety (2016)
The 2016 paper 'Concrete Problems in AI Safety' by researchers from OpenAI and Google Brain transitioned the discussion of artificial intelligence safety from speculative philosophy to empirical engineering. Before this work, concerns about AI risk were often framed through the lens of 'superintelligence' or sci-fi scenarios that lacked a clear connection to modern machine learning. The authors argued that safety is not a separate domain of ethics, but a fundamental property of robust system design. By identifying specific, tractable failure modes - such as reward hacking and unintended side effects - they provided a technical roadmap for building systems that remain predictable and beneficial as they scale.























































































