Ensuring smarter-than-human intelligence has a positive outcome.
The field of artificial intelligence has made major strides in recent years, but there is a growing movement to consider the implications of machines that can rival humans in general problem-solving abilities. Nate Soares outlines the underresearched fundamental technical obstacles to building AI that can reliably learn to be "aligned" with human values.
Talk Title | Ensuring smarter-than-human intelligence has a positive outcome. |
Speakers | Nate Soares (MIRI) |
Conference | Artificial Intelligence Conference |
Conf Tag | Put AI to Work |
Location | San Francisco, California |
Date | September 18-20, 2017 |
URL | Talk Page |
Slides | Talk Slides |
Video | |
The field of artificial intelligence has made major strides in recent years. The social and cultural prospects of general artificial intelligence are often left to philosophers, ethicists, and science fiction authors to hash out, with relatively little involvement from researchers in the field. As a result, discussion of the field’s long-term impact is often divorced from the field’s realities and constraints and pays too little attention to the key technical questions for ensuring that AI systems are safe and reliable in practice. But there is a growing movement instigated by luminaries in science and industry to consider the implications of machines that can rival humans in general problem-solving abilities and our capacity to reason, learn, and devise plans. Nate Soares argues that the key questions for developing safe general AI are not always the obvious ones. Intuitively, we would expect more capable systems to be better at doing what we want and to therefore be safer than less capable systems. In many cases, however, capability gains (e.g., higher scores in machine learning tasks) can introduce surprising new failure modes. There are a number of underresearched technical obstacles to building machines that can reliably learn to promote our goals over time, over and above the technical obstacles to building machines that can reliably learn about strictly factual questions. Building on work by Nick Bostrom, Stuart Russell, and others, Nate outlines four basic propositions about general AI systems and shares current research into the problem of aligning such systems with our values.