November 30, 2019

411 words 2 mins read

Reliable prediction: Handling uncertainty

Reliable prediction: Handling uncertainty

Reliable prediction is the ability of a predictive model to explicitly measure the uncertainty involved in a prediction without feedback. Robin Senge shares two approaches to measure different types of uncertainty involved in a prediction.

Talk Title Reliable prediction: Handling uncertainty
Speakers Robin Senge (inovex)
Conference Strata Data Conference
Conf Tag Making Data Work
Location London, United Kingdom
Date May 23-25, 2017
URL Talk Page
Slides Talk Slides
Video

Using machine learning to create predictive models enables many new use cases that traditional software engineering approaches would have never been able to. This is great. In the last few years, we have begun to understand that we are able to utilize these models in almost all possible fields, including driving a car or diagnosing disease. However, unlike traditional (bug-free) computer systems, systems that are based on predictive models always include a threat: uncertainty. For instance, how will a trained deep neural network behave in a certain difficult situation when driving a car? The way that these systems are created and trained is to some extent similar to the way a human is trained: through experience. And as we all know, humans can certainly fail while learning a new task. Likewise these new systems can fail in situations they have not experienced before. Even worse, they will never be free from mistakes no matter how hard we train them, just as we ourselves won’t. Thus, a prerequisite for using and dealing with the uncertainty involved in an automated decision is being able to measure it. As Drucker notes, “If you cannot measure it, you cannot control it.“ Typically, uncertainty or the probability of error is measured by a loss function on a hold-out set of validation examples. A probability calculated this way enables us to decide whether to accept the evaluated model or decline its application in a productive environment. Reliable prediction is the ability of a predictive model to explicitly measure the uncertainty involved in a prediction without feedback. Robin Senge shares two approaches to measure different types of uncertainty involved in a prediction: conformal prediction by Shafer and Vovk and reliable classification by Senge and Hüllermeier. Besides precisely informing us about the overall uncertainty the prediction is contaminated with, both approaches identify different sources of uncertainty. Being able to distinguish these sources provides valuable information that can be used during model selection, feature selection, and even active learning scenarios. Both methods are implemented in Spark and are ready for use.

comments powered by Disqus