Distribution shift and OOD detection

Last updated on 2024-05-16 | Edit this page

Overview

Questions

  • What is distribution shift and what are its implications in machine learning models?
  • How can AI systems be deployed while ensuring that they do not break when encountered with new data?
  • What are some of the latest methods in detecting out-of-distribution data during inference?

Objectives

  • Understand the concept of distribution shift in the context of machine learning.
  • Identify the factors that can lead to distribution shift.
  • Explain the implications of distribution shift on the performance and reliability of machine learning models.
  • Discuss strategies for detecting and mitigating distribution shift in machine learning applications.
  • Identify best practices for robust model deployment and monitoring.
  • Examine the importance of detecting out-of-distribution data in machine learning applications.
  • Explore the limitations of traditional methods for detecting out-of-distribution data.
  • Investigate state-of-the-art techniques for out-of-distribution detection.

Key Points

  • TODO

Lesson content is being drafted here: https://docs.google.com/document/d/1BpNwJK4jRhnNJCzZSFaueMOB7jKJJHupZ92bLmQlqYQ/edit?usp=sharing

Distribution shift and its implications


Overview of out-of-distribution (OOD) detection methods


Glossary


  • ID/OOD: In-distribution, out-of-distribution. Generally, the OOD instances can be defined as instances (x, y) sampled from an underlying distribution other than the training distribution P(Xtrain, Ytrain), where Xtrain and Ytrain are the training corpus and training label set, respectively.
  • OOD instances with semantic shift: OOD instances with semantic shift refer to instances that do not belong to y_train. More specifically, instances with semantic shift may come from unknown categories or irrelevant tasks.
  • OOD instances with non-semantic shift: OOD instances with non-semantic shift refer to the instances that belong to Ytrain but are sampled from a distribution other than Xtrain, e.g., a different corpus.
  • Closed-world assumption: an assumption that the training and test data are sampled from the same distribution. However, training data can rarely capture the entire distribution. In real-world scenarios, out-of-distribution (OOD) instances, which come from categories that are not known to the model, can often be present in inference phases

References / Learn more