Dynamic Distribution Shifts: OoD Detection with Dynamic Thresholds H/F

  • New computing paradigms, including quantum,
  • Internship
  • 3
  • CEA-List
  • Paris – Saclay
  • Level 7
  • 1900-01-01
Apply

The detection of out-of-distribution (OoD) samples is crucial for deploying deep learning (DL) models in real-world scenarios. OoD samples pose a challenge to DL models as they are not represented in the training data and can naturally arrive during deployment (i.e., a distribution shift), increasing the risk of obtaining wrong predictions. Consequently, OoD samples detection is crucial in safety-critical tasks, such as healthcare or automated vehicles, where trustworthy models are required.   The existing literature for the OoD detection problem focuses on the development of confidence scores where a threshold is applied to build a binary classifier to tell if a sample is in-distribution (InD) or OoD. In particular, the confidence score threshold is typically set using the values that correspond to InD samples, such that 95% of the confidence score values from InD samples fall above the selected thresholds, i.e., 95% True Positive Rate. However, setting a fixed threshold can lead to high False Positive Rate (FPR) values. In addition, even if the InD remains the same after deployment, the OoD could vary, resulting in FPR fluctuations. These two situations are of high interest in safety-critical applications, as misclassifying the confidence score value of an OoD sample as InD (False Positive) can result in more catastrophic consequences than misclassifying the confidence score value of an InD as OoD (False Negative).   To address the limitations and impact of a single fixed threshold selection, recent works propose using adaptive thresholds or a set of candidate thresholds to tackle the problem of dynamic distribution shifts. Specifically, in this internship position, we propose building on the work of Timans et al., who proposed a framework that leverages game theory and sequential hypothesis testing to assess the validity of a set of candidate thresholds. Therefore, the internship position aims to extend this work by exploring one or multiple of the following directions of improvement: Dynamic threshold selection (vs. fixed thresholds) Adaptive betting strategies (vs. static betting strategy) Adaptive windowing/batching (vs. fixed windows/batches size) Game theory methods: e.g., use of market-making algorithms (for threshold selection, and finding the optimal size of windows/batches)

Master students (M1/M2 – France) Proficiency in Python, NumPy, SciPy, sciki-tlearn, PyTorch,… Solid background in math, probability & statistics

Bac - Baccalauréat général

English Fluent

Related media

en_USEN

Contact us

We will reply as soon as possible...