[Paper] Anomaly Detection with Adaptive and Aggressive Rejection for Contaminated Training Data

Published: (November 26, 2025 at 08:25 AM EST)
1 min read
Source: arXiv

Source: arXiv

Abstract

Handling contaminated data poses a critical challenge in anomaly detection, as traditional models assume training on purely normal data. Conventional methods mitigate contamination by relying on fixed contamination ratios, but discrepancies between assumed and actual ratios can severely degrade performance, especially in noisy environments where normal and abnormal data distributions overlap. To address these limitations, we propose Adaptive and Aggressive Rejection (AAR), a novel method that dynamically excludes anomalies using a modified z‑score and Gaussian mixture model‑based thresholds. AAR effectively balances the trade‑off between preserving normal data and excluding anomalies by integrating hard and soft rejection strategies. Extensive experiments on two image datasets and thirty tabular datasets demonstrate that AAR outperforms the state‑of‑the‑art method by 0.041 AUROC. By providing a scalable and reliable solution, AAR enhances robustness against contaminated datasets, paving the way for broader real‑world applications in domains such as security and healthcare.

Subjects

  • Machine Learning (cs.LG)
  • Artificial Intelligence (cs.AI)

Citation

arXiv: 2511.21378 (cs.LG)

DOI

10.48550/arXiv.2511.21378

Submission History

  • v1 – Wed, 26 Nov 2025 13:25:36 UTC (96 KB)
Back to Blog

Related posts

Read more »

It’s code red for ChatGPT

A smidge over three years ago, OpenAI threw the rest of the tech industry into chaos. When ChatGPT launched, even billed as a 'low-key research preview,' it bec...