This survey introduces a novel taxonomy for online anomaly detection in multivariate time series, making a distinction between online training and online inference. It presents an extensive overview of the state-of-the-art model-based online semi- and unsupervised anomaly detection approaches, categorizing them into different model families and other properties.
The survey also provides a detailed analysis of the most popular benchmark data sets used in the literature, highlighting their fundamental flaws, such as triviality, unrealistic anomaly density, uncertain labels, and run-to-failure bias. Additionally, it presents an extensive overview and analysis of the proposed evaluation metrics, discussing their strengths, weaknesses, and the need for parameter-free and interpretable metrics.
The biggest research challenge revolves around benchmarking, as currently there is no reliable way to compare different approaches against one another. This problem is two-fold: on the one hand, public data sets suffer from at least one fundamental flaw, while on the other hand, there is a lack of intuitive and representative evaluation metrics in the field. Moreover, the way most publications choose a detection threshold disregards real-world conditions, which hinders the application in the real world. To allow for tangible advances in the field, these issues must be addressed in future work.
Para Outro Idioma
do conteúdo original
arxiv.org
Principais Insights Extraídos De
by Luca... às arxiv.org 09-20-2024
https://arxiv.org/pdf/2408.03747.pdfPerguntas Mais Profundas