This survey introduces a novel taxonomy for online anomaly detection in multivariate time series, making a distinction between online training and online inference. It presents an extensive overview of the state-of-the-art model-based online semi- and unsupervised anomaly detection approaches, categorizing them into different model families and other properties.
The survey also provides a detailed analysis of the most popular benchmark data sets used in the literature, highlighting their fundamental flaws, such as triviality, unrealistic anomaly density, uncertain labels, and run-to-failure bias. Additionally, it presents an extensive overview and analysis of the proposed evaluation metrics, discussing their strengths, weaknesses, and the need for parameter-free and interpretable metrics.
The biggest research challenge revolves around benchmarking, as currently there is no reliable way to compare different approaches against one another. This problem is two-fold: on the one hand, public data sets suffer from at least one fundamental flaw, while on the other hand, there is a lack of intuitive and representative evaluation metrics in the field. Moreover, the way most publications choose a detection threshold disregards real-world conditions, which hinders the application in the real world. To allow for tangible advances in the field, these issues must be addressed in future work.
Sang ngôn ngữ khác
từ nội dung nguồn
arxiv.org
Thông tin chi tiết chính được chắt lọc từ
by Luca... lúc arxiv.org 09-20-2024
https://arxiv.org/pdf/2408.03747.pdfYêu cầu sâu hơn