核心概念
CARLA is a novel two-stage self-supervised contrastive representation learning approach that effectively detects anomalies in time series data by learning discriminative representations that distinguish normal and anomalous patterns.
摘要
The paper introduces CARLA, a novel two-stage self-supervised contrastive representation learning approach for time series anomaly detection.
In the Pretext Stage, CARLA employs anomaly injection techniques to learn similar representations for temporally proximate windows and distinct representations for windows and their corresponding anomalous windows. This helps the model capture normal behavior and learn deviations indicating anomalies.
In the Self-supervised Classification Stage, CARLA leverages the learned representations to classify time series windows based on the proximity of their nearest and furthest neighbors in the representation space. This enhances the model's ability to differentiate between normal and anomalous patterns.
The key highlights of CARLA include:
- It addresses the challenge of lack of labeled data through a contrastive approach that leverages existing knowledge about different types of time series anomalies.
- It proposes an effective contrastive method to learn feature representations by injecting various types of anomalies as negative samples.
- It introduces a self-supervised classification method that utilizes the learned representations to classify time series windows based on their nearest and furthest neighbors.
- Extensive experiments on seven real-world benchmark datasets show CARLA's superior performance over a range of state-of-the-art unsupervised, semi-supervised, and self-supervised contrastive learning models for both univariate and multivariate time series.
統計資料
"One main challenge in time series anomaly detection (TSAD) is the lack of labelled data in many real-life scenarios."
"Most of the existing anomaly detection methods focus on learning the normal behaviour of unlabelled time series in an unsupervised manner."
"The normal boundary is often defined tightly, resulting in slight deviations being classified as anomalies, consequently leading to a high false positive rate and a limited ability to generalise normal patterns."
引述
"We argue that these assumptions are limited as augmentation of time series can transform them to negative samples, and a temporally distant window can represent a positive sample."
"Existing approaches to contrastive learning for time series have directly copied methods developed for image analysis. We argue that these methods do not transfer well."
"Our contrastive approach leverages existing generic knowledge about time series anomalies and injects various types of anomalies as negative samples."