The paper presents the Sub-Adjacent Transformer, a novel approach for unsupervised time series anomaly detection. The key idea is to focus the attention mechanism on the sub-adjacent neighborhoods of each time point, rather than the immediate vicinity. This is based on the observation that anomalies typically exhibit more pronounced differences from their sub-adjacent neighborhoods compared to normal points.
The authors introduce two key concepts: sub-adjacent neighborhoods and sub-adjacent attention contribution. The sub-adjacent neighborhoods refer to the regions not immediately adjacent to the target point. The sub-adjacent attention contribution is defined as the sum of the attention weights in the corresponding column of the attention matrix, within the pre-defined sub-adjacent span.
To achieve the desired attention matrix pattern, the authors leverage linear attention instead of the traditional Softmax-based self-attention. They also propose a learnable mapping function within the linear attention framework to further enhance performance.
The Sub-Adjacent Transformer is evaluated on six real-world datasets and one synthetic benchmark, demonstrating state-of-the-art performance across various anomaly detection metrics. Ablation studies are conducted to validate the effectiveness of the key components, including the sub-adjacent attention mechanism, linear attention, and dynamic Gaussian scoring.
На другой язык
из исходного контента
arxiv.org
Дополнительные вопросы