The paper presents the Sub-Adjacent Transformer, a novel approach for unsupervised time series anomaly detection. The key idea is to focus the attention mechanism on the sub-adjacent neighborhoods of each time point, rather than the immediate vicinity. This is based on the observation that anomalies typically exhibit more pronounced differences from their sub-adjacent neighborhoods compared to normal points.
The authors introduce two key concepts: sub-adjacent neighborhoods and sub-adjacent attention contribution. The sub-adjacent neighborhoods refer to the regions not immediately adjacent to the target point. The sub-adjacent attention contribution is defined as the sum of the attention weights in the corresponding column of the attention matrix, within the pre-defined sub-adjacent span.
To achieve the desired attention matrix pattern, the authors leverage linear attention instead of the traditional Softmax-based self-attention. They also propose a learnable mapping function within the linear attention framework to further enhance performance.
The Sub-Adjacent Transformer is evaluated on six real-world datasets and one synthetic benchmark, demonstrating state-of-the-art performance across various anomaly detection metrics. Ablation studies are conducted to validate the effectiveness of the key components, including the sub-adjacent attention mechanism, linear attention, and dynamic Gaussian scoring.
Sang ngôn ngữ khác
từ nội dung nguồn
arxiv.org
Thông tin chi tiết chính được chắt lọc từ
by Wenzhen Yue,... lúc arxiv.org 05-01-2024
https://arxiv.org/pdf/2404.18948.pdfYêu cầu sâu hơn