The paper presents a novel unsupervised anomaly detection framework called MFRNet that combines the advantages of feature-based and reconstruction-based methods. The key components are:
Multi-scale Feature Aggregator: A pre-trained model is used to extract multi-scale feature maps of the input image, which capture both low-level and high-level information.
Crossed-Mask Restoration Network: A restoration network is trained to recover the masked regions of the multi-scale feature maps. The masking is done using complementary crossed masks to ensure all potential anomalous regions are covered.
Hybrid Loss: A combination of contextual loss, SSIM loss, and gradient magnitude similarity loss is used to guide the training of the restoration network and measure the discrepancy between the input and reconstructed features.
The proposed MFRNet is able to learn more discriminative representations and prevent the model from over-generalizing to anomalies, leading to superior anomaly detection performance compared to state-of-the-art methods. Extensive experiments on five datasets, including a newly introduced Fabric-US dataset, demonstrate the effectiveness and generalization ability of MFRNet.
Para Outro Idioma
do conteúdo original
arxiv.org
Principais Insights Extraídos De
by Junpu Wang,G... às arxiv.org 04-23-2024
https://arxiv.org/pdf/2404.13273.pdfPerguntas Mais Profundas