The paper presents a Retrieval-Augmented Transformer (RAT) model for click-through rate (CTR) prediction. The key insights are:
Traditional CTR prediction methods focus on modeling feature interactions within individual samples, but overlook the potential cross-sample relationships that can serve as a reference context to enhance the prediction.
To address this, the paper develops the RAT model, which retrieves similar samples as context and then builds Transformer layers with cascaded attention to capture both intra- and cross-sample feature interactions.
The cascaded attention design not only improves the efficiency compared to joint modeling, but also enhances the robustness of RAT.
Extensive experiments on real-world datasets demonstrate the effectiveness of RAT and suggest its advantage in long-tail scenarios, indicating its capability in addressing feature sparsity and cold start issues.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Yushen Li,Ji... at arxiv.org 04-04-2024
https://arxiv.org/pdf/2404.02249.pdfDeeper Inquiries