The paper proposes a new Markov Decision Process (MDP) model to capture the user's response to the quality of ads in ad auctions. The key idea is to model the user state as a user-specific click-through rate (CTR) that changes in the next round based on the set of ads shown to the user in the current round.
The authors first characterize the optimal mechanism for this MDP setting as a Myerson's auction with a notion of modified virtual value, which takes into account both the current revenue and the future impact of showing the ad to the user. This optimal mechanism balances the short-term revenue considerations and the long-term effects on the user's propensity to click ads.
The authors then propose a simple second-price auction with personalized reserves as an approximation to the optimal mechanism. They show that this simple mechanism can achieve a constant-factor approximation to the optimal long-term discounted revenue, while maintaining the same user state transitions as the optimal mechanism. The key technical challenge is to design the personalized reserves in a way that controls the user state transitions and trades off the current round revenue with the long-term impact.
Finally, the authors provide experimental results comparing various natural auctions that incorporate user state.
Naar een andere taal
vanuit de broninhoud
arxiv.org
Belangrijkste Inzichten Gedestilleerd Uit
by Yang Cai,Zhe... om arxiv.org 05-07-2024
https://arxiv.org/pdf/2302.08108.pdfDiepere vragen