Core Concepts
State space models offer an efficient solution for sequential recommendation, addressing the effectiveness-efficiency dilemma.
Abstract
Mamba4Rec introduces selective state space models for efficient sequential recommendation. It leverages a Mamba block with sequential modeling techniques to enhance performance and maintain efficiency. Experiments show superiority over RNN- and attention-based baselines in both effectiveness and efficiency. The model achieves Transformer-quality performance with better efficiency, especially on long sequences. Mamba4Rec demonstrates strong computational efficiency and memory cost reduction, making it effective for recommendation tasks.
Stats
Mamba4Rec outperforms RNN- and attention-based baselines in terms of both effectiveness and efficiency.
GPU memory usage: SASRec - 14.98GB, BERT4Rec - 15.48GB, Mamba4Rec - 4.82GB.
Training time per epoch: SASRec - 131.24s, BERT4Rec - 207.60s, Mamba4Rec - 75.52s.
Inference time per batch: SASRec - 0.27s, BERT4Rec - 0.90s, Mamba4Rec - 0.14s.
Quotes
"Sequential recommendation aims to estimate the dynamic user preferences and sequential dependencies among historical user behaviors."
"Mamba4Rec is able to well address the effectiveness-efficiency dilemma."
"The structured state matrix addresses long-range dependencies with HiPPO."
"Mamba achieves Transformer-quality performance with better efficiency."
"Mamba4Rec demonstrates superior performance compared to both RNN- and attention-based baselines."