Core Concepts
Instruction tuning large language models as rankers can significantly improve the performance of top-k recommendations by leveraging high-quality training data, position-aware prompts, and the integration of signals from conventional recommender systems.
Abstract
The paper introduces RecRanker, a framework that applies instruction-tuned large language models (LLMs) as rankers for top-k recommendations. The key highlights are:
Adaptive User Sampling: RecRanker employs importance-aware sampling, clustering-based sampling, and penalty for repetitive sampling to construct a high-quality, representative, and diverse instruction-tuning dataset.
Prompt Enhancement: RecRanker incorporates position shifting and augments prompts with signals from conventional recommendation models to enhance the LLM's contextual understanding and reasoning.
Hybrid Ranking: RecRanker utilizes a hybrid ranking method that ensembles pointwise, pairwise, and listwise ranking tasks to improve the model's performance and reliability.
The authors conduct extensive experiments on three real-world datasets, demonstrating the effectiveness of RecRanker in both direct and sequential recommendation scenarios. The proposed framework outperforms backbone models by a large margin, showcasing its significant superiority.
Stats
The number of users in the ML-100K, ML-1M, and BookCrossing datasets are 943, 6,040, and 1,820, respectively.
The number of items in the ML-100K, ML-1M, and BookCrossing datasets are 1,682, 3,706, and 2,030, respectively.
The number of ratings in the ML-100K, ML-1M, and BookCrossing datasets are 100,000, 1,000,209, and 41,456, respectively.
The density of the ML-100K, ML-1M, and BookCrossing datasets are 0.063046, 0.044683, and 0.011220, respectively.