Core Concepts
ILCiteR introduces evidence-grounded local citation recommendation for interpretability, leveraging distant supervision and pre-trained models.
Abstract
ILCiteR proposes a novel approach to local citation recommendation by focusing on interpretability through evidence spans. It utilizes a distantly-supervised evidence retrieval system and pre-trained Transformer-based Language Models for recommendations without explicit model training. The system re-ranks evidence spans based on query similarity and ranks associated papers for each span. Key contributions include a new dataset, conditional neural rank ensembling, and improved downstream paper recommendation performance.
Stats
Over 200,000 unique evidence spans in the dataset.
No explicit model training required for ILCiteR.
Performance improvements over lexical and semantic similarity based methods.