Core Concepts
The author explores the significance of negative sampling methods in Knowledge Graph Representation Learning (KGRL) and categorizes them into static, dynamic, external model-based, and auxiliary data-based approaches to enhance the training process.
Abstract
The content delves into the importance of negative sampling techniques in KGRL for generating high-quality samples. It discusses various methods like Random, Probabilistic, External Model-Based, and Auxiliary Data-Based NS. The article highlights the pros and cons of each approach and their impact on efficiency, effectiveness, stability, independence, and quality.
The review covers a range of negative sampling strategies such as Uniform, Bernoulli, Nearest Neighbor (NN), Adaptive Negative Sampling (ANS), Entity-aware Negative Sampling (EANS), ϵ-Truncated UNS, Truncated NS, Distributional Negative Sampling (DNS), among others. Each method is analyzed based on its efficiency, effectiveness, stability, independence from side information, and quality of generated negatives.
Stats
Uniform [20] negative sampling is a prevalent approach.
Bernoulli [21] method substitutes head or tail entities based on relation mapping.
Probabilistic NS uses a fixed distribution to select negatives efficiently.
Nearest Neighbor [60] selects negatives close to positive triples in embedding space.
Adaptive NS [66] divides entities into clusters for better negative selection.
DNS technique utilizes entity type similarities for effective negative generation.
Quotes
"Generating high-quality negatives is essential in improving semantic learning."
"External Model-based NS generates semantically meaningful negatives."
"Probabilistic NS exhibits greater stability compared to Random NS methods."