核心概念
LazyDP, an algorithm-software co-design, enables high-throughput training of differentially private recommendation models by addressing the compute and memory bottlenecks of the noise sampling and noisy gradient update operations in DP-SGD.
要約
The paper presents LazyDP, an algorithm-software co-design for training differentially private recommendation models. It first characterizes the computational challenges of training recommendation models using the standard DP-SGD algorithm, identifying the noise sampling and noisy gradient update stages as the key performance bottlenecks.
To address these bottlenecks, LazyDP proposes two key innovations:
Lazy noise update: LazyDP delays the noise update process for embedding vectors that are not accessed in the current training iteration, reducing the memory bandwidth required for the noisy gradient update. It ensures that the delayed noise updates are properly conducted before the embedding vectors are actually accessed, preserving the privacy guarantee of the baseline DP-SGD.
Aggregated noise sampling: LazyDP introduces a novel noise sampling algorithm that dramatically reduces the compute overhead of noise sampling by exploiting the mathematical properties of normally distributed random variables.
Through these algorithm-software co-design techniques, LazyDP provides an average 119x training throughput improvement over a state-of-the-art DP-SGD training system, while ensuring mathematically equivalent, differentially private recommendation models.
統計
The paper does not provide any specific numerical data or statistics to support the key claims. The focus is on the algorithmic and architectural innovations of LazyDP.
引用
The paper does not contain any direct quotes that are particularly striking or support the key arguments.