The proposed DiceSGD algorithm uses a clipped error-feedback mechanism to eliminate the constant clipping bias in differentially private stochastic gradient descent (DPSGD) while maintaining the same level of privacy guarantee.
Larger subsampling rates (batch sizes) in differentially private stochastic gradient descent (DP-SGD) reduce the effective total gradient variance, explaining the empirical benefits of using large batch sizes.