Core Concepts
DPP-based coresets can provably outperform independently drawn coresets in machine learning tasks by achieving better accuracy guarantees with a smaller coreset size.
Stats
The paper demonstrates the effectiveness of DPP-based coresets on the k-means problem, achieving a faster decay rate of QS(0.9) (quantile function of the supremum of relative error) compared to independent sampling methods.
In experiments with a synthetic uniform dataset, the DPP-based methods and Gaussian m-DPP achieved a faster decay rate (around m−3/4) than the m−1/2 rate of independent sampling.
The stratified sampling baseline, suitable for uniformly-spread datasets, achieved the best performance with a m−1 rate.
For a trimodal dataset, DPP-based samplers still outperformed independent ones, showcasing their effectiveness in capturing data distribution characteristics.
Experiments with the MNIST dataset (projected to 4 dimensions) showed a faster decay for DPP-based methods, but the advantage decreased as dimensionality increased.
Quotes
"DPPs can provably outperform independently drawn coresets."
"DPP-based coresets actually can achieve cardinality m = O(ε−2/(1+δ))."
"sampling with this DPP gives |LS(f)/L(f)−1| ≤m−(1/2+1/(2d))−, ∀f ∈F, where (1/2+1/(2d))− denotes any positive number strictly smaller than 1/2+1/(2d). Meanwhile, for i.i.d. sampling, the accuracy rate ε is at best m−1/2."