Core Concepts
This research paper presents the first provably efficient algorithm for learning a single neuron under both adversarial label noise and distributional shifts, achieving an error bound of O(OPT) + ε by leveraging a novel primal-dual framework with sharpness and concentration analysis.
Stats
The algorithm achieves an error bound of O(OPT) + ε.
The sample complexity of the algorithm is ˜Ω(d/ε²).
The algorithm requires e^(O(d log(1/ε))) iterations to converge.
Quotes
"We study the problem of learning a single neuron with respect to the L2 loss in the presence of adversarial distribution shifts, where the labels can be arbitrary, and the goal is to find a “best-fit” function."
"Our algorithm follows a primal-dual framework and is designed by directly bounding the risk with respect to the original, nonconvex L2 loss."
"From an optimization standpoint, our work opens new avenues for the design of primal-dual algorithms under structured nonconvexity."