The SBGD method introduces a novel swarm-based approach for global optimization of non-convex functions. The key aspects are:
Agents: Each agent is characterized by its position, x, and mass, m. The total mass of the swarm is conserved at 1.
Communication: Agents dynamically adjust their masses based on their relative heights. Agents at higher positions shed more mass, which is transferred to the current global minimizer. This creates a distinction between 'heavier' agents, which take smaller steps and are expected to converge to local minima, and 'lighter' agents, which take larger steps to explore the search space.
Time-stepping: The step size for each agent is determined by a backtracking line search protocol, where the step size is adjusted based on the agent's relative mass. Heavier agents take smaller steps, while lighter agents take larger steps.
The communication-based dynamics of SBGD allows it to effectively avoid local minima traps and explore the search space for the global minimum, outperforming traditional gradient descent methods, especially when the global minimum is located away from the initial swarm distribution.
The convergence analysis shows that the sequence of SBGD minimizers converges to a band of local minima, with a quantified convergence rate. Numerical experiments on one-, two-, and 20-dimensional benchmark problems demonstrate the effectiveness of SBGD as a global optimizer compared to other gradient descent methods.
לשפה אחרת
מתוכן המקור
arxiv.org
תובנות מפתח מזוקקות מ:
by Jingcheng Lu... ב- arxiv.org 05-01-2024
https://arxiv.org/pdf/2211.17157.pdfשאלות מעמיקות