Core Concepts
Large language models (LLMs) struggle to coordinate the movements of multiple agents to achieve desired flocking patterns, due to their lack of spatial reasoning and understanding of collaborative behaviors.
Abstract
The paper investigates the challenges faced by large language models (LLMs) in solving the multi-agent flocking problem. Flocking is a behavior where multiple agents in a system attempt to stay close to each other while avoiding collision and maintaining a desired formation.
The authors set up various test scenarios with different numbers of agents and desired flocking patterns, such as circles, triangles, and lines. They use the GPT-3.5-Turbo model as the decision-maker for each agent, providing prompts that include the agent's role, game rules, and dynamic round descriptions.
The experiments reveal that LLMs like GPT-3.5-Turbo face significant challenges in solving the multi-agent flocking problem. Instead of maintaining the desired distance and formation, the agents typically opt to converge on the average of their initial positions or diverge from each other. The authors attribute this failure to the LLMs' lack of spatial awareness and reasoning, as well as their inability to understand the concept of maintaining a specific shape or distance in a meaningful way.
The paper highlights the urgent need for enhancing LLMs' capabilities in spatial reasoning and collaborative decision-making to address complex, real-world problems effectively.
Stats
The agents typically converge on the average of their initial positions or diverge from each other, failing to maintain the desired distance and formation.