The paper investigates the challenges faced by large language models (LLMs) in solving the multi-agent flocking problem. Flocking is a behavior where multiple agents in a system attempt to stay close to each other while avoiding collision and maintaining a desired formation.
The authors set up various test scenarios with different numbers of agents and desired flocking patterns, such as circles, triangles, and lines. They use the GPT-3.5-Turbo model as the decision-maker for each agent, providing prompts that include the agent's role, game rules, and dynamic round descriptions.
The experiments reveal that LLMs like GPT-3.5-Turbo face significant challenges in solving the multi-agent flocking problem. Instead of maintaining the desired distance and formation, the agents typically opt to converge on the average of their initial positions or diverge from each other. The authors attribute this failure to the LLMs' lack of spatial awareness and reasoning, as well as their inability to understand the concept of maintaining a specific shape or distance in a meaningful way.
The paper highlights the urgent need for enhancing LLMs' capabilities in spatial reasoning and collaborative decision-making to address complex, real-world problems effectively.
Til et annet språk
fra kildeinnhold
arxiv.org
Viktige innsikter hentet fra
by Peihan Li,Vi... klokken arxiv.org 04-09-2024
https://arxiv.org/pdf/2404.04752.pdfDypere Spørsmål