Core Concepts
A Deep Reinforcement Learning agent dynamically adjusts the cost function weights of a Social Force Window planner to enable socially compliant robot navigation in diverse environments.
Abstract
The paper proposes an adaptive social navigation system that combines a Social Force Window (SFW) planner with a Deep Reinforcement Learning (DRL) agent. The SFW planner integrates the classic Dynamic Window Approach (DWA) with a social cost based on the Social Force Model (SFM) to generate safe and human-aware trajectories.
The key innovation is the use of a DRL agent to dynamically adjust the weights of the cost function used by the SFW planner. The agent learns an optimal policy to set the weights of the social, obstacle, velocity, and other cost terms based on the local environmental conditions and task-specific features. This allows the planner to adapt its behavior to different social scenarios, such as pedestrian passing, overtaking, and crossing tasks in narrow and open spaces.
The authors extensively evaluate the proposed SFW-SAC (Soft Actor-Critic) approach in various Gazebo simulation environments and compare it to the baseline DWA and static SFW planners. The results demonstrate that the adaptive SFW-SAC planner outperforms the baselines in terms of success rate, navigation efficiency, and social compliance, as measured by metrics like clearance time, path length, average velocity, and social work. The adaptive planner is able to find a better trade-off between the different objectives, achieving more socially compliant navigation without compromising the overall performance.
The paper highlights the benefits of integrating classical navigation algorithms with learning-based methods to enhance the versatility and adaptability of service robots in complex social environments.
Stats
The robot has a maximum linear velocity of 0.6 m/s and a maximum angular velocity of 1.5 rad/s.
The LiDAR sensor has a maximum range of 3 m.
The robot is able to perceive up to 4 nearby people within a 5 m radius.
Quotes
"Machine Learning (ML) techniques represent a potential solution to this problem. ML models can leverage data to learn behaviors that enhance mobile robots' adaptability to new situations without being explicitly programmed for a specific task."
"The key idea of the proposed method lies in learning an optimal policy to dynamically set the weights of each objective function term used by the SFM local planner to score the simulated circular trajectories and select the next velocity command (v, w)."