Bibliographic Information: Hu, X., Zhang, L., Meng, D., Han, Y., & Yuan, L. (2024). GITSR: Graph Interaction Transformer-based Scene Representation for Multi Vehicle Collaborative Decision-making. arXiv preprint arXiv:2411.01608.
Research Objective: This paper aims to address the challenge of effective scene representation and interaction modeling for multi-vehicle collaborative decision-making in autonomous driving, particularly within mixed traffic environments where Connected Automated Vehicles (CAVs) and Human Driving Vehicles (HDVs) coexist.
Methodology: The researchers propose the GITSR framework, which utilizes an agent-centric approach to scene representation. The framework employs a Transformer architecture to encode local scene information from an agent-centric perspective, capturing interactions between vehicles and their surroundings. Simultaneously, a GNN models the spatial interaction behaviors between vehicles based on their motion information. These representations are then fed into a Multi-Agent Deep Q-Network (MADQN) to generate collaborative driving actions. The framework is evaluated in a simulated highway dual-ramp exit scenario.
Key Findings: The GITSR framework demonstrates superior performance compared to baseline methods, including MADQN and MADQN with Transformer encoding only. GITSR achieves a higher task success rate, indicating its effectiveness in guiding CAVs to successfully navigate the designated ramps. Additionally, GITSR exhibits a lower number of collisions, highlighting its ability to promote safe driving behaviors. The ablation study reveals that agent-centric scene representation contributes to a safer driving strategy compared to scene-centric representation.
Main Conclusions: The integration of Transformer and GNN architectures within the GITSR framework effectively enhances scene representation and interaction modeling for multi-vehicle collaborative decision-making in autonomous driving. The agent-centric approach to scene representation proves beneficial for improving safety, while scene-centric representation demonstrates advantages in task completion efficiency.
Significance: This research contributes to the field of autonomous driving by proposing a novel framework for multi-vehicle collaborative decision-making that effectively addresses the challenges of scene representation and interaction modeling in mixed traffic environments.
Limitations and Future Research: The study acknowledges the computational burden associated with agent-centric scene representation, particularly in large-scale scenarios. Future research could explore more computationally efficient methods for scene representation without compromising safety and effectiveness. Additionally, investigating the framework's performance in more complex and realistic driving environments would be valuable.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Xingyu Hu, L... at arxiv.org 11-05-2024
https://arxiv.org/pdf/2411.01608.pdfDeeper Inquiries