Core Concepts
Evaluating the real-world performance of state-of-the-art perception systems that utilize Behavior Cloning (BC) for lateral control in autonomous driving.
Abstract
The paper addresses the gap between simulation-based testing and real-world validation of autonomous driving systems that use Behavior Cloning (BC) for lateral control. It presents a comprehensive evaluation of different BC-based methods, including Autoencoder-based Behavioral Cloning (AutoBC), Vision Transformers (ViT), and Spatial Attention mechanisms, in a real-world setting using a scaled research vehicle.
The key highlights and insights are:
The authors collected a dataset of 20,000 (image, steering angle) tuples using the scaled research vehicle on a designed racetrack.
They implemented and evaluated three BC-based methods: AutoBC, ViT with and without MLP, and AutoBC with Spatial Attention.
The ViT model without MLP and Spatial Attention (ViT without MLP and SIT) achieved the best performance, with the lowest Mean Absolute Error (MAE) of 0.0795, Mean Squared Error (MSE) of 0.0117, and Root Mean Squared Error (RMSE) of 0.1082.
The AutoBC model showed moderate performance, better than ViT with MLP and without SIT but not as accurate as ViT without MLP and SIT.
The AutoBC with Spatial Attention model performed poorly, likely due to the attention mask highlighting only the lane boundaries, which occupy a small portion of the image.
The models were further tested on unseen track configurations, such as a round 'O' map, to assess their generalization capabilities. The ViT model without MLP and SIT maintained the best performance, while the other methods showed deteriorated accuracy.
The results indicate that the ViT model is capable of making precise steering angle predictions, with a high percentage of predictions falling within acceptable error margins, even on unseen track configurations.
The study provides valuable insights into the real-world applicability and limitations of different BC-based methods for autonomous driving, contributing to the broader understanding of their performance and guiding future research in this field.
Stats
The dataset consists of 20,000 (image, steering angle) tuples collected using a scaled research vehicle on a designed racetrack.
The steering angle values range from -0.5 to +0.5 radians, with a maximum value of 0.5 radians (approximately 28.65 degrees to the right) and a minimum value of -0.5 radians (28.65 degrees to the left).
Quotes
"While numerous vision-based autonomous vehicle systems have been trained and evaluated in simulated environments, there is a notable lack of real-world validation for these systems."
"Despite advancements in simulation-based testing of BC models using end-to-end CNN algorithms for steering control, and other approaches employing BC as an image classification task, these studies do not fully capture real-world driving complexities."