Core Concepts
eTraM is a first-of-its-kind, fully event-based dataset that provides over 10 hours of annotated data from static traffic monitoring scenarios, covering a diverse range of traffic participants, lighting conditions, and weather conditions.
Abstract
The eTraM dataset is a comprehensive event-based traffic monitoring dataset that offers several key insights:
Data Acquisition:
The dataset was captured using the high-resolution Prophesee EVK4 HD event camera, strategically positioned at traffic intersections, roadways, and local streets.
The data collection process spanned 8 months, covering diverse lighting conditions (daytime, nighttime, twilight) and weather conditions (sunny, overcast, rainy).
Annotations and Statistics:
The dataset contains over 2 million bounding box annotations for 8 distinct classes of traffic participants, including vehicles (cars, trucks, buses, trams), pedestrians, and micro-mobility (bikes, bicycles, wheelchairs).
The annotations include object IDs, enabling the evaluation of multi-object tracking.
The dataset is split into 70% training, 15% validation, and 15% testing, ensuring proportional representation of each scene.
Baseline Evaluation:
The performance of state-of-the-art tensor-based methods (RVT and RED) and a frame-based method (YOLOv8) was evaluated on eTraM.
The results demonstrate the effectiveness of event-based models, particularly in nighttime conditions, where they outperform the frame-based method.
The evaluation also highlights the challenges and strengths of various traffic monitoring scenarios and categories.
Generalization Evaluation:
Experiments were conducted to assess the ability of event-based models to generalize to nighttime conditions and unseen traffic scenes.
The results show that models trained on a combination of daytime and nighttime data outperform those trained solely on daytime data, emphasizing the need for labeled nighttime data.
The models also exhibit strong generalization capabilities, performing similarly on held-in and held-out test sets, validating their transferability to new traffic environments.
Overall, eTraM stands as a valuable resource for the research community, enabling the exploration of event-based methods for traffic monitoring and paving the way for advancements in intelligent transportation systems.
Stats
The dataset contains over 2 million bounding box annotations for 8 distinct classes of traffic participants.
The average duration spent by objects from different classes ranges from 5 seconds for trams to 25 seconds for pedestrians and wheelchairs.
The performance of event-based models is impacted by the size of the objects, with medium-sized instances exhibiting the best performance across pedestrian and vehicle categories.
Quotes
"eTraM offers 10 hr of data from different traffic scenarios in various lighting and weather conditions, providing a comprehensive overview of real-world situations."
"eTraM's utility has been assessed using state-of-the-art methods for traffic participant detection, including RVT, RED, and YOLOv8."
"Our findings substantiate the compelling potential of leveraging event cameras for traffic monitoring, opening new avenues for research and application."