PAVE: An End-to-End Dataset for Production Autonomous Vehicle Evaluation

arXiv — cs.CVWednesday, November 19, 2025 at 5:00:00 AM
  • The PAVE dataset has been introduced as a groundbreaking resource for evaluating the safety and performance of autonomous vehicles, collected exclusively through autonomous driving in real
  • The introduction of the PAVE dataset is pivotal for the advancement of autonomous vehicle technology, as it provides a comprehensive framework for assessing the real
  • This development highlights a growing trend in the field of autonomous driving, where datasets like PAVE and nuCarla are addressing previous limitations in data collection methods. The emphasis on real
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
nuCarla: A nuScenes-Style Bird's-Eye View Perception Dataset for CARLA Simulation
PositiveArtificial Intelligence
The nuCarla dataset has been introduced as a large-scale, nuScenes-style bird's-eye view perception dataset designed for the CARLA simulation environment. This dataset addresses the limitations of existing datasets that primarily support open-loop learning by providing a closed-loop simulation framework. nuCarla is fully compatible with the nuScenes format, allowing for the transfer of real-world perception models, and offers a scale comparable to nuScenes, enhancing the training of end-to-end autonomous driving models.
Towards Sharper Object Boundaries in Self-Supervised Depth Estimation
PositiveArtificial Intelligence
Accurate monocular depth estimation is essential for understanding 3D scenes, yet current methods often produce blurred depth at object boundaries, leading to erroneous 3D points. This study introduces a self-supervised approach that models per-pixel depth as a mixture distribution, allowing for sharp depth discontinuities without fine-grained supervision. The method integrates variance-aware loss functions and uncertainty propagation, achieving up to 35% higher boundary sharpness and improved point cloud quality on KITTI and VKITTIv2 datasets.
Divide and Merge: Motion and Semantic Learning in End-to-End Autonomous Driving
PositiveArtificial Intelligence
The article discusses a novel approach to end-to-end autonomous driving that separates semantic and motion learning to improve detection and tracking performance. The proposed method, Neural-Bayes motion decoding, utilizes learned motion queries in parallel with detection and tracking queries, enhancing information exchange through interactive semantic decoding. This addresses the negative transfer issue seen in multi-task learning, which can hinder performance in autonomous driving tasks.
Decoupling Scene Perception and Ego Status: A Multi-Context Fusion Approach for Enhanced Generalization in End-to-End Autonomous Driving
PositiveArtificial Intelligence
The article discusses the limitations of current end-to-end autonomous driving systems, which overly depend on ego status, affecting their ability to generalize and understand scenes robustly. It introduces AdaptiveAD, a new architectural solution that employs a dual-branch structure to decouple scene perception from ego status. This approach aims to enhance the performance of autonomous driving systems by allowing for more effective scene-driven reasoning without the influence of ego status.
RTS-Mono: A Real-Time Self-Supervised Monocular Depth Estimation Method for Real-World Deployment
PositiveArtificial Intelligence
RTS-Mono is a newly proposed real-time self-supervised monocular depth estimation method aimed at enhancing autonomous driving and intelligent robot navigation. Traditional monocular depth estimation models often require significant computing resources, which can hinder their practical application. RTS-Mono addresses this issue with a lightweight encoder-decoder architecture, utilizing a Lite-Encoder and a multi-scale sparse fusion framework to optimize performance and inference speed, making it suitable for real-world deployment.
CARScenes: Semantic VLM Dataset for Safe Autonomous Driving
PositiveArtificial Intelligence
CAR-Scenes is a frame-level dataset designed for autonomous driving, facilitating the training and evaluation of vision-language models (VLMs) for scene-level understanding. The dataset comprises 5,192 annotated images from sources like Argoverse, Cityscapes, KITTI, and nuScenes, utilizing a comprehensive 28-key category/sub-category knowledge base. The annotations are generated through a GPT-4o-assisted pipeline with human verification, providing detailed attributes and supporting semantic retrieval and risk-aware scenario mining.