nuCarla: A nuScenes-Style Bird's-Eye View Perception Dataset for CARLA Simulation

arXiv — cs.CVWednesday, November 19, 2025 at 5:00:00 AM
  • The nuCarla dataset has been launched to enhance closed
  • The introduction of nuCarla is significant as it enables the development of more advanced end
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended Readings
Enhancing LLM-based Autonomous Driving with Modular Traffic Light and Sign Recognition
PositiveArtificial Intelligence
Large Language Models (LLMs) are being enhanced for autonomous driving with the introduction of TLS-Assist, a modular layer that improves traffic light and sign recognition. This innovation addresses the current limitations of LLM-based driving agents, which often struggle to detect critical safety objects. TLS-Assist translates detections into structured natural language messages, ensuring that safety cues are prioritized. The framework is adaptable to various camera setups and has been evaluated in a closed-loop environment using the LangAuto benchmark in CARLA.
PAVE: An End-to-End Dataset for Production Autonomous Vehicle Evaluation
PositiveArtificial Intelligence
The PAVE dataset represents a significant advancement in the evaluation of autonomous vehicles (AVs), being the first end-to-end benchmark dataset collected entirely through autonomous driving in real-world conditions. It includes over 100 hours of naturalistic data from various production AV models, segmented into 32,727 key frames with synchronized camera images and high-precision GNSS/IMU data. This dataset aims to enhance the understanding of AV behavior and safety, providing crucial insights for future developments in autonomous driving technology.
Divide and Merge: Motion and Semantic Learning in End-to-End Autonomous Driving
PositiveArtificial Intelligence
The article discusses a novel approach to end-to-end autonomous driving that separates semantic and motion learning to improve detection and tracking performance. The proposed method, Neural-Bayes motion decoding, utilizes learned motion queries in parallel with detection and tracking queries, enhancing information exchange through interactive semantic decoding. This addresses the negative transfer issue seen in multi-task learning, which can hinder performance in autonomous driving tasks.
Decoupling Scene Perception and Ego Status: A Multi-Context Fusion Approach for Enhanced Generalization in End-to-End Autonomous Driving
PositiveArtificial Intelligence
The article discusses the limitations of current end-to-end autonomous driving systems, which overly depend on ego status, affecting their ability to generalize and understand scenes robustly. It introduces AdaptiveAD, a new architectural solution that employs a dual-branch structure to decouple scene perception from ego status. This approach aims to enhance the performance of autonomous driving systems by allowing for more effective scene-driven reasoning without the influence of ego status.
CARScenes: Semantic VLM Dataset for Safe Autonomous Driving
PositiveArtificial Intelligence
CAR-Scenes is a frame-level dataset designed for autonomous driving, facilitating the training and evaluation of vision-language models (VLMs) for scene-level understanding. The dataset comprises 5,192 annotated images from sources like Argoverse, Cityscapes, KITTI, and nuScenes, utilizing a comprehensive 28-key category/sub-category knowledge base. The annotations are generated through a GPT-4o-assisted pipeline with human verification, providing detailed attributes and supporting semantic retrieval and risk-aware scenario mining.