MapRF: Weakly Supervised Online HD Map Construction via NeRF-Guided Self-Training

arXiv — cs.CVWednesday, November 26, 2025 at 5:00:00 AM
  • MapRF has been introduced as a weakly supervised framework for online high-definition (HD) map construction, utilizing Neural Radiance Fields (NeRF) to generate 3D maps from 2D image labels. This approach aims to overcome the limitations of existing methods that require expensive 3D map annotations, enhancing scalability and generalization across various driving environments.
  • The development of MapRF is significant as it allows for the iterative refinement of map networks through self-training, reducing dependency on additional supervision and potentially lowering the costs associated with HD map creation for autonomous driving systems.
  • This advancement aligns with ongoing efforts in the field of autonomous driving to improve mapping accuracy and efficiency, as seen in related innovations like PriorDrive and UniFlow, which also focus on enhancing mapping and scene understanding through novel methodologies and data integration.
— via World Pulse Now AI Editorial System

Was this article worth reading? Share it

Recommended apps based on your readingExplore all apps
Continue Readings
BEVDilation: LiDAR-Centric Multi-Modal Fusion for 3D Object Detection
PositiveArtificial Intelligence
A new framework named BEVDilation has been introduced, focusing on the integration of LiDAR and camera data for enhanced 3D object detection. This approach emphasizes LiDAR information to mitigate performance degradation caused by the geometric discrepancies between the two sensors, utilizing image features as implicit guidance to improve spatial alignment and address point cloud limitations.
DGGT: Feedforward 4D Reconstruction of Dynamic Driving Scenes using Unposed Images
PositiveArtificial Intelligence
The Driving Gaussian Grounded Transformer (DGGT) has been introduced as a novel framework for fast and scalable 4D reconstruction of dynamic driving scenes using unposed images, addressing the limitations of existing methods that require known camera calibration and per-scene optimization. This approach allows for reconstruction directly from sparse images and supports long sequences with multiple views.
LiDARCrafter: Dynamic 4D World Modeling from LiDAR Sequences
PositiveArtificial Intelligence
LiDARCrafter has been introduced as a unified framework for dynamic 4D world modeling from LiDAR sequences, addressing challenges in controllability, temporal coherence, and evaluation standardization. The framework utilizes natural language inputs to generate structured scene graphs, which guide a tri-branch diffusion network in creating object structures and motion trajectories.
nuScenes Revisited: Progress and Challenges in Autonomous Driving
PositiveArtificial Intelligence
The nuScenes dataset has been revisited, highlighting its pivotal role in the advancement of autonomous vehicles (AVs) and advanced driver assistance systems (ADAS). This dataset is notable for being the first to incorporate radar data and diverse urban driving scenes from multiple continents, collected using fully autonomous vehicles on public roads.
PolarGuide-GSDR: 3D Gaussian Splatting Driven by Polarization Priors and Deferred Reflection for Real-World Reflective Scenes
PositiveArtificial Intelligence
The introduction of PolarGuide-GSDR marks a significant advancement in 3D Gaussian Splatting (3DGS) by integrating polarization-aware techniques to enhance the rendering of reflective scenes. This method addresses existing challenges such as slow training and inefficient rendering, while also improving reflection reconstruction through a novel bidirectional coupling mechanism between polarization and 3DGS.
Alligat0R: Pre-Training Through Co-Visibility Segmentation for Relative Camera Pose Regression
PositiveArtificial Intelligence
A novel pre-training approach named Alligat0R has been introduced, focusing on co-visibility segmentation for relative camera pose regression, replacing the previous cross-view completion method. This technique enhances performance in both covisible and non-covisible regions by predicting pixel visibility across images, supported by the large-scale Cub3 dataset containing 5 million image pairs with dense annotations.