Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
120 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

WayveScenes101: A Dataset and Benchmark for Novel View Synthesis in Autonomous Driving (2407.08280v1)

Published 11 Jul 2024 in cs.CV, cs.GR, and cs.RO

Abstract: We present WayveScenes101, a dataset designed to help the community advance the state of the art in novel view synthesis that focuses on challenging driving scenes containing many dynamic and deformable elements with changing geometry and texture. The dataset comprises 101 driving scenes across a wide range of environmental conditions and driving scenarios. The dataset is designed for benchmarking reconstructions on in-the-wild driving scenes, with many inherent challenges for scene reconstruction methods including image glare, rapid exposure changes, and highly dynamic scenes with significant occlusion. Along with the raw images, we include COLMAP-derived camera poses in standard data formats. We propose an evaluation protocol for evaluating models on held-out camera views that are off-axis from the training views, specifically testing the generalisation capabilities of methods. Finally, we provide detailed metadata for all scenes, including weather, time of day, and traffic conditions, to allow for a detailed model performance breakdown across scene characteristics. Dataset and code are available at https://github.com/wayveai/wayve_scenes.

Citations (1)

Summary

  • The paper presents WayveScenes101, a dataset of 101,000 images from 101 diverse driving scenes to benchmark novel view synthesis models.
  • It employs a five-camera rig and a unique off-axis evaluation protocol to assess model generalization under challenging real-world conditions.
  • The dataset overcomes limitations of existing benchmarks by capturing dynamic scenes, variable lighting, and complex geometries for improved autonomous driving research.

WayveScenes101: A Dataset and Benchmark for Novel View Synthesis in Autonomous Driving

The paper introduces WayveScenes101, a comprehensive dataset specifically curated to advance research in novel view synthesis within the context of autonomous driving. This dataset encompasses 101 distinct driving scenes captured under diverse environmental conditions, thereby presenting a broad spectrum of challenges including dynamic and deformable elements, changing geometries, and various lighting conditions. Such features make it a robust benchmark for evaluating state-of-the-art novel view synthesis models aimed at reconstructing real-world driving scenarios.

Key Features and Contributions

WayveScenes101 provides a substantial number of images (101,000) derived from 101 different scenes, each of 20 seconds duration. These are recorded using a five-camera rig positioned around a vehicle, capturing images at 10 frames per second. The dataset offers not only raw imagery but also COLMAP-derived camera poses and detailed metadata on each scene's conditions, enhancing its utility for model evaluation. An innovative evaluation protocol is proposed, which includes a held-out camera view for off-axis tests, focusing on the generalization ability of synthesis methods. Such a focus is crucial given the complexity associated with real-world driving scenes, where occlusion, rapid exposure changes, and dynamic obstacles are prevalent.

Comparison with Existing Datasets

This work situates itself amidst a landscape of existing datasets for novel view synthesis and autonomous driving assessment. While many current datasets like KITTI, Waymo, and Argoverse offer rich data for driving scenes, they do not adequately focus on view synthesis, often lacking the diversity and comprehensive camera setups necessary for this task. WayveScenes101 overcomes these limitations by specifically catering to the needs of novel view synthesis in the autonomous driving domain. It provides a wider array of scene types and camera configurations, emphasizing difficult reconstruction scenarios due to factors such as sun glare and pedestrian movement.

Evaluation Protocol and Metrics

The proposed evaluation protocol heavily relies on the assessment of off-axis reconstruction capability using the frontal camera, which is not part of the training set. This setup poses significant challenges due to the baseline distance between training and evaluation cameras, ensuring that models must effectively generalize across perspectives. Standard metrics such as PSNR, SSIM, LPIPS, and FID are used for quantitative evaluation, allowing for a detailed comparison of method efficacy. The dataset's metadata enables targeted evaluations across varied conditions, such as night or rain scenes.

Implications and Future Directions

WayveScenes101 holds substantial implications for the field of AI and computer vision, particularly within autonomous vehicle technology. By enabling more effective and realistic novel view synthesis, this work supports the refinement of driving models through improved synthetic scene generation, ultimately contributing to safer and more efficient autonomous systems. Practically, the dataset could enhance the design of systems that require rapid scene understanding from diverse viewpoints, a critical need in dynamic traffic environments.

Theoretically, WayveScenes101 encourages deeper exploration into handling dynamic elements and challenging exposure variations in novel view synthesis tasks. It has the potential to catalyze further research into improving the baseline synthesis methods and extending them to handle more heterogeneous datasets that encompass various global locational conditions, reinforcing their robustness and applicability.

In conclusion, WayveScenes101 is a significant asset to the automated driving and computer vision research communities, facilitating advancements in visual scene understanding and synthesis, which are pivotal to the evolution of autonomous technologies. Its application in future studies could spearhead notable improvements in both scene reconstruction methods and the operational capabilities of autonomous vehicles.

Github Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com