Papers
Topics
Authors
Recent
2000 character limit reached

HD Maps are Lane Detection Generalizers: A Novel Generative Framework for Single-Source Domain Generalization (2311.16589v2)

Published 28 Nov 2023 in cs.CV, cs.LG, and cs.RO

Abstract: Lane detection is a vital task for vehicles to navigate and localize their position on the road. To ensure reliable driving, lane detection models must have robust generalization performance in various road environments. However, despite the advanced performance in the trained domain, their generalization performance still falls short of expectations due to the domain discrepancy. To bridge this gap, we propose a novel generative framework using HD Maps for Single-Source Domain Generalization (SSDG) in lane detection. We first generate numerous front-view images from lane markings of HD Maps. Next, we strategically select a core subset among the generated images using (i) lane structure and (ii) road surrounding criteria to maximize their diversity. In the end, utilizing this core set, we train lane detection models to boost their generalization performance. We validate that our generative framework from HD Maps outperforms the Domain Adaptation model MLDA with +3.01%p accuracy improvement, even though we do not access the target domain images.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. F. Qiao, L. Zhao, and X. Peng, “Learning to learn single domain generalization,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 12 556–12 565.
  2. J. Wang, Y. Ma, S. Huang, T. Hui, F. Wang, C. Qian, and T. Zhang, “A keypoint-based global association network for lane detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 1392–1401.
  3. D. Li, Y. Yang, Y.-Z. Song, and T. M. Hospedales, “Deeper, broader and artier domain generalization,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 5542–5550.
  4. K. Muandet, D. Balduzzi, and B. Schölkopf, “Domain generalization via invariant feature representation,” in International conference on machine learning.   PMLR, 2013, pp. 10–18.
  5. H. Li, S. J. Pan, S. Wang, and A. C. Kot, “Domain generalization with adversarial feature learning,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 5400–5409.
  6. Q. Dou, D. Coelho de Castro, K. Kamnitsas, and B. Glocker, “Domain generalization via model-agnostic learning of semantic features,” Advances in neural information processing systems, vol. 32, 2019.
  7. Y. Li, X. Tian, M. Gong, Y. Liu, T. Liu, K. Zhang, and D. Tao, “Deep domain generalization via conditional invariant adversarial networks,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 624–639.
  8. G. Kang, L. Jiang, Y. Yang, and A. G. Hauptmann, “Contrastive adaptation network for unsupervised domain adaptation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 4893–4902.
  9. J. Liang, D. Hu, and J. Feng, “Do we really need to access the source data? source hypothesis transfer for unsupervised domain adaptation,” in International conference on machine learning.   PMLR, 2020, pp. 6028–6039.
  10. Z. Lu, Y. Yang, X. Zhu, C. Liu, Y.-Z. Song, and T. Xiang, “Stochastic classifiers for unsupervised domain adaptation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 9111–9120.
  11. C. Chen, W. Xie, W. Huang, Y. Rong, X. Ding, Y. Huang, T. Xu, and J. Huang, “Progressive feature alignment for unsupervised domain adaptation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 627–636.
  12. C. Hu, S. Hudson, M. Ethier, M. Al-Sharman, D. Rayside, and W. Melek, “Sim-to-real domain adaptation for lane detection and classification in autonomous driving,” in 2022 IEEE Intelligent Vehicles Symposium (IV).   IEEE, 2022, pp. 457–463.
  13. M.-Y. Liu, T. Breuel, and J. Kautz, “Unsupervised image-to-image translation networks,” Advances in neural information processing systems, vol. 30, 2017.
  14. C. Li, B. Zhang, J. Shi, and G. Cheng, “Multi-level domain adaptation for lane detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 4380–4389.
  15. X. Peng, F. Qiao, and L. Zhao, “Out-of-domain generalization from a single source: An uncertainty quantification approach,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022.
  16. S. Seo, Y. Suh, D. Kim, G. Kim, J. Han, and B. Han, “Learning to optimize domain specific normalization for domain generalization,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXII 16.   Springer, 2020, pp. 68–83.
  17. Q. Xu, R. Zhang, Y.-Y. Wu, Y. Zhang, N. Liu, and Y. Wang, “Simde: A simple domain expansion approach for single-source domain generalization,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 4797–4807.
  18. Z. Su, K. Yao, X. Yang, K. Huang, Q. Wang, and J. Sun, “Rethinking data augmentation for single-source domain generalization in medical image segmentation,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 2, 2023, pp. 2366–2374.
  19. Z. Wang, Y. Luo, R. Qiu, Z. Huang, and M. Baktashmotlagh, “Learning to diversify for single domain generalization,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 834–843.
  20. R. Volpi, H. Namkoong, O. Sener, J. C. Duchi, V. Murino, and S. Savarese, “Generalizing to unseen domains via adversarial data augmentation,” Advances in neural information processing systems, vol. 31, 2018.
  21. T. Gokhale, R. Anirudh, J. J. Thiagarajan, B. Kailkhura, C. Baral, and Y. Yang, “Improving diversity with adversarially learned transformations for domain generalization,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2023, pp. 434–443.
  22. X. Zhang, Q. Wang, J. Zhang, and Z. Zhong, “Adversarial autoaugment,” arXiv preprint arXiv:1912.11188, 2019.
  23. D. Hendrycks, N. Mu, E. D. Cubuk, B. Zoph, J. Gilmer, and B. Lakshminarayanan, “Augmix: A simple data processing method to improve robustness and uncertainty,” arXiv preprint arXiv:1912.02781, 2019.
  24. S. Yun, D. Han, S. J. Oh, S. Chun, J. Choe, and Y. Yoo, “Cutmix: Regularization strategy to train strong classifiers with localizable features,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 6023–6032.
  25. Z. Xu, D. Liu, J. Yang, C. Raffel, and M. Niethammer, “Robust and generalizable visual representation learning via random convolutions,” arXiv preprint arXiv:2007.13003, 2020.
  26. J. Jeong, Y. Cho, and A. Kim, “Hdmi-loc: Exploiting high definition map image for precise localization via bitwise particle filter,” IEEE Robotics and Automation Letters, vol. 5, no. 4, pp. 6310–6317, 2020.
  27. N. Deo, E. Wolff, and O. Beijbom, “Multimodal trajectory prediction conditioned on lane-graph traversals,” in Conference on Robot Learning.   PMLR, 2022, pp. 203–212.
  28. M. Heo, J. Kim, and S. Kim, “Hd map change detection with cross-domain deep metric learning,” in 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2020, pp. 10 218–10 224.
  29. Q. Li, Y. Wang, Y. Wang, and H. Zhao, “Hdmapnet: An online hd map construction and evaluation framework,” in 2022 International Conference on Robotics and Automation (ICRA).   IEEE, 2022, pp. 4628–4634.
  30. L. Van der Maaten and G. Hinton, “Visualizing data using t-sne.” Journal of machine learning research, vol. 9, no. 11, 2008.
  31. Z. Feng, S. Guo, X. Tan, K. Xu, M. Wang, and L. Ma, “Rethinking efficient lane detection via curve modeling,” in Computer Vision and Pattern Recognition, 2022.
  32. D. Jin, W. Park, S.-G. Jeong, H. Kwon, and C.-S. Kim, “Eigenlanes: Data-driven lane descriptors for structurally diverse lanes,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 17 163–17 171.
  33. Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE transactions on image processing, vol. 13, no. 4, pp. 600–612, 2004.
  34. Z. Wang, E. P. Simoncelli, and A. C. Bovik, “Multiscale structural similarity for image quality assessment,” in The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003, vol. 2.   Ieee, 2003, pp. 1398–1402.
  35. J. Snell, K. Ridgeway, R. Liao, B. D. Roads, M. C. Mozer, and R. S. Zemel, “Learning to generate images with perceptual similarity metrics,” in 2017 IEEE International Conference on Image Processing (ICIP).   IEEE, 2017, pp. 4277–4281.
  36. T.-H. Vu, H. Jain, M. Bucher, M. Cord, and P. Pérez, “Advent: Adversarial entropy minimization for domain adaptation in semantic segmentation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 2517–2526.
  37. Q. Lian, F. Lv, L. Duan, and B. Gong, “Constructing self-motivated pyramid curriculums for cross-domain semantic segmentation: A non-adversarial approach,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 6758–6767.
  38. M. Chen, H. Xue, and D. Cai, “Domain adaptation for semantic segmentation with maximum squares loss,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 2090–2099.
  39. V. Sushko, E. Schönfeld, D. Zhang, J. Gall, B. Schiele, and A. Khoreva, “You only need adversarial supervision for semantic image synthesis,” arXiv preprint arXiv:2012.04781, 2020.
  40. T.-C. Wang, M.-Y. Liu, J.-Y. Zhu, A. Tao, J. Kautz, and B. Catanzaro, “High-resolution image synthesis and semantic manipulation with conditional gans,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 8798–8807.
  41. E. Romera, J. M. Alvarez, L. M. Bergasa, and R. Arroyo, “Erfnet: Efficient residual factorized convnet for real-time semantic segmentation,” IEEE Transactions on Intelligent Transportation Systems, vol. 19, no. 1, pp. 263–272, 2017.
  42. X. Huang, M.-Y. Liu, S. Belongie, and J. Kautz, “Multimodal unsupervised image-to-image translation,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 172–189.
  43. A. Dosovitskiy, G. Ros, F. Codevilla, A. Lopez, and V. Koltun, “Carla: An open urban driving simulator,” in Conference on robot learning.   PMLR, 2017, pp. 1–16.

Summary

  • The paper proposes a single-source domain generalization framework that enhances lane detection diversity using HD maps and generative models.
  • The framework decomposes lane detection into lane structures and surroundings, optimizing data diversity with a tailored selection algorithm.
  • Experiments on CULane and TUSimple benchmarks show that the method outperforms existing techniques under diverse and challenging environmental conditions.

Improving Lane Detection Generalization: A Novel Framework using HD Maps for Boosting Diversity

The paper, "Improving Lane Detection Generalization: A Novel Framework using HD Maps for Boosting Diversity" by Daeun Lee, Minhyeok Heo, and Jiwon Kim, introduces an innovative framework designed to enhance the generalization capability of lane detection algorithms in autonomous driving systems. The authors focus on addressing the domain discrepancy that impacts lane detection performance when deployed in diverse and previously unseen road environments.

Summary

The core contribution of this work lies in the proposed framework for Single-Source Domain Generalization (SSDG) in lane detection. Traditional lane detection models suffer from performance degradation when tested on out-of-distribution data due to inherent differences in road environments like camera mounting positions, road widths, and varying lighting conditions. The framework presented leverages High-Definition (HD) maps and generative models to introduce a more diverse training dataset without relying on multiple source domains.

Key Contributions

  1. Novel Training Framework: The authors propose a novel framework aimed at generalizing a lane detection model using only a single-source domain. This approach is crucial for real-world autonomous driving applications where obtaining multi-source domain data is impractical.
  2. Decomposition into Lane Structures and Surroundings: The proposed method decomposes lane detection data into two primary components: lane structures and surrounding environments. Diversity within these components is enhanced separately using HD maps for lane data augmentation and generative models for surroundings augmentation.
  3. Data Selection Optimization: Instead of expanding the entire dataset, the framework identifies a core subset of data that maximizes diversity, thus optimizing the training data without redundant samples.
  4. Experimental Validation: Extensive experiments demonstrate the efficacy of the proposed framework, achieving superior generalization performance compared to existing domain adaptation techniques without using any prior target domain information.

Methodology

The framework meticulously addresses the problem of domain discrepancy by enhancing the diversity of lane structures and the visual surroundings. The method comprises several key phases:

  1. Lane Mask Extraction from HD Maps:
    • Lane markings are extracted and projected onto RGB images using camera parameters from HD maps.
    • This stage ensures that a consistent and diversified set of lane label masks is available without extensive manual annotations.
  2. Surrounding Diversity via Generative Models:
    • Generative models like OASIS and Pix2PixHD are employed to create realistic and varied surroundings for the lane label masks.
    • The synthetic images are selected based on their perceptual diversity using metrics like MS-SSIM to ensure the training data effectively captures diverse environmental conditions.
  3. Data Selection Algorithm:
    • A data selection algorithm based on graph theory is employed to identify the subset of images with the minimum sum of similarities, thus emphasizing diversity.

Results

The paper presents strong numerical results, illustrating significant improvements in generalization performance on benchmark datasets CULane and TUSimple. Specifically:

  • CULane Dataset: The framework achieved higher F1-scores compared to state-of-the-art domain adaptation methods, especially in challenging scenarios such as night scenes and dazzle light conditions.
  • TUSimple Dataset: The proposed method significantly outperformed existing methods, highlighting its effectiveness even when limited to a single source domain.

Implications and Future Developments

The implications of this research are twofold:

  1. Practical for Autonomous Driving:
    • The proposed SSDG approach is highly practical for autonomous driving, where pre-defining target domains is infeasible.
    • By enhancing the diversity of training data through methodical decomposition and augmentation, the framework provides a robust model capable of handling diverse road scenarios.
  2. Scalability and Flexibility:
    • The framework's reliance on HD maps and image synthesis models makes it scalable across various geographic regions without extensive manual data collection.
    • Future research could explore the integration of this framework with other autonomous driving tasks such as object detection and traffic sign recognition, potentially leading to a more comprehensive solution for domain generalization in autonomous systems.

Conclusion

This paper presents a methodically crafted framework to tackle the enduring challenge of domain discrepancy in lane detection. By leveraging HD maps and generative models to enhance training data diversity, the authors demonstrate that significant improvements in generalization performance can be achieved. The proposed framework is a notable contribution to the field of autonomous driving, providing a scalable and practical solution to ensure consistent lane detection across varying road environments.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.