Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 188 tok/s
Gemini 2.5 Pro 49 tok/s Pro
GPT-5 Medium 39 tok/s Pro
GPT-5 High 39 tok/s Pro
GPT-4o 78 tok/s Pro
Kimi K2 207 tok/s Pro
GPT OSS 120B 446 tok/s Pro
Claude Sonnet 4.5 35 tok/s Pro
2000 character limit reached

VR-based generation of photorealistic synthetic data for training hand-object tracking models (2401.17874v2)

Published 31 Jan 2024 in cs.CV

Abstract: Supervised learning models for precise tracking of hand-object interactions (HOI) in 3D require large amounts of annotated data for training. Moreover, it is not intuitive for non-experts to label 3D ground truth (e.g. 6DoF object pose) on 2D images. To address these issues, we present "blender-hoisynth", an interactive synthetic data generator based on the Blender software. Blender-hoisynth can scalably generate and automatically annotate visual HOI training data. Other competing approaches usually generate synthetic HOI data compeletely without human input. While this may be beneficial in some scenarios, HOI applications inherently necessitate direct control over the HOIs as an expression of human intent. With blender-hoisynth, it is possible for users to interact with objects via virtual hands using standard Virtual Reality hardware. The synthetically generated data are characterized by a high degree of photorealism and contain visually plausible and physically realistic videos of hands grasping objects and moving them around in 3D. To demonstrate the efficacy of our data generation, we replace large parts of the training data in the well-known DexYCB dataset with hoisynth data and train a state-of-the-art HOI reconstruction model with it. We show that there is no significant degradation in the model performance despite the data replacement.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (24)
  1. “Model based training, detection and pose estimation of texture-less 3d objects in heavily cluttered scenes,” International Journal of Computer Vision, vol. 97, no. 2, pp. 238–252, 2012.
  2. “Posecnn: A convolutional neural network for 6d object pose estimation in cluttered scenes,” 2018.
  3. “T-less: An rgb-d dataset for 6d pose estimation of texture-less objects,” IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), vol. 41, no. 12, pp. 2990–3004, 2019.
  4. “6-dof pose estimation of household objects for robotic manipulation: An accessible dataset and benchmark,” in International Conference on Intelligent Robots and Systems (IROS), 2022.
  5. “Homebreweddb: Rgb-d dataset for 6d pose estimation of 3d objects,” International Conference on Computer Vision (ICCV) Workshops, 2019.
  6. “DexYCB: A benchmark for capturing hand grasping of objects,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021.
  7. “Keypoint transformer: Solving joint identification in challenging hands and object interactions for accurate 3d pose estimation,” in CVPR, 2022.
  8. “gSDF: Geometry-Driven signed distance functions for 3D hand-object reconstruction,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023.
  9. “Chalet: Cornell house agent learning environment,” in Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), 2018.
  10. “Home: a household multimodal environment,” 2017.
  11. “Ai2-thor: An interactive 3d environment for visual ai,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 6195–6204.
  12. “Minos: Multimodal indoor simulator for navigation in complex environments,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 6402–6410.
  13. “Gibson env: Real-world perception for embodied agents,” in Proceedings of the IEEE Computer Vision and Pattern Recognition (CVPR), 2018.
  14. “Blenderproc2: A procedural pipeline for photorealistic rendering,” Journal of Open Source Software, vol. 8, no. 82, pp. 4901, 2023.
  15. “Hoisim: Synthesizing realistic 3d human-object interaction data for human activity recognition,” in 2021 30th IEEE International Conference on Robot & Human Interactive Communication (RO-MAN), 2021, pp. 1124–1131.
  16. “Virtualhome: Simulating household activities via programs,” in 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 8494–8503.
  17. “Letś play for action: Recognizing activities of daily living by learning from life simulation video games,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2021, pp. 8563–8569.
  18. “Eldersim: A synthetic data generation platform for human action recognition in eldercare applications,” IEEE Access, vol. 11, pp. 9279–9294, 2023.
  19. “Modular open robots simulation engine: Morse,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2011, pp. 1742–1748.
  20. “Are synthetic data useful for egocentric hand-object interaction detection? an investigation and the hoi-synth domain adaptation benchmark,” CoRR, vol. abs/2312.02672, 2023.
  21. “Vrkitchen: an interactive 3d virtual environment for task-oriented learning,” arXiv, vol. abs/1903.05757, 2019.
  22. “Threedworld: A platform for interactive multi-modal physical simulation,” 2021.
  23. “UnrealROX: An extremely photorealistic virtual reality environment for robotics simulations and synthetic data generation,” Virtual Reality, 2019.
  24. “Deephandmesh: A weakly-supervised deep encoder-decoder framework for high-fidelity hand mesh modeling,” in European Conference on Computer Vision (ECCV), 2020.

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.