Affordance Labeling and Exploration: A Manifold-Based Approach (2407.15479v1)
Abstract: The advancement in computing power has significantly reduced the training times for deep learning, fostering the rapid development of networks designed for object recognition. However, the exploration of object utility, which is the affordance of the object, as opposed to object recognition, has received comparatively less attention. This work focuses on the problem of exploration of object affordances using existing networks trained on the object classification dataset. While pre-trained networks have proven to be instrumental in transfer learning for classification tasks, this work diverges from conventional object classification methods. Instead, it employs pre-trained networks to discern affordance labels without the need for specialized layers, abstaining from modifying the final layers through the addition of classification layers. To facilitate the determination of affordance labels without such modifications, two approaches, i.e. subspace clustering and manifold curvature methods are tested. These methods offer a distinct perspective on affordance label recognition. Especially, manifold curvature method has been successfully tested with nine distinct pre-trained networks, each achieving an accuracy exceeding 95%. Moreover, it is observed that manifold curvature and subspace clustering methods explore affordance labels that are not marked in the ground truth, but object affords in various cases.
- Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, 1988.
- A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” Advances in neural information processing systems, vol. 25, 2012.
- K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” pp. 1–14, 2015.
- C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, “Going deeper with convolutions,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1–9, 2015.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016.
- J. Gibson, The Senses Considered as Perceptual Systems. Bloomsbury Academic, 1966.
- J. Gibson, The Ecological Approach to Visual Perception. Houghton Mifflin Harcourt (HMH), 1979.
- E. Ugur, Y. Nagai, E. Sahin, and E. Oztop, “Staged development of robot skills: Behavior formation, affordance learning and imitation with motionese,” IEEE Transactions on Autonomous Mental Development, vol. 7, no. 2, pp. 119–139, 2015.
- J. Modayil and B. Kuipers, “The initial development of object knowledge by a learning robot,” Robotics and autonomous systems, vol. 56, no. 11, pp. 879–890, 2008.
- H. S. Koppula, R. Gupta, and A. Saxena, “Learning human activities and object affordances from rgb-d videos,” vol. 32, pp. 951–970, 2013.
- R. Hartson, “Cognitive , physical , sensory , and functional affordances in interaction design,” Behavior & Information Technology, vol. 3001, pp. 315–338, 2003.
- A. Iriondo, E. Lazkano, and A. Ansuategi, “Affordance-based grasping point detection using graph convolutional networks for industrial bin-picking applications,” Sensors, vol. 21, no. 3, p. 816, 2021.
- H. Wu and G. S. Chirikjian, “Can i pour into it? robot imagining open containability affordance of previously unseen objects via physical simulations,” IEEE ROBOTICS AND AUTOMATION LETTERS, 8 2020.
- P. Ardón, M. E. Cabrera, E. Pairet, R. P. Petrick, S. Ramamoorthy, K. S. Lohan, and M. Cakmak, “Affordance-aware handovers with human arm mobility constraints,” IEEE Robotics and Automation Letters, vol. 6, no. 2, pp. 3136–3143, 2021.
- Z. Khalifa and S. A. A. Shah, “A large scale multi-view rgbd visual affordance learning dataset,” pp. 1325–1329, 2023.
- C.-Y. Chuang, J. Li, A. Torralba, and S. Fidler, “Learning to act properly: Predicting and explaining affordances from images,” arXiv preprint arXiv:1712.07576, 2017.
- W. Chen, H. Liang, Z. Chen, F. Sun, and J. Zhang, “Learning 6-dof task-oriented grasp detection via implicit estimation and visual affordance,” pp. 762–769, 2022.
- G. Li, V. Jampani, D. Sun, and L. Sevilla-Lara, “Locate: Localize and transfer object parts for weakly supervised affordance grounding,” in 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (Los Alamitos, CA, USA), pp. 10922–10931, IEEE Computer Society, jun 2023.
- Y. Yang, W. Zhai, H. Luo, Y. Cao, J. Luo, and Z. Zha, “Grounding 3d object affordance from 2d interactions in images,” in 2023 IEEE/CVF International Conference on Computer Vision (ICCV), (Los Alamitos, CA, USA), pp. 10871–10881, IEEE Computer Society, oct 2023.
- Y. Yang, W. Zhai, H. Luo, Y. Cao, and Z.-J. Zha, “Lemon: Learning 3d human-object interaction relation from 2d images,” 2023.
- T. V. Vo, M. N. Vu, B. Huang, T. Nguyen, N. Le, T. Vo, and A. Nguyen, “Open-vocabulary affordance detection using knowledge distillation and text-point correlation,” pp. 13508–13517, 2022.
- G. Li, D. Sun, L. Sevilla-Lara, and V. Jampani, “One-shot open affordance learning with foundation models,” 2023.
- E. Ragusa, S. Dosen, R. Zunino, and P. Gastaldo, “Affordance segmentation using tiny networks for sensing systems in wearable robotic devices,” IEEE Sensors Journal, vol. PP, pp. 1–1, 10 2023.
- C. Chen, A. Seff, A. Kornhauser, and J. Xiao, “Deepdriving: Learning affordance for direct perception in autonomous driving,” in Proceedings of the IEEE international conference on computer vision, pp. 2722–2730, 2015.
- M. Andries, A. Dehban, and J. Santos-Victor, “Automatic generation of object shapes with desired affordances using voxelgrid representation,” Frontiers in Neurorobotics, vol. 14, 5 2020.
- E. Ruiz and W. Mayol-Cuevas, “Geometric affordance perception: Leveraging deep 3d saliency with the interaction tensor,” Frontiers in Neurorobotics, vol. 14, p. 45, 2020.
- B. Moldovan, P. Moreno, M. Van Otterlo, J. Santos-Victor, and L. De Raedt, “Learning relational affordance models for robots in multi-object manipulation tasks,” in 2012 ieee international conference on robotics and automation, pp. 4373–4378, IEEE, 2012.
- S. Thermos, P. Daras, and G. Potamianos, “A deep learning approach to object affordance segmentation,” in ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2358–2362, IEEE, 2020.
- E. Ugur, E. Sahin, and E. Öztop, “Affordance learning from range data for multi-step planning.,” in EpiRob, 2009.
- E. Ugur, E. Oztop, and E. Sahin, “Goal emulation and planning in perceptual space using learned a ff ordances,” 2011.
- E. Ugur, M. Cakmak, and E. Sahin, “The learning and use of traversability affordance using range images on a mobile robot,” 2007.
- S. Bozeat, M. A. L. Ralph, K. Patterson, and J. R. Hodges, “When objects lose their meaning: What happens to their use?,” Cognitive, Affective, & Behavioral Neuroscience, vol. 2, pp. 236–251, 2002.
- G. Federico and M. A. Brandimonte, “Looking to recognise: the pre-eminence of semantic over sensorimotor processing in human tool use,” Scientific Reports, vol. 10, 12 2020.
- E. Şahin, M. Cakmak, M. R. Doğar, E. Uğur, and G. Üçoluk, “To afford or not to afford: A new formalization of affordances toward affordance-based robot control,” Adaptive Behavior, vol. 15, no. 4, pp. 447–472, 2007.
- D. Xu, A. Mandlekar, R. Martín-Martín, Y. Zhu, S. Savarese, and L. Fei-Fei, “Deep affordance foresight: Planning through what can be done in the future,” in 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 6206–6213, 2021.
- K. Ransikarbum, N. Kim, S. Ha, R. A. Wysk, and L. Rothrock, “A highway-driving system design viewpoint using an agent-based modeling of an affordance-based finite state automata,” IEEE Access, vol. 6, pp. 2193–2205, 2018.
- M. Hassanin, S. Khan, and M. Tahtali, “A new localization objective for accurate fine-grained affordance segmentation under high-scale variations,” IEEE Access, vol. 8, pp. 28123–28132, 2020.
- A. Myers, C. L. Teo, C. Fermüller, and Y. Aloimonos, “Affordance detection of tool parts from geometric features,” in 2015 IEEE International Conference on Robotics and Automation (ICRA), pp. 1374–1381, 2015.
- A. K. Pandey and R. Alami, “Affordance graph: A framework to encode perspective taking and effort based affordances for day-to-day human-robot interaction,” in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 2180–2187, 2013.
- A. Nguyen, D. Kanoulas, D. G. Caldwell, and N. G. Tsagarakis, “Detecting object affordances with convolutional neural networks,” in 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2765–2770, 2016.
- E. Ragusa, C. Gianoglio, S. Dosen, and P. Gastaldo, “Hardware-aware affordance detection for application in portable embedded systems,” IEEE Access, vol. 9, pp. 123178–123193, 2021.
- S. Thermos, G. Potamianos, and P. Daras, “Joint object affordance reasoning and segmentation in rgb-d videos,” IEEE Access, vol. 9, pp. 89699–89713, 2021.
- O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei, “ImageNet Large Scale Visual Recognition Challenge,” International Journal of Computer Vision (IJCV), vol. 115, no. 3, pp. 211–252, 2015.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” 2015.
- S. Xie, R. Girshick, P. Dollár, Z. Tu, and K. He, “Aggregated residual transformations for deep neural networks,” 2017.
- I. Radosavovic, R. P. Kosaraju, R. Girshick, K. He, and P. Dollár, “Designing network design spaces,” 2020.
- M. Tan and Q. V. Le, “Efficientnetv2: Smaller models and faster training,” 2021.
- A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, “An image is worth 16x16 words: Transformers for image recognition at scale,” 2021.
- A. Sekmen and B. Bilgin, “Manifold curvature estimation for neural networks,” in 2022 IEEE International Conference on Big Data (Big Data), pp. 3903–3908, IEEE, 2022.
- S. A. A. Shah and Z. Khalifa, “Hierarchical transformer for visual affordance understanding using a large-scale dataset,” in 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 11371–11376, 2023.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.