Papers
Topics
Authors
Recent
Search
2000 character limit reached

Online Action Representation using Change Detection and Symbolic Programming

Published 19 May 2024 in cs.CV | (2405.11511v1)

Abstract: This paper addresses the critical need for online action representation, which is essential for various applications like rehabilitation, surveillance, etc. The task can be defined as representation of actions as soon as they happen in a streaming video without access to video frames in the future. Most of the existing methods use predefined window sizes for video segments, which is a restrictive assumption on the dynamics. The proposed method employs a change detection algorithm to automatically segment action sequences, which form meaningful sub-actions and subsequently fit symbolic generative motion programs to the clipped segments. We determine the start time and end time of segments using change detection followed by a piece-wise linear fit algorithm on joint angle and bone length sequences. Domain-specific symbolic primitives are fit to pose keypoint trajectories of those extracted segments in order to obtain a higher level semantic representation. Since this representation is part-based, it is complementary to the compositional nature of human actions, i.e., a complex activity can be broken down into elementary sub-actions. We show the effectiveness of this representation in the downstream task of class agnostic repetition detection. We propose a repetition counting algorithm based on consecutive similarity matching of primitives, which can do online repetition counting. We also compare the results with a similar but offline repetition counting algorithm. The results of the experiments demonstrate that, despite operating online, the proposed method performs better or on par with the existing method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (11)
  1. A. Sabater, L. Santos, J. Santos-Victor, A. Bernardino, L. Montesano, and A. C. Murillo, “One-shot action recognition in challenging therapy scenarios,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).   Los Alamitos, CA, USA: IEEE Computer Society, jun 2021, pp. 2771–2779.
  2. Z. Zhao, S. Kiciroglu, H. Vinzant, Y. Cheng, I. Katircioglu, M. Salzmann, and P. Fua, “3d pose based feedback for physical exercises,” in Computer Vision – ACCV 2022, L. Wang, J. Gall, T.-J. Chin, I. Sato, and R. Chellappa, Eds.   Cham: Springer Nature Switzerland, 2023, pp. 189–205.
  3. M. Fieraru, M. Zanfir, S. C. Pirlea, V. Olaru, and C. Sminchisescu, “Aifit: Automatic 3d human-interpretable feedback models for fitness training,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 9914–9923.
  4. B. Dittakavi, D. Bavikadi, S. V. Desai, S. Chakraborty, N. Reddy, V. N. Balasubramanian, B. Callepalli, and A. Sharma, “Pose tutor: An explainable system for pose correction in the wild,” in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2022, pp. 3539–3548.
  5. S. Kulal, J. Mao, A. Aiken, and J. Wu, “Hierarchical motion understanding via motion programs,” 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6564–6572, 2021.
  6. O. Levy and L. Wolf, “Live repetition counting,” in 2015 IEEE International Conference on Computer Vision (ICCV), 2015, pp. 3020–3028.
  7. D. Dwibedi, Y. Aytar, J. Tompson, P. Sermanet, and A. Zisserman, “Counting out time: Class agnostic video repetition counting in the wild,” 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10 384–10 393, 2020.
  8. C. Lugaresi, J. Tang, H. Nash, C. McClanahan, E. Uboweja, M. Hays, F. Zhang, C.-L. Chang, M. G. Yong, J. Lee, W.-T. Chang, W. Hua, M. Georg, and M. Grundmann, “Mediapipe: A framework for building perception pipelines,” ArXiv, vol. abs/1906.08172, 2019.
  9. E. S. PAGE, “CONTINUOUS INSPECTION SCHEMES,” Biometrika, vol. 41, no. 1-2, pp. 100–115, 06 1954.
  10. M. Duarte, “detecta: A python module to detect events in data,” https://github.com/demotu/detecta, 2020.
  11. S. Kulal, J. Mao, A. Aiken, and J. Wu, “Programmatic concept learning for human motion description and synthesis,” in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).   Los Alamitos, CA, USA: IEEE Computer Society, jun 2022, pp. 13 833–13 842.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.