Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Sign Language Recognition System with Pepper, Lightweight-Transformer, and LLM (2309.16898v1)

Published 28 Sep 2023 in cs.RO, cs.CL, cs.CV, and cs.HC

Abstract: This research explores using lightweight deep neural network architectures to enable the humanoid robot Pepper to understand American Sign Language (ASL) and facilitate non-verbal human-robot interaction. First, we introduce a lightweight and efficient model for ASL understanding optimized for embedded systems, ensuring rapid sign recognition while conserving computational resources. Building upon this, we employ LLMs for intelligent robot interactions. Through intricate prompt engineering, we tailor interactions to allow the Pepper Robot to generate natural Co-Speech Gesture responses, laying the foundation for more organic and intuitive humanoid-robot dialogues. Finally, we present an integrated software pipeline, embodying advancements in a socially aware AI interaction model. Leveraging the Pepper Robot's capabilities, we demonstrate the practicality and effectiveness of our approach in real-world scenarios. The results highlight a profound potential for enhancing human-robot interaction through non-verbal interactions, bridging communication gaps, and making technology more accessible and understandable.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. Bbc-oxford british sign language dataset. arXiv preprint arXiv:2111.03635.
  2. Language models for human-robot interaction. In ACM/IEEE International Conference on Human-Robot Interaction, March 13–16, 2023, Stockholm, Sweden, pages 905–906. ACM Digital Library.
  3. Sign pose-based transformer for word-level sign language recognition. In Proceedings of the IEEE/CVF winter conference on applications of computer vision, pages 182–191.
  4. CDC (2010). Identifying infants with hearing loss. https://www.cdc.gov/mmwr/preview/mmwrhtml/mm5908a2.htm [Accessed: (28 Sep 2023)].
  5. Moving away from robotic interactions: Evaluation of empathy, emotion and sentiment expressed and detected by computer systems. In 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), pages 1365–1370. IEEE.
  6. A survey on efficient convolutional neural networks and hardware acceleration. Electronics, 11(6):945.
  7. Hierarchical lstm for sign language translation. In Proceedings of the AAAI conference on artificial intelligence, volume 32.
  8. Language deprivation syndrome: a possible neurodevelopmental disorder with sociocultural origins. Soc. Psychiatry Psychiatr. Epidemiol., 52(6):761–776.
  9. Signbert: pre-training of hand-model-aware representation for sign language recognition. In Proceedings of the IEEE/CVF international conference on computer vision, pages 11087–11096.
  10. The effect of robot attentional behaviors on user perceptions and behaviors in a simulated health care interaction: randomized controlled trial. Journal of medical Internet research, 21(10):e13667.
  11. Technological solutions for sign language recognition: a scoping review of research trends, challenges, and opportunities. IEEE Access, 10:40979–40998.
  12. Method for multimodal recognition of one-handed sign language gestures through 3d convolution and lstm neural networks. In International Conference on Speech and Computer, pages 191–200. Springer.
  13. Social robot tutoring for child second language learning. In 2016 11th ACM/IEEE international conference on human-robot interaction (HRI), pages 231–238. IEEE.
  14. Artificial cognition for social human–robot interaction: An implementation. Artificial Intelligence, 247:45–69.
  15. Personalizing robot tutors to individuals’ learning differences. In Proceedings of the 2014 ACM/IEEE international conference on Human-robot interaction, pages 423–430.
  16. Word-level deep sign language recognition from video: A new large-scale dataset and methods comparison. In Proceedings of the IEEE/CVF winter conference on applications of computer vision, pages 1459–1469.
  17. Accessible options for deaf people in e-learning platforms: technology solutions for sign language translation. Procedia Computer Science, 67:263–272.
  18. Chasing the mythical ten percent: Parental hearing status of deaf and hard of hearing students in the united states. Sign language studies, 4(2):138–163.
  19. Deep convolutional neural networks for sign language recognition. In 2018 conference on signal processing and communication engineering systems (SPACES), pages 194–197. IEEE.
  20. Lsa64: An argentinian sign language dataset. In XXII Congreso Argentino de Ciencias de la Computación (CACIC 2016).
  21. Towards real-time sign language interpreting robot: Evaluation of non-manual components on recognition accuracy. In CVPR Workshops.
  22. Scheutz, M. (2011). 13 the inherent dangers of unidirectional emotional bonds between humans and social robots. Robot ethics: The ethical and social implications of robotics, page 205.
  23. Hospisign: an interactive sign language platform for hearing impaired. Journal of Naval Sciences and Engineering, 11(3):75–92.
  24. A new robotic platform for sign language tutoring: Humanoid robots as assistive game companions for teaching sign language. International Journal of Social Robotics, 7:571–585.
  25. Language to rewards for robotic skill synthesis.
  26. Rasa: A low-cost upper-torso social robot acting as a sign language teaching assistant. In Social Robotics: 8th International Conference, ICSR 2016, Kansas City, MO, USA, November 1-3, 2016 Proceedings 8, pages 630–639. Springer.
  27. Natural language-assisted sign language recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14890–14900.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. JongYoon Lim (8 papers)
  2. Inkyu Sa (24 papers)
  3. Bruce MacDonald (10 papers)
  4. Ho Seok Ahn (12 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.