METHODS FOR RECOGNISING HAND GESTURES USING COMPUTER VISION TO ASSIST PEOPLE WITH MOVEMENT OR SPEECH IMPAIRMENTS
DOI:
https://doi.org/10.26906/SUNZ.2026.1.135Keywords:
gesture, recognition, computer vision, MediaPipe Hands, neural networks, assistive technologiesAbstract
Relevance. Real-time hand gesture recognition systems play an important role in the development of modern technologies for human interaction with the “smart environment.” They create conditions for contactless device control and open up new communication opportunities for people with speech or motor impairments. The object of research. The object of research is a real-time hand gesture recognition system using computer vision and machine learning methods. The research problem is to create an accessible and effective means of communication for people with speech or motor impairments, as well as to expand the possibilities of controlling devices and elements of a “smart home” without physical contact or voice commands. Purpose of the article. The purpose of this article is to investigate and compare different methods for real-time hand gesture recognition using computer vision and machine learning. Specifically, the study evaluates the performance of MediaPipe Hands versus YOLOv8 for detecting and classifying hand gestures. The aim is to identify which approach provides higher accuracy and better adaptability for applications such as assistive communication for people with speech or motor impairments. Research results. In the course of the work, a prototype system was developed that combines video processing technologies (GStreamer or FFmpeg), the MediaPipe Hands library for determining key points of the hand, and CNN or LSTM deep learning models for classifying gestures and movement sequences. The results demonstrated the possibility of accurate gesture recognition in real time, as well as flexible adaptation to individual user characteristics. Conclusions. Interpretation of the results showed that the effectiveness of the system is due to the use of a comprehensive approach: the combination of 3D coordinate detection of key hand points with a neural network allowed high stability to be achieved even under changing lighting or background conditions. A distinctive feature of the proposed solution is the ability to recognise sequences of gestures that form phrases or commands (for example, “I want to drink”), as well as a function for teaching user gestures to expand the system’s individual vocabulary. The developed technology can be used in assistive communication systems for people with disabilities, in medical and rehabilitation facilities, as well as for controlling smart home elements. Its implementation contributes to increasing the accessibility of digital technologies and improving the quality of life of users.Downloads
References
1. Barkovska, O., Shapiro, A., Mavrynskyi, O., & Zhebin, P. (2025). Research on the sensitivity of the distance measurement method based on the facemesh algorithm. Системи управління, навігації та зв’язку. Збірник наукових праць, 2(80), 76-82. https://doi.org/10.26906/SUNZ.2025.2.076 DOI: https://doi.org/10.26906/SUNZ.2025.2.076
2. J. Shin, A. S. M. Miah, M. H. Kabir, M. A. Rahim and A. Al Shiam, "A Methodological and Structural Review of Hand Gesture Recognition Across Diverse Data Modalities," in IEEE Access, vol. 12, pp. 142606-142639, 2024, https://doi.org/10.1109/ACCESS.2024.3456436. DOI: https://doi.org/10.1109/ACCESS.2024.3456436
3. Zhang, T.; Wang, Y.; Zhou, X.; Liu, D.; Ji, J.; Feng, J. Intelligent Human–Computer Interaction for Building Information Models Using Gesture Recognition. Inventions 2025, 10, 5. https://doi.org/10.3390/inventions10010005. DOI: https://doi.org/10.3390/inventions10010005
4. Al Farid, F.; Hashim, N.; Abdullah, J.; Bhuiyan, M.R.; Shahida Mohd Isa, W.N.; Uddin, J.; Haque, M.A.; Husen, M.N. A Structured and Methodological Review on Vision-Based Hand Gesture Recognition System. Journal of Imaging 2022, 8, 153. https://doi.org/10.3390/jimaging8060153. DOI: https://doi.org/10.3390/jimaging8060153
5. Marques, P.; Váz, P.; Silva, J.; Martins, P.; Abbasi, M. Real-Time Gesture-Based Hand Landmark Detection for Optimised Mobile Photo Capture and Synchronisation. Electronics 2025, 14, 704. https://doi.org/10.3390/electronics14040704. DOI: https://doi.org/10.3390/electronics14040704
6. Barkovska, O., Ruban, I., Tymoshenko, D., Holovchenko, O., & Yankovskyi , O. (2025). Research on mobile machine learning platforms for human gesture recognition in human-machine interaction systems. Technology Audit and Production Reserves, 2(2(82), 6–14. https://doi.org/10.15587/2706-5448.2025.325423 DOI: https://doi.org/10.15587/2706-5448.2025.325423
7. M. Al-Hammadi et al., "Deep Learning-Based Approach for Sign Language Gesture Recognition With Efficient Hand Gesture Representation," in IEEE Access, vol. 8, pp. 192527-192542, 2020, https://doi.org/10.1109/ACCESS.2020.3032140. DOI: https://doi.org/10.1109/ACCESS.2020.3032140
8. Li, H.-H., & Hsieh, C.-C. (2025). Dynamic Hand Gesture Recognition Using MediaPipe and Transformer. Engineering Proceedings, 108(1), 22. https://doi.org/10.3390/engproc2025108022 DOI: https://doi.org/10.3390/engproc2025108022
9. Sun, Y., Zhang, Y., Wang, H. et al. SES-YOLOv8n: automatic driving object detection algorithm based on improved YOLOv8. SIViP 18, 3983–3992 (2024). https://doi.org/10.1007/s11760-024-03003-9 DOI: https://doi.org/10.1007/s11760-024-03003-9
10. R. Varghese and S. M., "YOLOv8: A Novel Object Detection Algorithm with Enhanced Performance and Robustness," 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 2024, pp. 1-6, https://doi.org/10.1109/ADICS58448.2024.10533619 DOI: https://doi.org/10.1109/ADICS58448.2024.10533619
Downloads
Published
Issue
Section
License
Copyright (c) 2026 Valerii Yarovyi, Olesia Barkovska, Dmytro Maksymov, Yana Ni, Daniil Raptanov

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.