COMPARATIVE ANALYSIS OF REAL-TIME GESTURE RECOGNITION METHODS BASED ON MEDIAPIPE, OPENCV, AND YOLOV8

Authors

  • Olha Yeroshenko
  • Vadym Tsipkovskyi

DOI:

https://doi.org/10.26906/SUNZ.2025.4.062

Keywords:

gesture recognition, computer vision, MediaPipe, OpenCV, YOLOv8, FPS, Detection Rate

Abstract

The subject of the article is real-time gesture recognition methods based on computer vision, designed for integration into human-computer interaction (HCI) systems, particularly for device control via a webcam. The purpose of the work is a comparative analysis of the efficiency of three modern methods MediaPipe, OpenCV, and YOLOv8 by evaluating their performance using key metrics (FPS and Detection Rate) for detecting basic gestures. The article solves the following tasks: identifying the impact of external factors (lighting, background) on gesture recognition, implementing algorithms for three basic gestures (open palm, index finger up and down), conducting experimental testing and modeling in the Python environment. The following methods are used: computer vision and image processing (in particular, segmentation, keypoint tracking, and object detection); machine learning based on CNN (LeNet, YOLO); analysis of datasets and performance metrics (precision, recall, mAP); simulation modeling in real conditions using the libraries mediapipe, opencv-python, and ultralytics. The following results were obtained: a comparative analysis of the methods was conducted, where MediaPipe provided the highest accuracy (95% Detection Rate), OpenCV—the maximum speed (50.7 FPS), and YOLOv8—a balance for limited resources (73% Detection Rate); recommendations for hybrid approaches to optimization were proposed. Conclusions: The developed comparative analysis of real-time gesture recognition methods demonstrates that MediaPipe effectively eliminates interference from variable lighting and background, achieving a stable accuracy of 95%, while OpenCV optimizes processing speed to 50.7 FPS. Modeling was performed in Python with visualization of the results.

Downloads

Download data is not yet available.

References

1. Suresh M., Sinha A., R. P. A. (2019), “Real-Time Hand Gesture Recognition Using Deep Learning”, “IJIIE - International Journal of Innovations & Implementations in Engineering”, Vol. 1, December Edition, pp. 11-15. URL: https://www.ijiie.org/download/IJIIE_2019DEC1003VOL1.pdf

2. Nguyen T.-H., Ngo B.-V., Nguyen T.-N. (2025), “Vision-Based Hand Gesture Recognition Using a YOLOv8n Model for the Navigation of a Smart Wheelchair”, “Electronics”, Vol. 14, № 4, p. 734. URL: https://www.mdpi.com/2079-9292/14/4/734

3. Zengeler N., Kopinski T., Handmann U. (2019), “Hand Gesture Recognition in Automotive Human–Machine Interaction Using Depth Cameras”, “Sensors”, Vol. 19, № 1, p. 59. doi: https://doi.org/10.3390/s19010059

4. Miao, Q., Li, Y., Liu, X., Liu, R. Gesture Recognition: Theory and Applications. Elsevier, 2024. 420 p.

5. Fedorchenko V., Yeroshenko O., Shmatko O., Kolomiitsev O., Omarov M. (2024), “Password hashing methods and algorithms on the .Net platform”, “Advanced Information Systems”, № 8(4), pp. 82–92, doi: https://doi.org/10.20998/2522-9052.2024.4.11

6. Elgendy, M. Deep Learning for Vision Systems. – Manning, 2020. – 480 p. – ISBN: 9781617296192.

Published

2025-12-02