Hand Gesture Recognition System
Naren Kumar
Department Of Information Technology, Maharaja Agrasen Institute of Technology
ABSTRACT
Hand gesture recognition systems are increasingly common in applications that require natural, touch-free human-computer interaction. These systems integrate object detection and environmental awareness to interpret hand movements, enabling users to perform tasks efficiently. This report focuses on a Python-based Hand Gesture Recognition System, where the camera of the device captures real-time hand gestures to control system functionalities like volume and brightness adjustments. Such systems hold transformative potential, especially in assistive technologies, where they can interpret hand gestures for non-verbal communication, enabling individuals with speech impairments to convey information.
The development of this system leverages advanced libraries, notably MediaPipe and OpenCV, both of which contribute essential functionalities. MediaPipe provides a foundational framework to create an exo-skeleton structure over the hand, allowing the system to track the hand’s precise movements, finger positions, and gestures in real-time. OpenCV, a powerful open-source library stemming from machine learning and computer vision research, enables the program to draw over live video feeds, giving immediate feedback to users as gestures are recognized and processed. This synergy between MediaPipe's gesture tracking and OpenCV's visualization capabilities allows the system to perform complex tasks smoothly and in real time.
Additionally, basic Python libraries like `math`, `numpy`, and `time` play a supportive role by handling various essential operations. These libraries manage calculations, array manipulation, and time-based functions that streamline processing and enhance performance. Custom modules, including `HandTracingModule` and `MyDict`, are incorporated into the system to improve efficiency by preventing code redundancy and simplifying access to frequently used data. The
`HandTracingModule` specifically manages the hand-tracking process, ensuring consistent, accurate hand detection across frames, while `MyDict` provides a structured way to handle recurring data essential to the gesture recognition logic.