Skip to content

sri-dsa/Hand-Gesture-Recognition-Control

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

40 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

Hand Gesture Recognition & Control

Overview

Hand Gesture Recognition & Control is an advanced AI-powered system that enables real-time recognition and interpretation of hand gestures for various applications such as gesture-controlled drones, AI-based human-computer interaction, and assistive technologies. This project integrates multiple deep learning techniques, including Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTMs), MediaPipe, YOLO, and ONNX, to achieve robust gesture recognition and control.


Table of Contents


Features

๐Ÿ”น Real-Time Hand Gesture Recognition

  • Utilizes CNN-based models and MediaPipe for robust hand tracking.
  • Supports multiple hand gestures with precise landmark detection.

๐Ÿ”น Gesture-Controlled Drone Simulation

  • Enables drone control using recognized hand gestures in a simulated environment.
  • Implemented using V-REP.

๐Ÿ”น Dataset Collection and Training Pipeline

  • Includes a diverse dataset of hand gestures for training.
  • Covers various angles, lighting conditions, and backgrounds.

๐Ÿ”น Edge Device Compatibility

  • Optimized for deployment on embedded devices like Raspberry Pi and Jetson Nano.
  • Supports ONNX model conversion for efficient inference.

๐Ÿ”น Multi-Modal AI Fusion

  • Future integration plans with voice and facial recognition.

Installation

Prerequisites

Ensure that you have the following dependencies installed:

pip install -r requirements.txt

You may also need additional dependencies depending on the platform.

Clone the Repository

git clone https://github.com/your-repo/Hand-Gesture-Recognition-Control.git
cd Hand-Gesture-Recognition-Control

Dataset

The dataset used for training consists of multiple hand gesture images labeled for different actions.

Hand Gesture Dataset Samples

Dataset Sample 1 Dataset Sample 2

Gesture Classes

Gesture Classes


Model Architecture

The system architecture consists of:

  • Convolutional Neural Networks (CNNs): Extracts spatial features from hand images.
  • LSTMs: Processes sequences of hand movements for gesture recognition.
  • MediaPipe: Provides real-time hand tracking and landmark detection.
  • ONNX Models: Optimized models for deployment on edge devices.

Hand Landmarks Detection

Hand Landmarks


Training Process

The training pipeline includes dataset preprocessing, augmentation, model training, and evaluation.

python keypoint_classification.ipynb
python point_history_classification.ipynb

Training Visualization

Training Graph


Inference and Real-Time Gesture Recognition

To run the inference:

python app.py

Real-Time Gesture Detection

Real-Time Gesture Detection

Hand Gesture Recognition in Action

Hand Gesture Recognition


Gesture-Controlled Drone Simulation

This project includes a simulated drone that can be controlled using hand gestures. The simulation is implemented in V-REP.

Simulation Video

Drone Simulation

Drone Simulation Framework

Drone Framework


Real-Time Hand Pose Tracking

A continuous hand tracking system allows gesture-based AI interaction.

Pose Tracking GIF

Pose Tracking GIF


Edge Deployment

This system supports deployment on edge devices such as Raspberry Pi or other embedded systems for real-time low-power inference.

Edge Deployment Image

Edge Deployment


Future Improvements

  • Multi-Modal Fusion: Integrate voice and gesture recognition for better interaction.
  • Optimized Deployment: Convert models using TensorRT for improved efficiency.
  • Customizable Gestures: Enable users to define and train their own gesture sets.

Conclusion

Hand Gesture Recognition & Control is a robust foundation for real-time gesture recognition and AI-controlled applications. With future enhancements, it can be applied in robotics, AR/VR, smart environments, and assistive technologies. Contributions and feedback are welcome!

For any questions or contributions, feel free to open an issue or a pull request!

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published