Hand Gesture Controlled Robotic Hand
Presently,
robotic hands are used in variety of applications, including military, defense
medical surgery, and pick and place functions in industrial automation. The
robotic hand can move and execute the role based on the gestures of human
hands, which can mimic human hand movements.
This
project presents data, strategies, and procedures which are fundamental for
building a mechanical hand constrained by the development of typical human hand
where the information is procured by utilizing the sensor fusion strategy of
various sensors such as TCRT5000. The hands are extremely adaptable and can be
used in a variety of environments.
As
the individual touches the IR sensor, the fingers move likewise which catches
and handles the movement and exerts accordingly. The toiling of this hand
depends on the Arduino stage; in which all sensors are interfaced with one
another by utilizing remote correspondence. While NRF24L01 the transceiver
module is used to transmit and receive the IR sensor data. Thus, this framework
gives a chance to expand versatility and opens up a new application territory
and diminishes the hole between the actual world and the advanced world with a
yield more natural.
The proposed work is achieved by utilizing the
Arduino microcontroller, IR sensor and RF sender and receiver (transceiver). Two main benefactions are presented in this work. Firstly, we’ve
shown that the Robotic hand can be controlled with hand-gestures according to
the movement and position of the fingers. Secondly, the proposed Robotic hand
system can be further extended to and infused with servo motors to move
vertically and pick and place the objects, which would be shown on the LCD
infused on the Robotic hand. This Robotic
hand system is introduced to pick and place objects, mimic the human hand and
perform tasks where human lives are at risk. The proposed system is designed at
lab-scale prototype to experimentally validate the efficiency, accuracy, and affordability
of the system. We remark that the proposed system can be implemented under real
conditions at large-scale in the future that will be useful in automobiles and
robotics applications.
Scope of the Project
There are high requirements to develop artificial hands for
many inhuman situations where human interactions are presenting challenges for
not possible situations. This proposal presents information, methods, and
techniques which are essential for building a robotic hand controlled by the
movements of normal human hand (Gesture Robotic Hand) whose data is acquiring by
using the sensor fusion technique of various sensors such as TCRT5000.
It is an integrated approach in real time detections,
gesture-based data which controls vehicle movement and manipulation on hand
movements of the user. As the person moves their hand, and touches the IR
sensor the fingers of the Robotic hand moves accordingly. The motion is
captured by IR sensor and processed. The development of this hand is based on
the Arduino platform; in which all are interfaced with each other by communication
such as using NRF24L01 which initiates serial communication. Hence this system
provides an opportunity to increase adaptability and opens up new application
areas and reduces the gap between the physical world and the digital world with
an output more intuitive.
As today human machine interaction are moving away
from mouse and pen and is becoming pervasive and much more compatible with the
physical world. With each passing day, the gap between machines and human is
being reduced with the introduction of new technology, having future scope of
advanced robotic hands that are designed like the human hand and can easily be
controlled using hand gestures only. For appropriate control mechanism and for
reducing the noise amount, which is coming in from the sensors, a proper
averaging algorithm will be used for smoothing the IR sensor output. It also
has proposed utility in the field of construction, medical science, and radioactive
labs etc. Furthermore, the hands if mounted on a Rover can be moved around
through gestures anywhere.
Problem
Statement
· The main problem in the current
robot control is that the operator has to use any manual control system to
direct the controller or the joystick to move the robotic hand.
· Also
the range is not considerable.
· Most importantly the operator
should be able to control the Robot naturally through gestures and view the
informed environment conveniently.
· Mobility aid for persons with disabilities.
Aims and Objectives
The
main goal of this project is to create a system that provides more natural and
convenient ways of expressing intentions and interactions with the environment
remotely, and perform repetitive task of pick and place which can be controlled
through gestures. In this context following are the main objectives:
· To design a wireless robotic platform.
· To develop intuitive gesture based control strategy for
operator.
To implement a pick and place approach, with the focus primarily on the task of grasping objects of different shapes and manipulating or assembling objects, assisting impaired people.
Figure 1 Wireless Gesture control
Inclination
when user moves his hand in any direction then the robotic arm can move
accordingly to the direction of the user hand. The whole system is divided into
two sections one is transmitter section and other is receiver section. This
system provides an opportunity to increase adaptability and opens up new
application areas and reduces the gap between the physical world and the
digital world with an output more intuitive.
The main hardware components used in this project are:
3.1 Glove
3.2 TCRT5000 (IR sensor)
3.3 Servo Motors MG90s
3.4 NRF24L01
3.5 Arduino Nano
3.6 Arduino Nano shield
3.7 LCD
3.8 LIPO Battery
3.9 Custom 3D Printed Hand
Figure 2 Cotton Knitted Glove
Custom 3D
Printed Hand
This complete
Robotic hand is designed as shown in figure 3 through
SOLIDWORKS software which provides free symbols & footprints for Altium,
Eagle, Ki cad, PADS & Other 20+ Design Tools. It is 3D CAD design software
published by Dassault Systemes that runs primarily on Microsoft Windows.
Whereas, Ultimaker Cura
is used for printing CAM designs, which is World’s most popular 3D printing
software.
Figure 3 3D Printed Parts
Hardware Design
Hardware
control is divided into two parts.
i. Robotic Platform or Base Control and receiver
side
ii. Gloves infused with sensors for transmission
side
Steps for Designing and Implementation
Step 1. User moves his hand.
Step 2. Sensors detect movement.
Step 3. MCU reads sensor data and transmits the data to Arduino through Serial communication.
Step 4. Arduino Sends data to Robot node through Radio wave.
Step 5. NRF24L01 (Transceiver) gets incoming data.
Step 6. Sends data to MCU of the Robot.
Step 7. MCU analyzes incoming data to see if the condition matches.
Step 8. If yes, triggers an even to actuate robot accordingly.
For code and further queries regarding the project contact me at:
Email: easymathsforyou3@gmail.com
video Link:
https://www.youtube.com/watch?v=6GhK0qwlzfU



Great job
ReplyDeletegreat project
ReplyDelete