TY - GEN
T1 - Robotic Arm Handling Based on Real-Time Recognition of the Number of Raised Fingers Using Convolutional Neural Networks
AU - Hidalgo-Abril, John
AU - Benítez, Diego S.
AU - Pérez-Pérez, Noel
AU - Camacho, Oscar
AU - Grijalva, Felipe
AU - Herrera, Marco
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - This paper presents a system for operating a robotic arm based on the number of raised fingers detected in a human hand using a camera mounted on the robot. Leveraging advancements in physical human-robot interaction (pHRI), the system utilizes a convolutional neural network (CNN) to interpret hand gestures for intuitive control. Initially, the system uses the MediaPipe Framework to identify 21 landmarks of the hand, which are then used to define the bounding box of the hand. A convolutional neural network (CNN) processes this bounded hand image to determine the number of raised fingers. Implemented using a Jetson Nano, a Logitech Brio 4K webcam, and Python libraries such as OpenCV, I2C tools, and TensorFlow, the model was trained on 30,000 images, achieving a 92.7% accuracy during training and 94% during real-time testing. A voting strategy ensures robust predictions by considering the most frequent result from ten consecutive predictions, mitigating the impact of minor hand movements. The system demonstrates the potential for advanced applications in hand gesture-based robot manipulation and interaction.
AB - This paper presents a system for operating a robotic arm based on the number of raised fingers detected in a human hand using a camera mounted on the robot. Leveraging advancements in physical human-robot interaction (pHRI), the system utilizes a convolutional neural network (CNN) to interpret hand gestures for intuitive control. Initially, the system uses the MediaPipe Framework to identify 21 landmarks of the hand, which are then used to define the bounding box of the hand. A convolutional neural network (CNN) processes this bounded hand image to determine the number of raised fingers. Implemented using a Jetson Nano, a Logitech Brio 4K webcam, and Python libraries such as OpenCV, I2C tools, and TensorFlow, the model was trained on 30,000 images, achieving a 92.7% accuracy during training and 94% during real-time testing. A voting strategy ensures robust predictions by considering the most frequent result from ten consecutive predictions, mitigating the impact of minor hand movements. The system demonstrates the potential for advanced applications in hand gesture-based robot manipulation and interaction.
KW - convolutional neural networks
KW - hand recognition
KW - robotic arm con-trol
UR - http://www.scopus.com/inward/record.url?scp=85211081851&partnerID=8YFLogxK
U2 - 10.1109/ARGENCON62399.2024.10735915
DO - 10.1109/ARGENCON62399.2024.10735915
M3 - Contribución a la conferencia
AN - SCOPUS:85211081851
T3 - 2024 7th IEEE Biennial Congress of Argentina, ARGENCON 2024
BT - 2024 7th IEEE Biennial Congress of Argentina, ARGENCON 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 7th IEEE Biennial Congress of Argentina, ARGENCON 2024
Y2 - 18 September 2024 through 20 September 2024
ER -