Authors :
G. Balakrishnan; T. Mangaiyarkarasi; K. Sangeetha; M. Rathika
Volume/Issue :
Volume 10 - 2025, Issue 7 - July
Google Scholar :
https://tinyurl.com/r7v3wak6
Scribd :
https://tinyurl.com/4pw9tyw8
DOI :
https://doi.org/10.38124/ijisrt/25jul535
Note : A published paper may take 4-5 working days from the publication date to appear in PlumX Metrics, Semantic Scholar, and ResearchGate.
Note : Google Scholar may take 30 to 40 days to display the article.
Abstract :
Hand gestures are revolutionizing Human-Computer Interaction (HCI), moving beyond traditional interfaces to
offer a natural and intuitive means of controlling objects across diverse applications, from virtual reality to robotics. The
shift towards hand gestures represents a paradigm shift in HCI, driven by the demand for more immersive and accessible
interactions. Hand gestures, as a powerful non-verbal communication modality, offer naturalness, accessibility, and hands-
free control, crucial for enhancing user experience in fields like VR/AR and smart homes. Gestures can be broadly
categorized into static gestures (fixed hand shapes for discrete commands like "stop" or "grab") and dynamic gestures
(sequences of movements for continuous actions like "move" or "rotate"). Specific manipulations include translation,
rotation, scaling, grabbing/releasing, and selection/deselection. Applications are vast, spanning immersive VR/AR
environments, remote control of robotic arms, intuitive smart home device control, medical rehabilitation, and even in-car
infotainment systems. However, challenges persist, including variations in hand shape, lighting conditions, occlusion,
computational complexity, and user fatigue. Opportunities lie in advancements in sensor technology, deep learning, hybrid
approaches, and the potential for gesture standardization. A typical hand gesture recognition system for object movement
involves several key stages: data acquisition, preprocessing, hand detection/segmentation, feature extraction, gesture
classification/recognition, and object control mapping. Data acquisition employs various sensing modalities. Vision-based
approaches utilize RGB cameras (cost-effective but sensitive to lighting), depth cameras (providing 3D information, robust
to lighting, but potentially more expensive), and infrared (IR) cameras (effective in low light). Sensor-based approaches rely
on wearable devices like data gloves (highly accurate but intrusive), Inertial Measurement Units (IMUs) (less intrusive,
capturing orientation and movement), and Electromyography (EMG) sensors (detecting muscle activity, but requiring direct
skin contact and complex processing). Hybrid approaches combine these modalities to leverage their respective strengths,
enhancing overall robustness and accuracy.
References :
- Aggarwal, J. K., & Cai, Q. (1999). Human motion analysis: A review. Computer Vision and Image Understanding, 73(3), 428-440.
- Mitra, S., & Acharya, T. (2007). Gesture recognition: A survey. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), 37(3), 365-381.
- Holz, D., Nieuwenhuisen, M., & Kobbelt, L. (2015). A Survey on Hand Pose Estimation and Hand Gesture Recognition. Computers & Graphics, 49, 162-171.
- Ren, Z., Meng, J., & Zhang, Z. (2013). Hand gesture recognition with depth data. IEEE Transactions on Multimedia, 15(7), 1575-1588.
- Wang, J., Wang, Y., & Chen, J. (2018). Real-time hand gesture recognition for virtual object manipulation using a single RGB camera. Journal of Visual Communication and Image Representation, 53, 217-227.
- Qian, C., Sun, X., Wei, Y., & Tang, X. (2019). PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. Advances in Neural Information Processing Systems, 31. (Relevant for depth-based point cloud processing)
- Zhang, H., Yu, Z., & Liu, Q. (2020). Hand gesture recognition based on 3D convolutional neural networks for human-robot interaction. Robotics and Autonomous Systems, 126, 103444.
- Oh, S. H., Park, J. H., Kim, K. H., & Park, J. O. (2012). Hand gesture recognition based on wearable sensors for robot control. Journal of Bionic Engineering, 9(3), 371-380.
- Srinivasan, S., Muneeswaran, A., & Palanisamy, S. (2018). IMU-based hand gesture recognition for human-robot interaction. Journal of Ambient Intelligence and Humanized Computing, 9(6), 1779-1788.
- Deep Learning in Hand Gesture Recognition for Object Control:
- Molchanov, P., et al. (2016). Online Deep Learning for Hand Gesture Recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 1-9.
- Liu, W., et al. (2020). EfficientPose: Scalable and Efficient 3D Hand Pose Estimation. Proceedings of the European Conference on Computer Vision (ECCV).
- Ge, L., Ren, Z., Yuan, Y., Xu, C., & Zhang, Z. (2018). 3D Hand Pose Estimation from a Single RGB Image with a Hierarchical Deep Network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 7818-7827.
- Lee, J., & Kim, H. (2017). Hand gesture interface for virtual object manipulation in an augmented reality environment. Journal of Supercomputing, 73(9), 4150-4166.
- Choi, J., & Kim, J. (2019). Real-time hand gesture recognition for drone control using deep learning. Sensors, 19(23), 5220.
- Wang, Y., Hu, K., & Lee, S. (2021). Hand Gesture Recognition for Human-Robot Collaboration in Industrial Settings. IEEE Robotics and Automation Letters, 6(2), 2419-2426.
Hand gestures are revolutionizing Human-Computer Interaction (HCI), moving beyond traditional interfaces to
offer a natural and intuitive means of controlling objects across diverse applications, from virtual reality to robotics. The
shift towards hand gestures represents a paradigm shift in HCI, driven by the demand for more immersive and accessible
interactions. Hand gestures, as a powerful non-verbal communication modality, offer naturalness, accessibility, and hands-
free control, crucial for enhancing user experience in fields like VR/AR and smart homes. Gestures can be broadly
categorized into static gestures (fixed hand shapes for discrete commands like "stop" or "grab") and dynamic gestures
(sequences of movements for continuous actions like "move" or "rotate"). Specific manipulations include translation,
rotation, scaling, grabbing/releasing, and selection/deselection. Applications are vast, spanning immersive VR/AR
environments, remote control of robotic arms, intuitive smart home device control, medical rehabilitation, and even in-car
infotainment systems. However, challenges persist, including variations in hand shape, lighting conditions, occlusion,
computational complexity, and user fatigue. Opportunities lie in advancements in sensor technology, deep learning, hybrid
approaches, and the potential for gesture standardization. A typical hand gesture recognition system for object movement
involves several key stages: data acquisition, preprocessing, hand detection/segmentation, feature extraction, gesture
classification/recognition, and object control mapping. Data acquisition employs various sensing modalities. Vision-based
approaches utilize RGB cameras (cost-effective but sensitive to lighting), depth cameras (providing 3D information, robust
to lighting, but potentially more expensive), and infrared (IR) cameras (effective in low light). Sensor-based approaches rely
on wearable devices like data gloves (highly accurate but intrusive), Inertial Measurement Units (IMUs) (less intrusive,
capturing orientation and movement), and Electromyography (EMG) sensors (detecting muscle activity, but requiring direct
skin contact and complex processing). Hybrid approaches combine these modalities to leverage their respective strengths,
enhancing overall robustness and accuracy.