Hand Gestures for Object Movements: A Comprehensive Exploration


Authors : G. Balakrishnan; T. Mangaiyarkarasi; K. Sangeetha; M. Rathika

Volume/Issue : Volume 10 - 2025, Issue 7 - July


Google Scholar : https://tinyurl.com/r7v3wak6

Scribd : https://tinyurl.com/4pw9tyw8

DOI : https://doi.org/10.38124/ijisrt/25jul535

Note : A published paper may take 4-5 working days from the publication date to appear in PlumX Metrics, Semantic Scholar, and ResearchGate.

Note : Google Scholar may take 30 to 40 days to display the article.


Abstract : Hand gestures are revolutionizing Human-Computer Interaction (HCI), moving beyond traditional interfaces to offer a natural and intuitive means of controlling objects across diverse applications, from virtual reality to robotics. The shift towards hand gestures represents a paradigm shift in HCI, driven by the demand for more immersive and accessible interactions. Hand gestures, as a powerful non-verbal communication modality, offer naturalness, accessibility, and hands- free control, crucial for enhancing user experience in fields like VR/AR and smart homes. Gestures can be broadly categorized into static gestures (fixed hand shapes for discrete commands like "stop" or "grab") and dynamic gestures (sequences of movements for continuous actions like "move" or "rotate"). Specific manipulations include translation, rotation, scaling, grabbing/releasing, and selection/deselection. Applications are vast, spanning immersive VR/AR environments, remote control of robotic arms, intuitive smart home device control, medical rehabilitation, and even in-car infotainment systems. However, challenges persist, including variations in hand shape, lighting conditions, occlusion, computational complexity, and user fatigue. Opportunities lie in advancements in sensor technology, deep learning, hybrid approaches, and the potential for gesture standardization. A typical hand gesture recognition system for object movement involves several key stages: data acquisition, preprocessing, hand detection/segmentation, feature extraction, gesture classification/recognition, and object control mapping. Data acquisition employs various sensing modalities. Vision-based approaches utilize RGB cameras (cost-effective but sensitive to lighting), depth cameras (providing 3D information, robust to lighting, but potentially more expensive), and infrared (IR) cameras (effective in low light). Sensor-based approaches rely on wearable devices like data gloves (highly accurate but intrusive), Inertial Measurement Units (IMUs) (less intrusive, capturing orientation and movement), and Electromyography (EMG) sensors (detecting muscle activity, but requiring direct skin contact and complex processing). Hybrid approaches combine these modalities to leverage their respective strengths, enhancing overall robustness and accuracy.

References :

  1. Aggarwal, J. K., & Cai, Q. (1999). Human motion analysis: A review. Computer Vision and Image Understanding, 73(3), 428-440.
  2. Mitra, S., & Acharya, T. (2007). Gesture recognition: A survey. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), 37(3), 365-381.
  3. Holz, D., Nieuwenhuisen, M., & Kobbelt, L. (2015). A Survey on Hand Pose Estimation and Hand Gesture Recognition. Computers & Graphics, 49, 162-171.
  4. Ren, Z., Meng, J., & Zhang, Z. (2013). Hand gesture recognition with depth data. IEEE Transactions on Multimedia, 15(7), 1575-1588.
  5. Wang, J., Wang, Y., & Chen, J. (2018). Real-time hand gesture recognition for virtual object manipulation using a single RGB camera. Journal of Visual Communication and Image Representation, 53, 217-227.
  6. Qian, C., Sun, X., Wei, Y., & Tang, X. (2019). PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. Advances in Neural Information Processing Systems, 31. (Relevant for depth-based point cloud processing)
  7. Zhang, H., Yu, Z., & Liu, Q. (2020). Hand gesture recognition based on 3D convolutional neural networks for human-robot interaction. Robotics and Autonomous Systems, 126, 103444.
  8. Oh, S. H., Park, J. H., Kim, K. H., & Park, J. O. (2012). Hand gesture recognition based on wearable sensors for robot control. Journal of Bionic Engineering, 9(3), 371-380.
  9. Srinivasan, S., Muneeswaran, A., & Palanisamy, S. (2018). IMU-based hand gesture recognition for human-robot interaction. Journal of Ambient Intelligence and Humanized Computing, 9(6), 1779-1788.
  10. Deep Learning in Hand Gesture Recognition for Object Control:
  11. Molchanov, P., et al. (2016). Online Deep Learning for Hand Gesture Recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 1-9.
  12. Liu, W., et al. (2020). EfficientPose: Scalable and Efficient 3D Hand Pose Estimation. Proceedings of the European Conference on Computer Vision (ECCV).
  13. Ge, L., Ren, Z., Yuan, Y., Xu, C., & Zhang, Z. (2018). 3D Hand Pose Estimation from a Single RGB Image with a Hierarchical Deep Network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 7818-7827.
  14. Lee, J., & Kim, H. (2017). Hand gesture interface for virtual object manipulation in an augmented reality environment. Journal of Supercomputing, 73(9), 4150-4166.
  15. Choi, J., & Kim, J. (2019). Real-time hand gesture recognition for drone control using deep learning. Sensors, 19(23), 5220.
  16. Wang, Y., Hu, K., & Lee, S. (2021). Hand Gesture Recognition for Human-Robot Collaboration in Industrial Settings. IEEE Robotics and Automation Letters, 6(2), 2419-2426.

Hand gestures are revolutionizing Human-Computer Interaction (HCI), moving beyond traditional interfaces to offer a natural and intuitive means of controlling objects across diverse applications, from virtual reality to robotics. The shift towards hand gestures represents a paradigm shift in HCI, driven by the demand for more immersive and accessible interactions. Hand gestures, as a powerful non-verbal communication modality, offer naturalness, accessibility, and hands- free control, crucial for enhancing user experience in fields like VR/AR and smart homes. Gestures can be broadly categorized into static gestures (fixed hand shapes for discrete commands like "stop" or "grab") and dynamic gestures (sequences of movements for continuous actions like "move" or "rotate"). Specific manipulations include translation, rotation, scaling, grabbing/releasing, and selection/deselection. Applications are vast, spanning immersive VR/AR environments, remote control of robotic arms, intuitive smart home device control, medical rehabilitation, and even in-car infotainment systems. However, challenges persist, including variations in hand shape, lighting conditions, occlusion, computational complexity, and user fatigue. Opportunities lie in advancements in sensor technology, deep learning, hybrid approaches, and the potential for gesture standardization. A typical hand gesture recognition system for object movement involves several key stages: data acquisition, preprocessing, hand detection/segmentation, feature extraction, gesture classification/recognition, and object control mapping. Data acquisition employs various sensing modalities. Vision-based approaches utilize RGB cameras (cost-effective but sensitive to lighting), depth cameras (providing 3D information, robust to lighting, but potentially more expensive), and infrared (IR) cameras (effective in low light). Sensor-based approaches rely on wearable devices like data gloves (highly accurate but intrusive), Inertial Measurement Units (IMUs) (less intrusive, capturing orientation and movement), and Electromyography (EMG) sensors (detecting muscle activity, but requiring direct skin contact and complex processing). Hybrid approaches combine these modalities to leverage their respective strengths, enhancing overall robustness and accuracy.

CALL FOR PAPERS


Paper Submission Last Date
31 - December - 2025

Video Explanation for Published paper

Never miss an update from Papermashup

Get notified about the latest tutorials and downloads.

Subscribe by Email

Get alerts directly into your inbox after each post and stay updated.
Subscribe
OR

Subscribe by RSS

Add our RSS to your feedreader to get regular updates from us.
Subscribe