Authors :
Sathya.C; Dhatchana.P
Volume/Issue :
Volume 10 - 2025, Issue 1 - January
Google Scholar :
https://tinyurl.com/ybvwpf44
Scribd :
https://tinyurl.com/2y2entex
DOI :
https://doi.org/10.5281/zenodo.14613833
Abstract :
Understanding and classifying student actions
within educational environments is a vital component of
boosting learning results and well-being. This study
presents a novel method to student activity categorisation
by employing facial expression detection technologies.
The technology is intended to record and evaluate pupils'
facial expressions, understand their emotional states, and
then classify their actions. This study investigates the
application of deep learning models for face emotion
identification using a dataset that includes both academic
and non-academic activities. The system can recognise
emotions such as happiness, sorrow, rage, and surprise.
The extracted emotion traits are then used to characterise
student actions, revealing whether a student is engaged,
attentive, puzzled, or indifferent, among other states. This
strategy has the potential to improve educational settings
by offering real-time insights into student conduct and
allowing for timely adjustments to improve learning
experiences and outcomes. It also offers up possibilities
for personalised educational support and the creation of
intelligent learning systems. In this research, we will
construct a system to extract face characteristics using the
Grassmann method. And identify the emotions of
students at certain times. Predict the active state using
emotion categorisation and provide reports to the
administrator. Furthermore, this technique shows
potential for the creation of adaptive learning systems
that react to students' emotional states, delivering extra
help or challenges as needed. For example, a virtual tutor
may modify the difficulty of exercises based on a student's
emotional reactions, producing a dynamic and responsive
learning experience.
References :
- Gautam, Chahak, and K. R. Seeja. "Facial emotion recognition using Handcrafted features and CNN." Procedia Computer Science 218 (2023): 1295-1303.
- Dirik, Mahmut. "Optimized anfis model with hybrid metaheuristic algorithms for facial emotion recognition." International Journal of Fuzzy Systems 25.2 (2023): 485-496.
- Punuri, Sudheer Babu, et al. "Efficient net-XGBoost: an implementation for facial emotion recognition using transfer learning." Mathematics 11.3 (2023): 776.
- Mehendale, Ninad. "Facial emotion recognition using convolutional neural networks (FERC)." SN Applied Sciences 2.3 (2020): 446.
- Chaudhari, Aayushi, et al. "Facial emotion recognition with inter-modality-attention-transformer-based self-supervised learning." Electronics 12.2 (2023): 288.
- Schoneveld, Liam, Alice Othmani, and Hazem Abdelkawy. "Leveraging recent advances in deep learning for audio-visual emotion recognition." Pattern Recognition Letters 146 (2021): 1-7.
- Sumathy, P., and Ahilan Chandrasekaran. "An Optimized Image Pre-Processing Technique for Face Emotion Recognition System." Annals of the Romanian Society for Cell Biology 25.6 (2021): 6247-6261.
- Kim, Jung Hwan, Alwin Poulose, and Dong Seog Han. "The extensive usage of the facial image threshing machine for facial emotion recognition performance." Sensors 21.6 (2021): 2026.
- Akhand, M. A. H., et al. "Facial emotion recognition using transfer learning in the deep CNN." Electronics 10.9 (2021): 1036.
- Sinha, Avigyan, and R. P. Aneesh. "Real time facial emotion recognition using deep learning." International Journal of Innovations and Implementations in Engineering 1 (2019).
- Zhong, Yuanchang, et al. "HOG-ESRs Face Emotion Recognition Algorithm Based on HOG Feature and ESRs Method." Symmetry 13.2 (2021): 228.
- Chang, Jia-Wei, et al. "Music recommender using deep embedding-based features and behavior-based reinforcement learning." Multimedia Tools and Applications 80.26 (2021): 34037-34064.
- Athavle, Madhuri,” Music Recommendation System Using Facial Expression Recognition Using Machine Learning, International Journal for Research in Applied Science & Engineering Technology (IJRASET), 2022
- Chowdary, M. Kalpana, Tu N. Nguyen, and D. Jude Hemanth. "Deep learning-based facial emotion recognition for human–computer interaction applications." Neural Computing and Applications (2021): 1-18.
- Ch, Satyanarayana. "An efficient facial emotion recognition system using novel deep learning neural network-regression activation classifier." Multimedia Tools and Applications 80.12 (2021): 17543-17568.
- Mehendale, Ninad. "Facial emotion recognition using convolutional neural networks (FERC)." SN Applied Sciences 2.3 (2020): 1-8.
- Ramírez, Jaime, and M. Julia Flores. "Machine learning for music genre: multifaceted review and experimentation with audioset." Journal of Intelligent Information Systems 55.3 (2020): 469-499.
- Liu, Jun, Yanjun Feng, and Hongxia Wang. "Facial expression recognition using pose-guided face alignment and discriminative features based on deep learning." IEEE Access 9 (2021): 69267-69277.
- Said, Yahia, and Mohammad Barr. "Human emotion recognition based on facial expressions via deep learning on high-resolution images." Multimedia Tools and Applications 80.16 (2021): 25241-25253.
- Ruiz-Garcia, Ariel, et al. "Deep learning for emotion recognition in faces." International Conference on Artificial Neural Networks. Springer, Cham, 2016.
Understanding and classifying student actions
within educational environments is a vital component of
boosting learning results and well-being. This study
presents a novel method to student activity categorisation
by employing facial expression detection technologies.
The technology is intended to record and evaluate pupils'
facial expressions, understand their emotional states, and
then classify their actions. This study investigates the
application of deep learning models for face emotion
identification using a dataset that includes both academic
and non-academic activities. The system can recognise
emotions such as happiness, sorrow, rage, and surprise.
The extracted emotion traits are then used to characterise
student actions, revealing whether a student is engaged,
attentive, puzzled, or indifferent, among other states. This
strategy has the potential to improve educational settings
by offering real-time insights into student conduct and
allowing for timely adjustments to improve learning
experiences and outcomes. It also offers up possibilities
for personalised educational support and the creation of
intelligent learning systems. In this research, we will
construct a system to extract face characteristics using the
Grassmann method. And identify the emotions of
students at certain times. Predict the active state using
emotion categorisation and provide reports to the
administrator. Furthermore, this technique shows
potential for the creation of adaptive learning systems
that react to students' emotional states, delivering extra
help or challenges as needed. For example, a virtual tutor
may modify the difficulty of exercises based on a student's
emotional reactions, producing a dynamic and responsive
learning experience.