Evaluation of Students’ Learning Engagement in Online Classes Based on Multimodal Vision Perspective
<p>Flow chart of the hierarchical framework for online classroom student engagement.</p> "> Figure 2
<p>Model of the hierarchical framework for online classroom student engagement. (<b>A</b>) The gaze estimation model. (<b>B</b>) The proposed facial expression recognition model. (<b>C</b>) The proposed action recognition model.</p> "> Figure 3
<p>Two detailed submodule of Inflated ResNet. (<b>a</b>) Detailed structure of the Res1. submodule. (<b>b</b>) Detailed structure of the Res2. submodule.</p> "> Figure 4
<p>Partial online classroom student behavior dataset.</p> "> Figure 5
<p>Distribution of the number of samples at four engagement levels in the dataset.</p> "> Figure 6
<p>Online classroom student gaze situation.</p> "> Figure 7
<p>Gaze estimation Euler angle prediction visualization. (<b>a</b>) Watching the screen. (<b>b</b>) Not looking at the screen.</p> "> Figure 8
<p>Visualization of facial recognition classification results. (<b>a</b>) Surprised. (<b>b</b>) Confused. (<b>c</b>) Happy. (<b>d</b>) Neutral. (<b>e</b>) Tired. (<b>f</b>) Boredom.</p> "> Figure 9
<p>Visualization of action recognition classification results. (<b>a</b>) Writing. (<b>b</b>) Reading. (<b>c</b>) Eating. (<b>d</b>) Looking around. (<b>e</b>) Sleeping. (<b>f</b>) Playing with mobile phone.</p> "> Figure 10
<p>Visualization of action recognition classification results.</p> "> Figure 11
<p>Confusion matrix for the classification of engagement levels at various grades.</p> ">
Abstract
:1. Introduction
2. Related Work
3. Proposed Method
3.1. Hierarchical Analysis Framework for Online Classroom Student Engagement
3.2. The Gaze Estimation Model
3.3. The Proposed Facial Expression Recognition Model
3.4. The Proposed Action Recognition Model
4. Design of Experiments
4.1. Datasets
4.2. Analysis of Student Engagement Based on Gaze Estimation
4.3. Analysis of Student Engagement Based on Facial Expression Recognition
4.4. Analysis of Student Engagement Based on Action Recognition
4.5. Online Classroom Student Engagement Rating
5. Experiments and Results Analysis
5.1. Additional Datasets
5.2. Experiment Details
5.3. Discussion of the Experimental Results
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Chen, X.; Wang, D. On the development process and main characteristic of MOOC. Mod. Educ. Technol. 2013, 23, 5–10. [Google Scholar]
- Reich, J.; Ruipérez-Valiente, J.A. The MOOC Pivot. Science 2019, 363, 130–131. [Google Scholar] [CrossRef] [PubMed]
- Jordan, K. MOOC Completion Rates: The Data. Available online: http://www.katyjordan.com/MOOCproject.html (accessed on 5 August 2023).
- Jiang, Z.; Zhang, Y.; Li, X. Learning behavior analysis and prediction based on MOOC data. J. Comput. Res. Dev. 2015, 52, 614–628. [Google Scholar]
- Hughes, M.; Salamonson, Y.; Metcalfe, L. Student Engagement Using Multiple-Attempt ‘Weekly Participation Task’ Quizzes with Undergraduate Nursing Students. Nurse Educ. Pract. 2020, 46, 102803. [Google Scholar] [CrossRef] [PubMed]
- Brozina, C.; Knight, D.B.; Kinoshita, T.; Johri, A. Engaged to Succeed: Understanding First-Year Engineering Students’ Course Engagement and Performance Through Analytics. IEEE Access 2019, 7, 163686–163699. [Google Scholar] [CrossRef]
- Wang, W.; Guo, L.; He, L.; Wu, Y.J. Effects of Social-Interactive Engagement on the Dropout Ratio in Online Learning: Insights from MOOC. Behav. Inf. Technol. 2019, 38, 621–636. [Google Scholar] [CrossRef]
- Oh, C.; Roumani, Y.; Nwankpa, J.K.; Hu, H.-F. Beyond Likes and Tweets: Consumer Engagement Behavior and Movie Box Office in Social Media. Inf. Manag. 2017, 54, 25–37. [Google Scholar] [CrossRef]
- Sun, Y.; Ni, L.; Zhao, Y.; Shen, X.; Wang, N. Understanding Students’ Engagement in MOOCs: An Integration of Self-determination Theory and Theory of Relationship Quality. Br. J. Educ. Technol. 2019, 50, 3156–3174. [Google Scholar] [CrossRef]
- Fredricks, J.A.; Blumenfeld, P.C.; Paris, A.H. School Engagement: Potential of the Concept, State of the Evidence. Rev. Educ. Res. 2004, 74, 59–109. [Google Scholar] [CrossRef]
- Wu, F.; Zhang, Q. Learning behavioral engagement: Definition, analysis framework and theoretical model. China Educ. Technol. 2018, 372, 35–41. [Google Scholar]
- Zhang, F.; Zhang, T.; Mao, Q.; Xu, C. Geometry Guided Pose-Invariant Facial Expression Recognition. IEEE Trans. Image Process. 2020, 29, 4445–4460. [Google Scholar] [CrossRef]
- Zhang, H. The Literature Review of Action Recognition in Traffic Context. J. Vis. Commun. Image Represent. 2019, 58, 63–66. [Google Scholar] [CrossRef]
- Larson, R.W.; Richards, M.H. Boredom in the Middle School Years: Blaming Schools versus Blaming Students. Am. J. Educ. 1991, 99, 418–443. [Google Scholar] [CrossRef]
- Shernoff, D.J.; Csikszentmihalyi, M.; Schneider, B.; Shernoff, E.S. Student Engagement in High School Classrooms from the Perspective of Flow Theory. Sch. Psychol. Q. 2003, 18, 158. [Google Scholar] [CrossRef]
- Whitehill, J.; Serpell, Z.; Lin, Y.-C.; Foster, A.; Movellan, J.R. The Faces of Engagement: Automatic Recognition of Student Engagementfrom Facial Expressions. IEEE Trans. Affect. Comput. 2014, 5, 86–98. [Google Scholar] [CrossRef]
- Grafsgaard, J.F.; Wiggins, J.B.; Boyer, K.E.; Wiebe, E.N.; Lester, J.C. Automatically Recognizing Facial Expression: Predicting Engagement and Frustration. In Proceedings of the 6th International Conference on Educational Data Mining (EDM 2013), Memphis, TN, USA, 6–9 July 2013. [Google Scholar]
- Monkaresi, H.; Bosch, N.; Calvo, R.A.; D’Mello, S.K. Automated Detection of Engagement Using Video-Based Estimation of Facial Expressions and Heart Rate. IEEE Trans. Affect. Comput. 2017, 8, 15–28. [Google Scholar] [CrossRef]
- Zhang, Z.; Li, Z.; Liu, H.; Cao, T.; Liu, S. Data-Driven Online Learning Engagement Detection via Facial Expression and Mouse Behavior Recognition Technology. J. Educ. Comput. Res. 2020, 58, 63–86. [Google Scholar] [CrossRef]
- Zhan, Z.H. An emotional and cognitive recognition model for distance learners based on intelligent agent-the coupling of eye tracking and expression recognition techniques. Mod. Dist. Educ. Res. 2013, 5, 100–105. [Google Scholar]
- Alkabbany, I.; Ali, A.M.; Foreman, C.; Tretter, T.; Hindy, N.; Farag, A. An Experimental Platform for Real-Time Students Engagement Measurements from Video in STEM Classrooms. Sensors 2023, 23, 1614. [Google Scholar] [CrossRef]
- Zhang, Y.H.; Pan, M.; Zhong, G.C.; Cao, X.M. Learning Engagement Detection Based on Face Dataset in the Mixed Scene. Mod. Educ. Technol. 2021, 31, 84–92. [Google Scholar]
- Liu, H.; Nie, H.; Zhang, Z.; Li, Y.-F. Anisotropic Angle Distribution Learning for Head Pose Estimation and Attention Understanding in Human-Computer Interaction. Neurocomputing 2021, 433, 310–322. [Google Scholar] [CrossRef]
- Chen, P.; Huangpu, D.P.; Luo, Z.Y.; Li, D.X. Visualization analysis of learning attention based on single-image PnP head posture estimation. J. Commun. 2018, 39, 141–150. [Google Scholar]
- Singh, T.; Mohadikar, M.; Gite, S.; Patil, S.; Pradhan, B.; Alamri, A. Attention Span Prediction Using Head-Pose Estimation with Deep Neural Networks. IEEE Access 2021, 9, 142632–142643. [Google Scholar] [CrossRef]
- Zhou, J.; Ye, J.M.; Li, C. Multimodal Learning Affective Computing: Motivations, Frameworks, and Recommendations. e-Educ. Res. 2021, 42, 26–32+46. [Google Scholar] [CrossRef]
- Xu, X.; Zhao, W.; Liu, H. Research on application and model of emotional analysis in blended learning environment: From perspective of meta-analysis. e-Educ. Res. 2018, 39, 70–77. [Google Scholar] [CrossRef]
- Loderer, K.; Pekrun, R.; Lester, J.C. Beyond Cold Technology: A Systematic Review and Meta-Analysis on Emotions in Technology-Based Learning Environments. Learn. Instr. 2020, 70, 101162. [Google Scholar] [CrossRef]
- Zhao, C.; Shu, H.; Gu, X. The Measurement and Analysis of Students’ Classroom Learning Behavior Engagement Based on Computer. Mod. Educ. Technol. 2021, 31, 96–103. [Google Scholar]
- Alkabbany, I.; Ali, A.; Farag, A.; Bennett, I.; Ghanoum, M.; Farag, A. Measuring student engagement level using facial information. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 3337–3341. [Google Scholar] [CrossRef]
- Langton, S.R.H.; Watt, R.J.; Bruce, V. Cues to the Direction of Social Attention. Trends Cogn. Sci. 2000, 4, 50–59. [Google Scholar] [CrossRef]
- Abdelrahman, A.A.; Hempel, T.; Khalifa, A.; Al-Hamadi, A. L2CS-Net: Fine-grained gaze estimation in unconstrained environments. arXiv 2022, arXiv:2203.03339. [Google Scholar]
- Ding, X.; Zhang, X.; Ma, N.; Han, J.; Ding, G.; Sun, J. RepVGG: Making VGG-Style ConvNets Great Again. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; IEEE: Nashville, TN, USA, 2021; pp. 13728–13737. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Las Vegas, NV, USA, 2016; pp. 770–778. [Google Scholar]
- Gupta, A.; D’Cunha, A.; Awasthi, K.; Balasubramanian, V. DAiSEE: Towards User Engagement Recognition in the Wild. arXiv 2022, arXiv:1609.01885. [Google Scholar]
- D’Mello, S.K.; Craig, S.D.; Graesser, A.C. Multimethod Assessment of Affective Experience and Expression during Deep Learning. Int. J. Learn. Technol. 2009, 4, 165. [Google Scholar] [CrossRef]
- Chen, Z.; Shi, B.E. Appearance-based gaze estimation using dilated-convolutions. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 309–324. [Google Scholar]
- Cheng, Y.; Zhang, X.; Lu, F.; Sato, Y. Gaze estimation by exploring two-eye asymmetry. IEEE Trans. Image Process. 2020, 29, 5259–5272. [Google Scholar] [CrossRef] [PubMed]
- Cheng, Y.; Huang, S.; Wang, F.; Qian, C.; Lu, F. A coarse-to-fine adaptive network for appearance-based gaze estimation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10623–10630. [Google Scholar]
- Wang, G.; Li, J.; Wu, Z.; Xu, J.; Shen, J.; Yang, W. EfficientFace: An Efficient Deep Network with Feature Enhancement for Accurate Face Detection. Multimed. Syst. 2023, 29, 2825–2839. [Google Scholar] [CrossRef]
- Zhang, Y.; Wang, C.; Ling, X.; Deng, W. Learn From All: Erasing Attention Consistency for Noisy Label Facial Expression Recognition Supplementary Material. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer Nature: Cham, Switzerland, 2022; pp. 418–434. [Google Scholar]
- Savchenko, A.V. HSEmotion: High-Speed Emotion Recognition Library. Softw. Impacts 2022, 14, 100433. [Google Scholar] [CrossRef]
- Donahue, J.; Hendricks, L.A.; Guadarrama, S.; Rohrbach, M.; Venugopalan, S.; Saenko, K.; Darrell, T. Long-Term Recurrent Convolutional Networks for Visual Recognition and Description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2625–2634. [Google Scholar]
- Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning Spatiotemporal Features with 3D Convolutional Networks. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; IEEE: Santiago, Chile, 2015; pp. 4489–4497. [Google Scholar]
- Simonyan, K.; Zisserman, A. Two-Stream Convolutional Networks for Action Recognition in Videos. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2014; Volume 27. [Google Scholar]
Attitude towards Learning | Expression |
---|---|
Positive | Neutral |
Surprised | |
Confused | |
Happy | |
Negative | Tired |
Boredom |
Attitude towards Learning | Expression |
---|---|
Positive | Writing |
Reading | |
Negative | Eating |
Looking around | |
Sleeping | |
Playing with mobile phone |
Methods | MPIIGaze |
---|---|
Dilated-Net [37] | 4.8° |
FAR-Net [38] | 4.3° |
CA-Net [39] | 4.1° |
) [32] | 3.96° |
) [32] | 3.92° |
Methods | Accuracy |
---|---|
EfficientFace [40] | 82.2 |
EAC [41] | 80.8 |
HSEmotion [42] | 84.7 |
Ours | 88.4 |
Methods | Accuracy |
---|---|
LRCN [43] | 82.7 |
C3D [44] | 85.2 |
Two Stream [45] | 88.0 |
Ours | 89.5 |
Algorithm | Accuracy |
---|---|
LRCN | 76.8 |
C3D | 74.9 |
Two Stream | 80.4 |
Ours | 83.7 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Qi, Y.; Zhuang, L.; Chen, H.; Han, X.; Liang, A. Evaluation of Students’ Learning Engagement in Online Classes Based on Multimodal Vision Perspective. Electronics 2024, 13, 149. https://doi.org/10.3390/electronics13010149
Qi Y, Zhuang L, Chen H, Han X, Liang A. Evaluation of Students’ Learning Engagement in Online Classes Based on Multimodal Vision Perspective. Electronics. 2024; 13(1):149. https://doi.org/10.3390/electronics13010149
Chicago/Turabian StyleQi, Yongfeng, Liqiang Zhuang, Huili Chen, Xiang Han, and Anye Liang. 2024. "Evaluation of Students’ Learning Engagement in Online Classes Based on Multimodal Vision Perspective" Electronics 13, no. 1: 149. https://doi.org/10.3390/electronics13010149
APA StyleQi, Y., Zhuang, L., Chen, H., Han, X., & Liang, A. (2024). Evaluation of Students’ Learning Engagement in Online Classes Based on Multimodal Vision Perspective. Electronics, 13(1), 149. https://doi.org/10.3390/electronics13010149