Classroom Student Emotions Classification from Facial Expressions and Speech Signals using Deep Learning
Archana Sharma1, Vibhakar Mansotra2

1Dr. Archana Sharma, Department of Computer Science, Government M.A.M College, Cluster University of Jammu, Jammu, India.
2Dr. Vibhakar Mansotra, Department of Computer Science and IT, University of Jammu, Jammu, India.

Manuscript received on 04 August 2019. | Revised Manuscript received on 08 August 2019. | Manuscript published on 30 September 2019. | PP: 6675-6683 | Volume-8 Issue-3 September 2019 | Retrieval Number: C5666098319/2019©BEIESP | DOI: 10.35940/ijrte.C5666.098319
Open Access | Ethics and Policies | Cite | Mendeley | Indexing and Abstracting
© The Authors. Blue Eyes Intelligence Engineering and Sciences Publication (BEIESP). This is an open access article under the CC-BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/)

Abstract: Classroom environment is a competent platform for the students to learn and improve their understanding of the subject. An instructor’s primary responsibility lies in managing the students in a way they feel interested and focused during the class. With the aid of automated systems based on artificial intelligence, an instructor can get feedback on the students’ attention span in the class by monitoring their emotions using learning algorithms that can prove to be effective to improve the teaching style of the instructor that can in turn have positive effects on the class. In this paper, we propose an LSTM recurrent neural network trained on an emotional corpus database to extract the speech features and convolutional neural networks trained on the FER2013 facial emotion recognition database were used to predict the speech and facial emotions of the students respectively, in real-time. The live video and audio sequence of the students captured is fed to the learned model to classify the emotions individually. Once the emotions such as anger, sadness, happiness, surprise, fear, disgust and neutral were identified, a decision-making mechanism was used to analyze the predicted emotions and choose the overall group emotion by virtue of the highest peak value achieved. This research approach has the potential to be deployed in video conferences, online classes etc. This implementation proposal should effectively improve the classification accuracy and the relatability of the detected student emotions and facilitate in the design of sophisticated automated learning systems that can be a valuable tool in evaluating both the students and the instructors. The adapted research methodologies and their results are discussed and found to perform suggestively better than the other research works used in the comparison.
Keywords: Deep Learning, Emotion Recognition, Convolutional Neural Network, Face Recognition, Speech Emotion, Recurrent Neural Network.

Scope of the Article: Deep Learning