An Intelligent Multimodal Emotion Recognition System for E-Learning
Main Article Content
Abstract
The purpose of this research paper is to introduce an Intelligent Multimodal Emotion Recognition System (IMERS) that aims to improve the e-learning process by accurately perceiving and reacting to the emotional states of learners. IMERS incorporates information through three primary modalities: facial expressions, voice, and text. The utilization of multimodal fusion in this technique effectively addresses the constraints of single-modality systems and yields a more extensive and precise comprehension of emotions. The paper highlight the following: Initially, we discuss the structure of multimodal fusion, specifically focusing on the architecture of IMERS. This includes an explanation of the many components involved, such as data preparation, feature extraction, decision fusion, and sentiment classification. Every approach employs distinct deep learning algorithms customized to its unique properties. Furthermore, our assessment of IMERS encompasses its proficiency in discerning emotions inside e-learning environments, as evidenced by its correct detection of primary emotions across diverse datasets. Another area of emphasis is personalized learning applications, in which we demonstrate how IMERS customize learning experiences by adapting instruction, offering specific feedback, and cultivating an emotionally nurturing learning environment..