Multimodal motivation modelling and computing towards motivationally intelligent E-learning systems

Ruijie Wang, Luke Chen, Aladdin Ayesh

Research output: Contribution to journalArticlepeer-review

Abstract

Motivation to engage in learning is essential for learning performance. Learners’ motivation is traditionally assessed using self-reported data, which is time-consuming, subjective, and interruptive to their learning process. To address this issue, this paper proposes a novel framework for multimodal assessment of learners’ motivation in e-learning environments with the ultimate purpose of supporting intelligent e-learning systems to facilitate dynamic, context-aware, and personalized services or interventions, thus sustaining learners’ motivation for learning engagement. We investigated the performance of the machine learning classifier and the most and least accurately predicted motivational factors. We also assessed the contribution of different electroencephalogram (EEG) and eye gaze features to motivation assessment. The applicability of the framework was evaluated in an empirical study in which we combined eye tracking and EEG sensors to produce a multimodal dataset. The dataset was then processed and used to develop a machine learning classifier for motivation assessment by predicting the levels of a range of motivational factors, which represented the multiple dimensions of motivation. We also proposed a novel approach to feature selection combining data-driven and knowledge-driven methods to train the machine learning classifier for motivation assessment, which has been proved effective in our empirical study at selecting predictors from a large number of extracted features from EEG and eye tracking data. Our study has revealed valuable insights for the role played by brain activities and eye movements on predicting the levels of different motivational factors. Initial results using logistic regression classifier have achieved significant predictive power for all the motivational factors studied, with accuracy of between 68.1% and 92.8%. The present work has demonstrated the applicability of the proposed framework for multimodal motivation assessment which will inspire future research towards motivationally intelligent e-learning systems.
Original languageEnglish
JournalCCF Transactions on Pervasive Computing and Interaction
Early online date30 May 2022
DOIs
Publication statusE-pub ahead of print - 30 May 2022

Keywords

  • E-learning
  • EEG
  • Eye tracking
  • Multimodal motivation modelling

Fingerprint

Dive into the research topics of 'Multimodal motivation modelling and computing towards motivationally intelligent E-learning systems'. Together they form a unique fingerprint.

Cite this