Abstract
Motivation to engage in learning is essential for learning performance. Learners’ motivation is traditionally assessed using self-reported data, which is time-consuming, subjective, and interruptive to their learning process. To address this issue, this paper proposes a novel framework for multimodal assessment of learners’ motivation in e-learning environments with the ultimate purpose of supporting intelligent e-learning systems to facilitate dynamic, context-aware, and personalized services or interventions, thus sustaining learners’ motivation for learning engagement. We investigated the performance of the machine learning classifier and the most and least accurately predicted motivational factors. We also assessed the contribution of different electroencephalogram (EEG) and eye gaze features to motivation assessment. The applicability of the framework was evaluated in an empirical study in which we combined eye tracking and EEG sensors to produce a multimodal dataset. The dataset was then processed and used to develop a machine learning classifier for motivation assessment by predicting the levels of a range of motivational factors, which represented the multiple dimensions of motivation. We also proposed a novel approach to feature selection combining data-driven and knowledge-driven methods to train the machine learning classifier for motivation assessment, which has been proved effective in our empirical study at selecting predictors from a large number of extracted features from EEG and eye tracking data. Our study has revealed valuable insights for the role played by brain activities and eye movements on predicting the levels of different motivational factors. Initial results using logistic regression classifier have achieved significant predictive power for all the motivational factors studied, with accuracy of between 68.1% and 92.8%. The present work has demonstrated the applicability of the proposed framework for multimodal motivation assessment which will inspire future research towards motivationally intelligent e-learning systems.
Original language | English |
---|---|
Pages (from-to) | 64-81 |
Number of pages | 18 |
Journal | CCF Transactions on Pervasive Computing and Interaction |
Volume | 5 |
Issue number | 1 |
Early online date | 30 May 2022 |
DOIs | |
Publication status | Published (in print/issue) - 31 Mar 2023 |
Bibliographical note
Funding Information:This work was jointly funded by De Montfort University, UK and Norwegian Computing Centre, Norway. The authors would like to thank the colleagues from both organisations for their support in conducting the present research, in particular Dr Wolfgang Leister for his valuable comments.
Publisher Copyright:
© 2022, The Author(s).
Keywords
- E-learning
- EEG
- Eye tracking
- Multimodal motivation modelling