Predicting Cognitive Load Using Sensor Data in a Literacy Game
Minghao Cai
EdTeKLA, Computing Science University of AlbertaEdmonton, Canada
minghaocai@ualberta.ca
Carrie Demmans Epp
EdTeKLA, Computing Science University of AlbertaEdmonton, Canada
cdemmansepp@ualberta.ca

ABSTRACT

Educational games are being increasingly used to support self-paced learning. However, educators and system designers often face challenges in monitoring student affect and cognitive load. Existing assessments in game-based learning environments (GBLEs) tend to focus more on outcomes rather than processes, potentially overlooking key aspects of the learning journey that include learner affect and cognitive load. To address this issue, we collected data and trained a model to track learner cognitive load while they used an online literacy game for English. We collected affect-related physiological data and pupil data during gameplay to enable the development of models that identify these latent characteristics of learner processes. Our model indicates the feasibility of using these data to track cognitive load in GBLEs. Our multimodal model distinguished different levels of cognitive load, achieving the highest Kappa (\(.417\)) and accuracy (\(70\%\)). Our model reveals the importance of including affect-related features (i.e., EDA and heart rate) when predicting cognitive load and extends recent findings suggesting the benefit of using multiple channels when modeling latent aspects of learner processes. Findings also suggest that cognitive load tracking could now be used to facilitate the creation of personalized learning experiences.

Keywords

Cognitive load, Student modeling, Game-based learning environments, Eye tracking, Physiological signals

1. INTRODUCTION

Prior studies have found that educational games can facilitate self-paced learning [12730]. Despite the recognized benefits of game-based learning environments (GBLEs), research examining how learners react when interacting with these environments is still limited [23]. Current evaluations of GBLEs often prioritize learning outcomes over the processes that lead to these outcomes [2130]. This approach may overlook elements of the learning experience, which limits our understanding of how learners react to these environments. Cognitive load is a critical aspect of learning experience that has been shown to have a complicated relationship with learner performance [19]. According to Cognitive Load Theory (CLT), cognitive load refers to the demands placed on working memory by learning tasks and activities [282019].

A key challenge in capturing the full spectrum of learner cognitive load in GBLEs is the reliance on traditional measures such as self-report questionnaires. These methods only provide retrospective assessments after learning or sporadic snapshots of learner experience [15]. In addition, they depend on learners’ ability to self-assess and express their internal states, which can be challenging [163]. These limitations underscore the need for continuous methods of tracking cognitive load so we can gain a deeper understanding of self-paced processes within GBLEs.

Some work has attempted to use brain-activity data to track variations in cognitive load, with electroencephalography (EEG) emerging as a prominent tool [2]. However, EEG relies on obtrusive equipment; the EEG cap that must be worn might distract users from the task, potentially causing unwanted cognitive load.

More recently, pupil diameter has emerged as an indicator of cognitive load [179]. This work built on foundational research that linked pupil dilation to increased problem difficulty [12], which is related to intrinsic load. Although using pupil diameter has been effective [179], it is sensitive to eye movements and several extraneous factors (e.g., ambient lighting conditions and the eye-tracker’s camera angle) [4].

Up until this point, the use of sensors for estimating cognitive load has relied on a single channel of data, like pupil diameter. Given that learning activities are often complex and involve the coordination of cognitive processes and affective states, relying solely on information from a single channel limits our understanding of how and when students learn best. Moving beyond existing models that primarily focus on unimodal data, our study explored a multimodal approach to enhance predictive performance.

Prior studies suggest that affect and cognitive load appear to interact during learning [2211146]. Yet, there has not been a consensus on how to effectively integrate affect into cognitive-load modeling. Following from this, we explored the effectiveness of incorporating physiological signals indicative of affective responses (i.e., electrodermal activity [EDA] and heart rate) with pupil diameter to predict cognitive load. This multi-modal model of cognitive load demonstrates the feasibility of using non-invasive sensor data to track cognitive load in GBLEs without interrupting the learner, thus enabling a dynamic evaluation of the learning process. Moreover, investigation of what the model learned revealed that features related to affect support the evaluation of cognitive load.

2. METHODS

We conducted a case study, where participants interacted with an English literacy game. This study was approved by our institutional research ethics board. Each participant consented to the study and was compensated with a $30 gift card. All participants interacted with the GBLE for roughly 60 minutes. None had previously used the GBLE.

Data were collected from 35 English-language learners (ELLs), but we only used data from 34 due to considerable sensor noise in one participant’s data. ELLs were aged 17 through 33 years (\(M = 24.1\)). The languages participants reported speaking at home included Mandarin (\(n = 24\)), Japanese (\(n = 3\)), and Persian (Farsi; \(n = 3\)). Gujarati, Korean, Spanish, and French were each spoken at home by one ELL.

2.1 Game-based Learning Environment

The employed GBLE, Dreamscape, was designed to provide adaptive learning experiences [6]. It blends reading-comprehension activities into a base-building game. Players engage in a fantasy world where they are tasked with building and safeguarding their virtual homeland against invading reveries (creatures akin to soldiers in traditional base-building games). In this system, learner progress in the game is directly linked to their learning-activity performance. Learners are required to correctly answer questions to facilitate their progress in the game. Incorrect answers do not result in penalties.

2.2 Data Collection

We collected both sensor and self-report data. The self-report data were used as labels for model training and evaluation.

The sensing system used two non-invasive wearable sensors. Pupil Core, a glasses-shaped eye tracker, recorded dual eye movements at 200 Hz and provided gaze tracking and pupil dynamics. A wireless wristband (Empatica E4) was used to provide continuous tracking of EDA data at 4Hz and heart rate using 10-second spans.

The self-report instrument captured participants’ subjective experiences of cognitive load. Following from cognitive load theory [25], a 10-item Likert scale questionnaire (1 - Strongly Disagree and 10 - Strongly Agree) was developed by adapting existing scales from other educational domains where Cronbach’s \(\alpha \) was greater than \(.80\). This instrument assesses the three sub-types of load (intrinsic with 3 items, extraneous with 2 items, and germane load with 2 items) in addition to cognitive load (1 item). We averaged the sub-type scores and then combined them with the cognitive load item score to compute the final cognitive load score. The final scores were then discretized into three levels: low, moderate, and high.

Self-report data were collected after a participant completed a task in the system (e.g., finished a game-play activity, submitted an answer) to ensure minimal interference with learning processes. The instrument was administered approximately every 5-minutes.

3. MODEL DEVELOPMENT

3.1 Data Cleaning and Preprocessing

Sensor data is noisy and requires pre-processing to enable its use. We report data cleaning by sensor type.

For each participant, the pupil signal includes pupil diameters along with time stamps and an estimated confidence level for each recorded diameter. The confidence ranges from 0.0 (could not be detected) to 1.0 (very high certainty). Because blinks can introduce substantial noise when measuring pupil size, we first removed data from 250 milliseconds before the onset and 250 milliseconds after the moment of a blink, as detected by the device. To further reduce the risk of incorporating erroneous or flawed data caused by the device itself, we implemented a confidence filtering step. Data points with a confidence level below \(.65\) were excluded. For data continuity, missing data due to removal were linearly interpolated. To further reduce noise, we applied a \(3^{rd}\)-order Butterworth filter with a 4 Hz cutoff to minimize high-frequency disturbances while retaining signal fidelity.

Given the sensitivity of wristband sensors to various types of noise, we applied a simple moving average (SMA) filter after exporting participants’ EDA and heart rate signal records.

After data cleaning, we aligned the sensor data with the timestamped participant self-report data. Data recorded from when participants were completing the self-report form were excluded. This step resulted in a dataset that contained 312 segments.

3.2 Feature Extraction

Following data cleaning and preparation, our research proceeded to feature extraction from the segmented physiological data.

Recognizing the lack of established evidence for the choice of time window for cognitive load prediction during literacy activities, we experimented with different window durations for extracting physiological features. This empirical approach allowed us to ascertain an appropriate window for cognitive load prediction. We tested the predictive performance of physiological data from 6 windows (30s, 60s, 90s, 120s, 150s, and 180s) and selected the window with the highest Kappa.

We extracted 7 features from the pupil diameter data for each segment: the average size of the pupil diameter (AvgPD); the maximum pupil size in the dataset (MaxPD); the minimum pupil size in the dataset (MinPD); the average speed at which the pupil diameter changes (AvgPV); the maximum speed at which the pupil diameter changes (MaxPV); the largest continuous change in pupil diameter without direction change (MaxPC); and the frequency of changes in pupil diameter (PCF).

We extracted 9 features from the EDA data for each segment: the average value of the EDA (AvgE); the standard deviation of the EDA (SDGE); the maximum value of the EDA (MaxE); the minimum value of the EDA (MinE); the difference between the maximum and minimum EDA (RngE); the average speed of change in EDA levels (AvgEV); the maximum speed at which EDA changes (MaxEV); the maximum continuous change in EDA without direction change (MaxEC); and the frequency of changes in EDA (ECF).

We extracted 9 features from the heart rate data for each segment: the average value of the heart rate (AvgH); the standard deviation of the heart rate (SDH); the maximum value of the heart rate (MaxH); the minimum value of the heart rate (MinH); the difference between the maximum and minimum heart rate (RngH); the the average speed of change in heart rate (AvgHV); the maximum speed at which the heart rate changes (MaxHV); the maximum continuous change in heart rate without direction change (MaxHC); and the frequency of changes in heart rate (HCF).

3.3 Classification Models

To achieve our research goal, we implemented and examined six conventional machine learning classifiers: Naive Bayes (NB), Decision Tree (DT), Support vector machine with linear kernel (Linear SVM), Support vector machine with radial basis function kernel (RBF SVM), logistic regression (LR), random forest (RF).

In addition to the above classifier types, two approaches to modelling were compared: a baseline unimodal approach and our multimodal approach. The unimodal baseline used Index of Pupillary Activity (IPA) [9], specifically measuring fluctuations in pupil diameter over time to indicate cognitive load. The multimodal approach augmented the baseline model by including the EDA and heart-rate features as inputs. This combination of classifiers and input features resulted in the training and evaluation of 12 models.

3.4 Model Training

We used Scikit-Learn (1.3.2) to develop our models. Model training was conducted on a desktop PC (i7-9700 CPU, 64GB RAM).

The data was split, with 80% allocated to training and 20% to testing.

We performed hyperparameter tuning on these models by applying 4-fold cross-validation to the training data. We used grid search to identify the optimal hyperparameters for each combination of classifier type and input features. Kappa served as the evaluation metric during the hyperparameter tuning process. The hyperparameter search space can be found in Appendix A. We report the selected hyperparameters below.

For NB, we used alpha = 1 with Laplace smoothing. DT was configured using ‘gini’. The Linear SVM and RBF SVM had a regularization parameter of 1.0 (i.e., C = 1.0) and used the ‘linear’ and ‘rbf’ kernels, respectively. For LR, the ‘lbfgs’ solver was used with an ‘l2’ penalty and C = 1.0. Finally, RF used 100 estimators.

In addition to tuning the standard hyperparameters reported above, we tuned the time window for sensor signals. The best performing time windows are reported in the results section (see Table 1).

3.5 Model Evaluation and Interpretation

We used mlxtend (0.23.1) to conduct statistical tests to examine model performance.

We report Cohen’s \(\kappa \) [7] alongside accuracy metrics due to the skewed distribution of class labels within our dataset. Cochran’s \(Q\) test was used to assess performance differences among models, and McNemar’s test was used to identify pairwise differences.

To explore what factors contributed to model predictions, we analyzed the Gini importance [5] of features in our best-performing multimodal model.

4. RESULTS

4.1 Model Performance

Table 1 reports the peak performance metrics for both the multimodal model and the unimodal model.

Cochran’s Q test of multimodal model performance indicated a difference in classifier performance when using all three feature sets (\(Q = 15.357\), \(p = .009\), \(\eta ^2_{Q} = .061\)). The subsequent McNemar’s tests (Table 2) identified differences between RF and NB as well as RBF SVM and NB. These statistical testing results and the descriptive statistics of multimodal model performance indicate that random forest best supported the prediction of learner cognitive-load levels when pupil, heart rate, and EDA information were used as inputs.

In contrast to the multimodal models, no differences were found across the unimodal models (\(Q = 8.391,\) \(p = .136\), \(\eta ^2_{Q} = .034\)).

Table 1: Model Performance and window size
Classifier
\(\kappa \) Accuracy Window (s)
Multimodal models
Naive Bayes .121 .417 150
Decision Tree .256 .583 180
Linear SVM .210 .517 60
RBF SVM .319 .633 210
Logistic Regression .242 .550 90
Random Forest .417 .700 210
Unimodal models
Naive Bayes .169 .550 180
Decision Tree .102 .517 180
Linear SVM .178 .600 210
RBF SVM .234 .683 90
Logistic Regression .159 .517 90
Random Forest .102 .517 180
Bold font indicates the best result

Table 2: McNemar’s test results comparing multimodal classifiers
DT
Linear SVM
RBF SVM
LR
RF
NB \(\chi ^2 = 3.125\), p \(= .077\), \(\eta ^2 = .063\) \(\chi ^2 = 1.636\), p \(= .201\), \(\eta ^2 = .033\) \(\boldsymbol {\chi }^2\)= 6.760, p = .009, \(\boldsymbol {\eta }^2 = \textbf { .135}\) \(\chi ^2 = 2.909\), p \(= .088\), \(\eta ^2 = .058\) \(\boldsymbol \chi ^2 \) = 8.257, p = .004, \(\boldsymbol {\eta }^2\) = .165
DT \(\chi ^2 = 0.615\), p \(= .433\), \(\eta ^2 = .012\) \(\chi ^2 = 0.333\), p \(= .564\), \(\eta ^2 = .007\) \(\chi ^2 = 0.167\), p \(= .683\), \(\eta ^2 = .003\) \(\chi ^2 = 2.579\), p \(= .108\), \(\eta ^2 = .052\)
Linear SVM \(\chi ^2 = 2.882\), p \(= .090\), \(\eta ^2 = .058\) \(\chi ^2 = 0.333\), p \(= .564\), \(\eta ^2 = .007\) \(\chi ^2 = 5.261\), p \(= .022\), \(\eta ^2 = .105\)
RBF SVM \(\chi ^2 = 1.316\), p \(= .251\), \(\eta ^2 = .026\) \(\chi ^2 = 1.600\), p \(= .206\), \(\eta ^2 = .032\)
LR \(\chi ^2 = 3.522\), p \(= .061\), \(\eta ^2 = .070\)
Bold font indicates significant differences

Table 3: McNemar’s test results comparing multimodal to unimodal models
NB
DT
Linear SVM
RBF SVM
LR
RF
\(\chi ^2\) 2.000 0.533 1.000 0.474 0.118 4.481
p .157 .465 .317 .491 .732 .034
\(\eta ^2\) .040 .011 .020 .009 .002 .090
Bold font indicates significant differences

To determine whether the multimodal model outperformed the unimodal one, we compared model performance across inputs for each algorithm. The multimodal version of RF outperformed the unimodal one (Table 3), showing the benefit of including complementary sensor signals.

4.2 Feature Importance

We analyzed the multimodal random forest model because it had the highest \(\kappa \). Our results showed all feature categories contributed to model predictions. We found that both the EDA and heart rate group of features played an important role in cognitive load prediction. We reviewed the categories and specific features in decreasing order and summarized the three most important features from each signal.

4.2.1 Pupil Features

Overall, the pupil feature group provided the most information to the model (36.61%). The most important feature category was frequency of changes in pupil diameter (PCF), followed by the average speed at which the pupil diameter changed (AvgPV) and the average size of the pupil diameter (AvgPD).

4.2.2 Heart Rate Features

The heart rate feature group provided a similar amount of information to the model (36.24%). The most important feature category was the maximum speed at which the heart rate changed (MaxHV), followed by the minimum (MinH) and average (AvgH) heart-rate values.

4.2.3 EDA Features

The EDA feature group provided less information to the model (27.15%). In this group, the most important feature category was the minimum value (MinE), followed by the frequency of changes (ECF) and the average value (AvgE).

5. DISCUSSION

Our research is driven by the primary goal of building a model to track learners’ cognitive-load levels in a GBLE for literacy. Our results demonstrate the feasibility of creating predictive models to classify levels of cognitive load in this context. Notably, models employing RBF SVM and RF emerged as top performers (Table 1). These models achieved Cohen’s \(\kappa \) scores exceeding .319 and .417 and demonstrated prediction accuracies ranging from 63% to 70%. Moreover, our multimodal strategy generally yielded higher Cohen’s kappa scores across classifiers, with the RF model showing particularly superior performance.

Participants’ physiological responses were recorded as they naturally occured without intentionally inducing affective or cognitive states. This approach provides a more authentic representation of how learners react to and process the game-based learning environment. This approach contrasts with studies where conditions are manipulated to elicit specific learner states [26], which, while valuable for controlled experiments, may not fully capture how learners behave under normal conditions.

Our multimodal model’s superior performance over that of the baseline unimodal classifier’s, highlights the added value of incorporating diverse physiological channels.

5.1 Feature Contributions to Prediction

Our results showed the feasibility of developing a predictive model and provided insight into the underlying factors influencing cognitive load.

While all features played a role in the prediction process, pupil-based metrics had the highest importance in our prediction model with the frequency, speed of changes in pupil diameter, and average pupil size being the most critical features. These findings align with existing research that indicated pupil dilation and constriction are controlled by the autonomic nervous system and can be influenced by cognitive effort [9].

Beyond the frequently used pupil features, our analysis highlights the contributions (63.76% in total) of EDA and heart rate features. Until now, these features have only been considered indicators of affective states [13]. Our findings suggest these sensor streams may also be indicative of cognitive states. This finding aligns with recent research highlighting the interplay between learner affect and cognitive load. Studies have suggested that emotions play a crucial role in cognitive processes, with different emotional states potentially enhancing or hindering learners’ ability to comprehend and retain information [1018]. More recently, Cai et al. conducted a study demonstrating that affect valence can predict cognitive load in game-based learning environments [6]. Our modeling study underscores the importance of considering both cognitive and affective factors when designing and evaluating educational content, suggesting that an integrated approach to monitoring and responding to both affective states and cognitive load could enhance student modeling.

Among the heart rate features, rapid heart rate changes as well as the heart rate’s mean and trough value helped predict learner cognitive load levels. These findings highlight the significance of incorporating heart rate, particularly its fluctuations, as an indicator of cognitive load. The identification of heart rate as an effective indicator advances our ability to detect changes in load, particularly with the widespread availability of wearable sensor technology such as smartwatches.

Similarly, the identification of EDA as another indicator advances our ability to detect cognitive load with EDA’s minimum and average values as well as fluctuations in EDA supporting this goal. This finding highlights the nuanced ways in which physiological responses correlate with cognitive demand. The present work expands the association between increases in EDA and emotional arousal [24] to cognitive load. The prominence of minimum EDA as an important feature suggests that moments of lower emotional arousal may offer insights into cognitive load states, possibly reflecting periods of reduced cognitive effort or enhanced understanding. The inclusion of EDA in student modeling goes beyond conventional cognitive assessments to support the diagnosis of cognitive load and learner affect that might then be used to adapt the system to enhance engagement and learning [8].

5.2 Towards Adaptation based on Cognitive Load

The best-performing model in our study supported the prediction of learner cognitive load in a literacy game context. These results illustrate the utility of physiological data for predicting cognitive load that could then be used to dynamically capture this aspect of learner experience so that we can better understand learner behaviors in ways that are not achievable through self-report measures.

Our model tracks cognitive load using real-time physiological responses that can provide continuous monitoring of learner cognitive state without disrupting the flow of learning. This capability should enable the dynamic adaptation of support and instructional strategies to enhance the effectiveness of GBLEs and improve learning experience [2029].

Our work provides a feasible way to track learner cognitive load that can then be used to inform adaptation. These models enable the identification of moments when cognitive load becomes overtaxed, which could result from overly challenging tasks or poor system feedback. This is the point at which automated adaptation would be implemented. Adaptations could include adjusted learning activities, feedback that encourages learners to take a break, or suggested alternative learning strategies. The identification of lower cognitive demand periods could be used to inform adaptation in a manner that enhances subsequent engagement and information retention [19].

6. CONCLUSIONS

Despite the growing adoption of GBLEs, the exploration of how learner cognitive load fluctuates within these complex interactive environments is still restricted. This limitation appears to result from a lack of methods for continuously monitoring such aspects of learner experience.

In this study, we created a predictive model that uses sensor data to represent learner physiological responses as input to a model that estimates cognitive load. Our results demonstrate that a combination of data on pupil diameter, EDA, and heart rate offers a reasonable estimation for three cognitive load levels, highlighting the added value of using mulitmodal data. The sensor data used as input to the multimodal classifiers highlights the importance of affect-related features. These results also indicate that both cognitive and affective factors need to be accounted for in the design and assessment of educational materials. We suggest that an approach combining the monitoring of affective states and cognitive load could improve student modeling.

7. ACKNOWLEDGMENTS

This work was supported in part by funding from the Social Sciences and Humanities Research Council of Canada and the Natural Sciences and Engineering Research Council of Canada (NSERC), [RGPIN-2018-03834].

8. REFERENCES

  1. E. O. Acquah and H. T. Katz. Digital game-based L2 learning outcomes for primary through high-school students: A systematic literature review. Computers & Education, 143:103667, Jan. 2020.
  2. P. Antonenko, F. Paas, R. Grabner, and T. van Gog. Using Electroencephalography to Measure Cognitive Load. Educational Psychology Review, 22(4):425–438, Dec. 2010.
  3. P. Ayres. Using subjective measures to detect variations of intrinsic cognitive load within problems. Learning and Instruction, 16(5):389–400, Oct. 2006.
  4. J. Beatty and B. Lucero-Wagoner. The pupillary system. In Handbook of psychophysiology, 2nd ed, pages 142–162. Cambridge University Press, New York, NY, US, 2000.
  5. L. Breiman, J. Friedman, R. Olshen, and C. J. Stone. Classification and Regression Trees. Routledge, New York, Oct. 2017.
  6. M. Cai, G. Rebolledo Mendez, G. Arevalo, S. S. Tang, Y. A Abdullah, and C. Demmans Epp. Toward Supporting Adaptation: Exploring Affect’s Role in Cognitive Load when Using a Literacy Game. In Proceedings of the CHI Conference on Human Factors in Computing Systems (CHI ’24), Honolulu, HI, USA, May 2024.
  7. J. Cohen. A Coefficient of Agreement for Nominal Scales. Educational and Psychological Measurement, 20(1):37–46, Apr. 1960. Publisher: SAGE Publications Inc.
  8. S. D’Mello. A Selective Meta-Analysis on the Relative Incidence of Discrete Affective States during Learning with Technology. Journal of Educational Psychology, 105(4):1082–1099, Nov. 2013. Publisher: American Psychological Association ERIC Number: EJ1054438.
  9. A. T. Duchowski, K. Krejtz, I. Krejtz, C. Biele, A. Niedzielska, P. Kiefer, M. Raubal, and I. Giannopoulos. The Index of Pupillary Activity: Measuring Cognitive Load vis-à-vis Task Difficulty with Pupil Oscillation. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, CHI ’18, pages 1–13, New York, NY, USA, Apr. 2018. Association for Computing Machinery.
  10. G. Egidi and R. J. Gerrig. How valence affects language processing: Negativity bias and mood congruence in narrative comprehension. Memory & Cognition, 37(5):547–555, July 2009.
  11. K. L. Fraser, P. Ayres, and J. Sweller. Cognitive Load Theory for the Design of Medical Simulations. Simulation in Healthcare, 10(5):295–307, Oct. 2015.
  12. E. H. Hess and J. M. Polt. Pupil Size in Relation to Mental Activity during Simple Problem-Solving. Science (New York, N.Y.), 143(3611):1190–1192, Mar. 1964.
  13. M. S. Hussain, O. AlZoubi, R. A. Calvo, and S. K. D’Mello. Affect Detection from Multichannel Physiology during Learning Sessions with AutoTutor. In G. Biswas, S. Bull, J. Kay, and A. Mitrovic, editors, Artificial Intelligence in Education, Lecture Notes in Computer Science, pages 131–138, Berlin, Heidelberg, 2011. Springer.
  14. S. Kalyuga and A.-M. Singh. Rethinking the Boundaries of Cognitive Load Theory in Complex Learning. Educational Psychology Review, 28(4):831–852, Dec. 2016.
  15. M. Krell, K. M. Xu, G. D. Rey, and F. Paas. Editorial: Recent Approaches for Assessing Cognitive Load From a Validity Perspective. Frontiers in Education, 6, 2022.
  16. J. Leppink, F. Paas, C. P. M. Van der Vleuten, T. Van Gog, and J. J. G. Van Merriënboer. Development of an instrument for measuring different types of cognitive load. Behavior Research Methods, 45(4):1058–1072, Dec. 2013.
  17. S. Marshall. The Index of Cognitive Activity: measuring cognitive workload. In Proceedings of the IEEE 7th Conference on Human Factors and Power Plants, pages 7–7, Sept. 2002.
  18. C. Mills, J. Wu, and S. D’Mello. Being Sad Is Not Always Bad: The Influence of Affect on Expository Text Comprehension. Discourse Processes, 56(2):99–116, Feb. 2019.
  19. F. Paas and P. Ayres. Cognitive Load Theory: A Broader View on the Role of Memory in Learning and Education. Educational Psychology Review, 26(2):191–195, June 2014.
  20. F. Paas, J. E. Tuovinen, H. Tabbers, and P. W. M. Van Gerven. Cognitive Load Measurement as a Means to Advance Cognitive Load Theory. Educational Psychologist, 38(1):63–71, Mar. 2003.
  21. G. Petri and C. Gresse von Wangenheim. How games for computing education are evaluated? A systematic literature review. Computers & Education, 107:68–90, Apr. 2017.
  22. J. L. Plass and U. Kaplan. Emotional Design in Digital Media for Learning. In S. Y. Tettegah and M. Gartmeier, editors, Emotions, Technology, Design, and Learning, Emotions and Technology, pages 131–161. Academic Press, San Diego, Jan. 2016.
  23. J. L. Plass, R. E. Mayer, and B. D. Homer. Handbook of Game-Based Learning. MIT Press, Feb. 2020. Google-Books-ID: _2fKDwAAQBAJ.
  24. H. F. Posada-Quintero and K. H. Chon. Innovations in Electrodermal Activity Data Collection and Signal Processing: A Systematic Review. Sensors, 20(2):479, Jan. 2020. Number: 2 Publisher: Multidisciplinary Digital Publishing Institute.
  25. G. Salomon. Television is "easy" and print is "tough": The differential investment of mental effort in learning as a function of perceptions and attributions. Journal of Educational Psychology, 76(4):647–658, 1984.
  26. S. Scrimin and L. Mason. Does mood influence text processing and comprehension? Evidence from an eye-movement study. British Journal of Educational Psychology, 85(3):387–406, 2015.
  27. K. D. Stiller and S. Schworm. Game-Based Learning of the Structure and Functioning of Body Cells in a Foreign Language: Effects on Motivation, Cognitive Load, and Performance. Frontiers in Education, 4, 2019.
  28. J. Sweller. Cognitive Load Theory. In J. P. Mestre and B. H. Ross, editors, Psychology of Learning and Motivation, volume 55, pages 37–76. Academic Press, Jan. 2011.
  29. J. J. G. van Merriënboer and J. Sweller. Cognitive Load Theory and Complex Learning: Recent Developments and Future Directions. Educational Psychology Review, 17(2):147–177, June 2005.
  30. Z. Xu, Z. Chen, L. Eutsler, Z. Geng, and A. Kogut. A scoping review of digital game-based technology on English language learning. Educational Technology Research and Development, 68(3):877–904, June 2020.

APPENDIX

A. HYPERPARAMETER SEARCH SPACE

For the decision tree, we tried tried ‘gini’ and ‘entropy’.

For Linear SVM, RBF SVM, and Logistic Regression (LR), we tuned C. The following values were explored: 0.1, 1, 10, and 100.

We tuned the number of trees in the random forest, exploring values in the range of 20 through 100 using a step size of 20.