A Conceptual Model for End-to-End Causal Discovery in Knowledge Tracing
Nischal Ashok Kumar, Wanyong Feng, Jaewook Lee, Hunter McNichols, Aritra Ghosh, Andrew Lan
University of Massachusetts Amherst
{nashokkumar,wanyongfeng,jaewooklee,wmcnichols,arighosh,andrewlan}@umass.edu

ABSTRACT

In this paper, we take a preliminary step towards solving the problem of causal discovery in knowledge tracing, i.e., finding the underlying causal relationship among different skills from real-world student response data. This problem is important since it can potentially help us understand the causal relationship between different skills without extensive A/B testing, which can potentially help educators to design better curricula according to skill prerequisite information. Specifically, we propose a conceptual solution, a novel causal gated recurrent unit (GRU) module in a modified deep knowledge tracing model, which uses i) a learnable permutation matrix for causal ordering among skills and ii) an optionally learnable lower-triangular matrix for causal structure among skills. We also detail how to learn the model parameters in an end-to-end, differentiable way. Our solution placed among the top entries in Task 3 of the NeurIPS 2022 Challenge on Causal Insights for Learning Paths in Education. We detail preliminary experiments as evaluated on the challenge’s public leaderboard since the ground truth causal structure has not been publicly released, making detailed local evaluation impossible.

Keywords

Causal Discovery, Knowledge Tracing, Response Data

1. INTRODUCTION

Knowledge Tracing (KT) [1] refers to the problem of estimating a student’s understanding or mastery of certain skills, concepts, or knowledge components through their responses to questions and using these estimates to predict future performance. KT methods are frequently utilized in modern online education platforms to determine the knowledge levels of many students to enable the platform to provide personalized feedback and recommendations, ultimately leading to better learning results [19]. KT methods are limited in how they represent the relationship between skills; One key limitation is that most do not model the causal relationships between skills. Most KT methods simply treat human expert-provided skill tags as a flat structure (with a few exceptions, such as [27], that organize skills hierarchically as trees). As a result, these models are not capable of providing meaningful pedagogical insights, i.e., predicting future student performance if a particular instructional plan is applied instead of the actual plan applied.

Causal analysis tools are a perfect fit to address these limitations in KT. The task of causal discovery, i.e., learning causal relationships among different skills from observational data, is especially important. First, it helps educators learn prerequisite relationships among skills. This can guide educators in ordering topics within their curriculum, and can guide students to review prerequisite information when they are stuck on a question [2]. Second, causal relationships among skills helps us with the task of causal inference, i.e., estimating the effect of a particular pedagogical treatment or intervention. Traditionally, these tasks are addressed through randomized controlled trials which are difficult to scale. Therefore, incorporating causal discovery into KT methods has the potential to become a scalable alternative since it can be done solely from observational student response data. Performing causal discovery directly from observational student response data is challenging since it is not straightforward to estimate treatment effects from observational data with incomplete or no knowledge of the causal relationship between skills. This problem is referred to as the end-to-end causal inference problem, where we discover the causal graph and estimate treatment effects together.

1.1 Contributions

In this paper, we take a preliminary step towards learning causal ordering among skills from student response data. This task is proposed in the NeurIPS 2022 Challenge on Causal Insights for Learning Paths in Education1. Our proposed conceptual solution is, to the best of our knowledge, the first KT method to learn the causal structure among human expert-provided skill tags directly from observational data in an end-to-end manner. Specifically, our contributions in this paper are as follows:

2. RELATED WORK

2.1 KT methods

Existing KT methods can be classified along several different axes, the first of which is how they represent the student knowledge representation variable \(h\). Classic Bayesian KT methods, such as those in [91528], treat student knowledge as a latent binary variable. Recent methods like deep learning-based KT methods, such as [414172229], treat student knowledge as hidden states in neural networks. This setup results in models that excel at predicting future performance but have limited interpretability [3]. Another major axis is how KT methods represent responses, questions, skills, and time steps. To represent student responses, most existing KT methods treat them as binary-valued indicating response correctness. However, a few methods, such as option tracing [5] and predict partial analysis [26], have characterized student responses as non-binary-valued by analyzing the specific options selected on multiple-choice questions. Another exception is [11], which uses large language models to predict open-ended student responses in a generative way. To represent questions and skills, most existing KT methods one-hot encode them based on question IDs or skill tags [24], except [1112]. To represent time steps, most existing KT methods treat each question as a discrete time step, with a few exceptions such as [25], which considers the exact, continuous time elapsed between responses.

2.2 Causal Analysis Methods

In the field of education, there exist very few works on causality and especially few in the context of KT. [8] is closely related to our work, where the authors study the relationship between courses in higher educational institutions using historical student performance data. They use matching methods and regression to determine the average treatment effect (ATE). Along similar lines, [20] and [21] developed theory and methods for analyzing A/B testing data and presented studied data collected from real-world randomized controlled trials. These works focus on causal inference, i.e., assuming that the structure is given and the focus is on estimating the treatment effect. However, the data we use from Eedi contains fine-grained skills, defined as the smallest elements of learning among primary/middle school students. Therefore, our work is different from these works in terms of both the goal and the educational context. The only existing works that study causal discovery in the context of KT are [10] and [13]. The former uses a special model structure that has some similarity to ours to model knowledge state transitions among uninterpretable latent skills. The authors showed that their method, while simple, is highly accurate in predicting unobserved student responses, but do not evaluate on whether the identified causal structure is valid. Our proposed method to learn latent causal structure is closely based on the structural equation model (SEM) [16]. SEM enables us to estimate the relationships between observed and latent variables, offering valuable insights into their underlying relationships. The hypothesized causal relationship among variables is represented as a directed acylic graph (DAG). In this work, our goal is to learn the causal structure graph \(\mathcal {G}\).

3. METHODOLOGY

We now detail our conceptual causal KT method.

3.1 Basic Setup

The basic KT model contains two components: \begin {align} \mathbf {h}_{j,t} \sim f(\mathbf {h}_{j,t-1},\mathbf {x}_{j,t}). \label {eq:ke} \\ p(Y_{j,t} \: | \: \mathbf {h}_{j,t}, {i}_{j,t}) \label {eq:rp}. \end {align}

For a student \(j\) at time step \(t\), The knowledge estimation component in Eq. \eqref{eq:ke} estimates the current knowledge state \(\mathbf {h}_{j,t}\) given the previous knowledge state \(\mathbf {h}_{j,t-1}\) and the student’s performance on the problem \(\mathbf {x}_{j,t}\) as inputs. The response prediction component in Eq. \eqref{eq:rp} outputs the prediction of the student’s likelihood of answering the next question \({Y}_{j, t}\) correctly given the current knowledge state \(\mathbf {h}_{j,t}\) and the next question index \({i}_{j,t}\) as input. During the learning process, the KT model needs to maximize the predicted likelihood across responses of all students, i.e., \(\sum _{j} \sum _{t} \log p(Y_{j,t} \:|\: Y_{j,1}, \ldots , Y_{j,t-1}).\)

We adopt the DKT setup detailed in [18] for consistency. Since incorporating causal learning into the base KT model introduces additional parameters, we use the gated recurrent unit (GRU) as the transition model instead of long short-term memory (LSTM) for computational efficiency. For response prediction, we simply use a single linear layer over the hidden states of the GRU. For causal discovery, i.e., learning the causal structure among skills, we use the causal GRU module detailed below.

Figure showing a casual GRU cell
Figure 1: The implementation of a causal GRU cell. All the GRU weight matrices, \(\mathbf {W}_z\), \(\mathbf {W}_r\), and \(\mathbf {W}\) are multiplied by the causal mask \(\mathbf {M} = \mathbf {P}\mathbf {L}\mathbf {P}^T\), resulting in \(\mathbf {W}_{z}'\), \(\mathbf {W}_{r}'\), and \(\mathbf {W}'\).

3.2 Causal GRU

We now detail the structure of the Causal GRU. From now on, we drop the student index \(j\) for notation simplicity. We define a permuted causal mask \(\mathbf {M}\) that represents the causal ordering and structure between skills. The \(\mathbf {L}\) matrix represents the causal structure/skill dependency, and the \(\mathbf {P}\) matrix represents the causal ordering. The permuted causal mask \(\mathbf {M}\) is calculated in Eq. \eqref{eq:tmatrix} as first multiplying \(\mathbf {P}\) by \(\mathbf {L}\) to obtain the updated causal structure and multiplying by \(\mathbf {P}^T\) to transform the causal structure into the original space.

The parameters in the Causal GRU are masked, i.e., element-wise multiplied by the permuted causal mask \(\mathbf {M}\) in Eq. \eqref{eq:mask}. By masking out some parameters, we zero out parameters that do not satisfy the causal graph. This step ensures that there is no relationship between the hidden states of the non-causally dependent skills in latent student knowledge states. The latest student knowledge state estimation \(\mathbf {h}_{t}\) is calculated in Eq. \eqref{eq:cgru}. The input \(\mathbf {x}_{t}\) is represented as a one-hot vector with the dimension size equals to the number of skills \(C\). The entry of value \(\pm 1\) represents whether the student can correctly answer the question corresponding to the skill. The input \(\mathbf {h}_{t-1}\) is the previous student knowledge state estimation. The implementation detail of a Causal GRU cell can be found in Fig. 1.

\begin {align} \mathbf {M} = \mathbf {P} \mathbf {L} \mathbf {P}^T, \label {eq:tmatrix}\\ \mathbf {W}^{'} = \mathbf {M} \odot \mathbf {W}, \label {eq:mask} \\ \mathbf {h}_t = GRU_c(\mathbf {h}_{t-1},\mathbf {x}_t). \label {eq:cgru} \end {align}

Figure showing the intuition behind casual GRU cell
Figure 2: The intuition behind causal GRU. Here, \(\mathbf {M} = \mathbf {P}\mathbf {L}\mathbf {P}^T\). \(M_{i,j} = 1\) if and only if \([\mathbf {h}_{t-1}]_j\) influences \([\mathbf {h}_{t}]_i\) where \(\mathbf {h}_t\) is the student’s knowledge state at time-step \(t\).

3.2.1 Causal Ordering

One important element of the causal GRU is the causal ordering matrix \(\mathbf {P}\), which we set to be a permutation matrix. By definition, a permutation matrix has exactly one entry of 1 in each row and each column and 0s elsewhere. Since multiplying a matrix by a permutation matrix permutes the order of the columns/rows of that matrix, the permutation matrix is naturally capable of sorting skills into order based on prerequisite relationships. However, the binary and discrete nature of the permutation matrix makes the learning process non-differentiable. To solve this problem, we introduce a relaxed version of the problem by approximating a permutation matrix with a doubly stochastic matrix, i.e., one where all entries are non-negative and the summation of each column/row is equal to 1, i.e., \( P_{i,k} \in [0,1], \:\sum _{k} P_{i,k} = 1 \: \forall i \: , \: \sum _{i} P_{i,k} = 1 \: \forall k \:\). Instead of learning a doubly stochastic matrix directly, which is very difficult, we learn a matrix of free parameters \(\bar {\mathbf {P}}\), from which we can obtain \(\mathbf {P}\) after applying the Sinkhorn operator [23] \(\mathbf {P} = \text {Sinkhorn}(\bar {\mathbf {P}})\).

The Sinkhorn operator works as follows: First, starting with the base matrix \(\bar {\mathbf {P}}\), we subtract the largest entry of the matrix from each entry, multiply each entry with the temperature hyper-parameter, and pass it through an exponential function. Second, we apply a series of row and column normalizations by dividing each entry of the column/row by the summation of all the entries in the column/row. In our implementation of the Sinkhorn operator, there are two hyper-parameters: temperature and unroll. The temperature hyper-parameter specifies the extent of the continuous relaxation: the larger the temperature hyper-parameter, the closer \(\mathbf {P}\)’s entries are to either \(0\) or \(1\). The unroll hyper-parameter specifies the number of times row/column normalization is carried out: the more times the normalization is applied, the closer \(\mathbf {P}\) is to satisfying the row/column normalization constraints.

3.2.2 Causal Structure

The other important element of the causal GRU is the causal structure/ skill dependency matrix \(\mathbf {L}\), which we set to be lower triangular. By definition, a lower triangular matrix is one in which all the elements above the principal diagonal of the matrix are 0. This matrix is important since it specifies the causal structure among the skills. Once the skills are ordered using the causal ordering matrix \(\mathbf {P}\), we apply the causal structure matrix \(\mathbf {L}\) to regularize student knowledge state transitions across time steps. Due to its lower-diagonal structure, an entry \(L_{i,k} > 0\) with \(i>k\) implies that skill \(k\) is a prerequisite of skill \(i\). Therefore, since the causal GRU weight matrices are masked by the \(\mathbf {PLP^T}\) matrix, at the next time step \(t\), the entry in the latent student knowledge vector that corresponds to skill \(i\), \([\mathbf {h}_t]_i\), depends only on the entries in the previous knowledge state that correspond to prerequisites of skill \(i\), i.e., \([\mathbf {h}_{t-1}]_k\,\forall k\) s.t. \(L_{i,k} > 0\).

The L matrix being lower diagonal ensures that the resulting causal structure is a DAG. This means that if skill \(C_1\) depends on \(C_2\) then \(C_2\) cannot depend on \(C_1\). As a concrete example, Fig. 2 visualizes the effect of applying a mask \(\mathbf {M} = \mathbf {P}\mathbf {L}\mathbf {P}^T\) on the skill matrix \(C\). The skill matrix \(C\) represents 5 skills where each skill is represented as a one-hot vector. The causal ordering matrix \(\mathbf {P}\) is applied to the skill matrix to give a skill ordering of \(C_3\), \(C_2\), \(C_1\), \(C_5\), \(C_4\) which is in the order of decreasing pre-requisites (\(C_3\) being the most pre-requisite of all skills). We further apply the \(\mathbf {L}\) matrix (in this case a lower diagonal matrix with all ones) that specifies that every subsequent skill depends on the preceding one. For example, it specifies that \(C_4\) depends on all of \(C_1\), \(C_2\), \(C_3\), and \(C_5\); \(C_5\) depends on \(C_1\), \(C_2\), and \(C_3\) and so on.

One easy choice for \(\mathbf {L}\) is to set its lower-diagonal part to be all ones; this setting means that every subsequent skill causally depends on all the previous skills. However, in practice, causal dependencies among skills nay not be this dense; most skills will only be causally related to a few other skills. To resolve this problem, we can make the \(\mathbf {L}\) matrix learnable by restricting the lower diagonal elements to be either 0 or 1. We do this by learning a matrix of free parameters \(\bar {\mathbf {L}}\), from which we can obtain \(\mathbf {L}\) after applying the element-wise sigmoid operator \(\mathbf {L} = \text {sigmoid}(\alpha \bar {\mathbf {L}})\). A large value of the temperature parameter \(\alpha > 0\) will push entries to be close to either 0 or 1 but not in between.

3.2.3 Skill Embeddings

We use a learnable dense embedding to represent each skill and alter both the input and the output layers of the causal GRU. We learn an embedding matrix \(\mathbf {E}\) where each column \(\mathbf {e}_c\) represents the embedding of skill \(c\). We treat the dimension of \(\mathbf {e}_c\) as a hyperparameter. For the input layer, we use another learnable embedding \(\mathbf {d}\), which is either added or subtracted from the skill embedding depending on the correctness of the previous answer. We then learn the input to the causal GRU using \(NN(\mathbf {e}_c \pm \mathbf {d})\) where \(NN\) is a single-layer neural network. For the output, we use \(p(Y_t) \sim NN_{o}([\mathbf {e}_c^T,\tilde {\mathbf {h}}_t^T]^T)\), where \(\tilde {\mathbf {h}}_t\) is a masked version of \(\mathbf {h}_t\) with the only non-zero entry being the one that corresponds to the skill of the next question that we are predicting. Here \(NN_{o}\) is a single-layer neural network that predicts the probability of the correct answer.

4. EXPERIMENTS

4.1 Data and Challenge Description

We participated in Task 3 of the NeurIPS Challenge co-hosted by Eedi, Microsoft Research, and Rice University [6]. The goal of this task is to discover the causal relationships between different skills, or constructs (as defined by Eedi, which means the smallest unit of learning; for example, “mental addition and subtraction” is a construct within the main topic “math”), and evaluate the effect of learning one skill on another. Questions in this dataset are multiple-choice, with a single correct option and three distractors that are designed to assess a single skill. The challenge hypothesis is that it is possible to discover the hidden relationship behind different skills through analyzing the responses to a large number of diagnostic questions. The challenge uses an \(F_1\) score-based metric which calculates the similarity between the predicted adjacency matrix \(\hat {A}\) and the true adjacency matrix \(A\).

Table 1: Results on different model variants.

Model

Variant

Leaderboard

\(F_{1}\) Score

No Embedding

No Adaptive

0.11

No Embedding

Adapative

0.17

Embedding (300D)

Adaptive

0.33

Embedding (300D)

Adaptive 

Learnable L

0.43

4.2 Model Learning and Hyperparameters

The dataset consists of 1855 skills and 6468 students. We set the default skill embedding dimension to 300. We use an adaptive strategy and start with small values of the temperature and unroll and linearly increase their values over a set of epochs. We set the initial temperature and unroll to 2 and 5 respectively and linearly increase the values with a factor of 2 and 5 respectively for every 10 epochs. We train the model for 50 epochs with a batch size of 64, and a learning rate of 5e-4 using four Nvidia Tesla 2080 GPUs with a GPU memory of 12GB each which takes about 6 hours. After training, to obtain the final causal structure matrix \(\mathbf {L}\) we apply a post-processing step. We define a hyperparameter \(\kappa \) such that all values of the \(\mathbf {L}\) matrix less than \(\kappa \) are set to 0 and all values greater than or equal to \(\kappa \) are set to 1.

4.3 Results and Discussion

In Table 1, we show the results of different model variants. We report the leaderboard \(F_1\) score obtained in our experiments. We see that the \(F_{1}\) score is 0.11 for the case where we are not using the skill embeddings. Using an adaptive strategy increases the \(F_{1}\) score by 0.06, which suggests that the adaptive strategy is helpful during model training. We also report the results corresponding to the skill embeddings and the learnable \(\mathbf {L}\). We see that using an embedding dimension of 300 almost doubles the \(F_{1}\) score. This observation confirms our hypothesis that using skill embeddings increases the representational capacity of the neural network model and hence performs better. When the causal structure matrix \(\mathbf {L}\) is learnable, we see that we get a further 0.1 increase in the \(F_{1}\) score. The increase in the \(F_{1}\) score on using the learnable \(\mathbf {L}\) configuration of the model shows that it is better to learn the explicit causal dependence of skills instead of assuming a dense representation where each skill depends on all the skills preceding it.

5. CONCLUSIONS AND FUTURE WORK

In this work, we proposed a conceptual method for learning causal structure among skills from student response data, as a part of our solution to the NeurIPS 2022 Challenge on Causal Insights for Learning Paths in Education. Our method is a novel causal knowledge tracing method that enables us to learn the causal structure in an end-to-end manner while performing knowledge tracing. Unfortunately, due to space limitations, we cannot show a qualitative example of the learned causal structure among skills. We believe that our work should inspire future works in the direction of building causal knowledge tracing methods on observational student response data. First, it is important to evaluate the accuracy of the the learned causal structure between skills, either against human domain experts or via A/B testing. Second, it is important to apply our causal module to more flexible knowledge tracing methods, such as attention-based methods, to see whether it is applicable and effective. Third, it is important to develop ways to leverage both the opinion of human experts and our data-driven causal discovery model, in a human-in-the-loop manner. The former may be less accurate but the latter requires extensive training data; a hybrid human-AI collaboration may be able to take the best from both sides.

6. ACKNOWLEDGEMENTS

The authors thank the NSF (under grants 1917713, 2118706, 2202506, 2215193) for partially supporting this work. We also thank the organizers of the 2022 NeurIPS Challenge on Causal Insights for Learning Paths in Education for proposing a new and meaningful task.

7. REFERENCES

  1. A. Corbett and J. Anderson. Knowledge tracing: Modeling the acquisition of procedural knowledge. User Model. User-adapted Interact., 4(4):253–278, Dec. 1994.
  2. M. C. Desmarais. Mapping question items to skills with non-negative matrix factorization. ACM SIGKDD Explorations Newsletter, 13(2):30–36, 2012.
  3. T. Gervet, K. Koedinger, J. Schneider, T. Mitchell, et al. When is deep learning the best approach to knowledge tracing? Journal of Educational Data Mining, 12(3):31–54, 2020.
  4. A. Ghosh, N. Heffernan, and A. S. Lan. Context-aware attentive knowledge tracing. In Proc. ACM SIGKDD, pages 2330–2339, 2020.
  5. A. Ghosh, J. Raspat, and A. Lan. Option tracing: Beyond correctness analysis in knowledge tracing. In Int. Conf. Artif. Intell. Educ., pages 137–149. Springer, 2021.
  6. W. Gong, D. Smith, Z. Wang, C. Barton, S. Woodhead, N. Pawlowski, J. Jennings, and C. Zhang. Instructions and guide: Causal insights for learning paths in education. arXiv preprint arXiv:2208.12610, 2022.
  7. A. Hyvärinen, K. Zhang, S. Shimizu, and P. O. Hoyer. Estimation of a structural vector autoregression model using non-gaussianity. Journal of Machine Learning Research, 11(5), 2010.
  8. P. Kaur, A. Polyzou, and G. Karypis. Causal inference in higher education: Building better curriculums. In Proceedings of the Sixth (2019) ACM Conference on Learning@ Scale, pages 1–4, 2019.
  9. M. Khajah, Y. Huang, J. González-Brenes, M. Mozer, and P. Brusilovsky. Integrating knowledge tracing and item response theory: A tale of two frameworks. In Proc. Int. Workshop Personalization Approaches Learn. Environ., volume 1181, pages 7–15, 2014.
  10. A. S. Lan, C. Studer, and R. G. Baraniuk. Time-varying learning and content analytics via sparse factor analysis. In Proc. ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), pages 452–461, Aug. 2014.
  11. N. Liu, Z. Wang, R. Baraniuk, and A. Lan. Open-ended knowledge tracing for computer science education. In Conference on Empirical Methods in Natural Language Processing, pages 3849–3862, 2022.
  12. Q. Liu, Z. Huang, Y. Yin, E. Chen, H. Xiong, Y. Su, and G. Hu. Ekt: Exercise-aware knowledge tracing for student performance prediction. IEEE Trans. Knowl. Data Eng., 33(1):100–115, 2019.
  13. S. Minn, J.-J. Vie, K. Takeuchi, H. Kashima, and F. Zhu. Interpretable knowledge tracing: Simple and efficient student modeling with causal relations. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 12810–12818, 2022.
  14. S. Pandey and J. Srivastava. Rkt: Relation-aware self-attention for knowledge tracing. arXiv preprint arXiv:2008.12736, 2020.
  15. Z. A. Pardos and N. T. Heffernan. Modeling individualization in a Bayesian networks implementation of knowledge tracing. In Proc. Int. Conf. User Model. Adaptation Personalization, pages 255–266, 2010.
  16. J. Pearl. Causal inference in statistics: An overview. Statistics surveys, 3:96–146, 2009.
  17. C. Piech, J. Bassen, J. Huang, S. Ganguli, M. Sahami, L. J. Guibas, and J. Sohl-Dickstein. Deep knowledge tracing. In Proc. NeurIPS, pages 505–513, 2015.
  18. C. Piech, J. Bassen, J. Huang, S. Ganguli, M. Sahami, L. J. Guibas, and J. Sohl-Dickstein. Deep knowledge tracing. Advances in neural information processing systems, 28, 2015.
  19. S. Ritter, J. R. Anderson, K. R. Koedinger, and A. Corbett. Cognitive tutor: Applied research in mathematics education. Psychonomic Bulletin & Review, 14(2):249–255, 2007.
  20. A. Sales, A. Botelho, T. Patikorn, and N. T. Heffernan. Using big data to sharpen design-based inference in a/b tests. In Proceedings of the Eleventh International Conference on Educational Data Mining, 2018.
  21. A. C. Sales and J. F. Pane. Student log-data from a randomized evaluation of educational technology: A causal case study. Journal of Research on Educational Effectiveness, 14(1):241–269, 2021.
  22. D. Shin, Y. Shim, H. Yu, S. Lee, B. Kim, and Y. Choi. Saint+: Integrating temporal features for ednet correctness prediction. In 11th Int. Learn. Analytics Knowl. Conf., pages 490–496, 2021.
  23. R. Sinkhorn. A relationship between arbitrary positive matrices and doubly stochastic matrices. The annals of mathematical statistics, 35(2):876–879, 1964.
  24. S. Sonkar, A. E. Waters, A. S. Lan, P. J. Grimaldi, and R. G. Baraniuk. qdkt: Question-centric deep knowledge tracing. In Proceedings of the International Conference on Educational Data Mining, 2020.
  25. C. Wang, W. Ma, M. Zhang, C. Lv, F. Wan, H. Lin, T. Tang, Y. Liu, and S. Ma. Temporal cross-effects in knowledge tracing. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining, pages 517–525, 2021.
  26. Y. Wang and N. Heffernan. Extending knowledge tracing to allow partial credit: Using continuous versus binary nodes. In Int. conf. artif. intell. educ., pages 181–188. Springer, 2013.
  27. Y. Yang, J. Shen, Y. Qu, Y. Liu, K. Wang, Y. Zhu, W. Zhang, and Y. Yu. Gikt: A graph-based interaction model for knowledge tracing. In Proc. Joint Eur. Conf. Mach. Learn. Knowl. Discovery Databases, 2020.
  28. M. V. Yudelson, K. R. Koedinger, and G. J. Gordon. Individualized bayesian knowledge tracing models. In Int. Conf. artif. intell. educ., pages 171–180. Springer, 2013.
  29. J. Zhang, X. Shi, I. King, and D.-Y. Yeung. Dynamic key-value memory networks for knowledge tracing. In Proc. Int. Conf. World Wide Web, pages 765–774, Apr. 2017.

APPENDIX

A. STRUCTURAL EQUATION MODELING

In SEM, given a collection of random variables \(\mathbf {x} = (x_1, \cdots , x_D)\) and a causal directed acyclic graph (DAG) \(\mathcal {G}\), each variable \(x_i\) is generated using its parents \(Pa(i;\mathcal {G})\) in the DAG and an exogenous noise variable \(\epsilon _i\): \(x_i = f_i(\{x_j\}_{j\in Pa(i;\mathcal {G})}) +\epsilon _i\). The structural vector autoregression (SVAR) model extends SEM to random variables that form a time series \(\mathbf {x}_t = (x_{1,t}, \cdots , x_{D,t})\) for time steps \(t\in \{1,\cdots ,T\}\) [7]. The influence of one variable on others can be instantaneous or lagged behind for a few time steps. The SEM model is given by \begin {align*} x_{i,t} = \sum _{\tau =0}^k f_{ i,\tau }(\{x_{j, t-\tau }\}_{j \in Pa(i, t,\tau ;\mathcal {G})}) +\epsilon _{i,t}, \end {align*}

where \(Pa(i, t,\tau ;\mathcal {G})\) are random variables from time step \(t-\tau \) that influence random variable \(x_{i,t}\). One can also use a single latent state \(\mathbf {h}{\cdot ,t}\) to model the influence of past random variables. The SEM becomes \begin {align*} \mathbf {h}_{i,t} = f_{i} (\{h_{j,t-1}\}_{j\in Pa(i;\mathcal {G})})+\epsilon _i. \end {align*}

B. SINKHORN OPERATION

Figure showing a the working of the Sinkhorn Operator
Figure 3: The Sinkhorn operator is a smooth operator that outputs an approximate permutation matrix.

Fig. 3 shows the working of the Sinkhorn operator.

C. ADDITIONAL RESULTS

Table 2: Results on varying the cutoff for the \(\mathbf {L}\) matrix, \(\kappa \).
\(\kappa \) \(F_1\) Score
0.42 0.43
0.45 0.43
0.4350.43
0.48 0.43
0.4950.43
0.51 0.33
0.5250.18

We report the results obtained using different hyperparameters of the learnable \(\mathbf {L}\) model configuration. In Table 2, we show the results across different \(\kappa \) values. We vary the values from 0.42 to 0.525 and observe that using any values out of this range gives a leaderboard \(F_{1}\) score of 0. Among the experimented values for \(\kappa \) we see that we obtain a maximum \(F_{1}\) score of 0.43 for all values in the range of 0.42 to 0.495. The maximum \(F_{1}\) score for \(\kappa \) values in the range of 0.42 to 0.495 means that using very large or very less values of \(\kappa \) does not give the optimal skill dependency.

Figure showing an example of learned causal ordering among skills in actual student response data
Figure 4: An example of learned causal ordering among skills in actual student response data.

We perform qualitiative analysis of our proposed method. Fig. 4 shows an example from the DAG obtained from the learned adjacency matrix for causal relations. Here, we represent the constructs based on their subject names, and an arrow from subject \(i\) to subject \(j\) implies that subject \(i\) is a pre-requisite of subject \(j\). Here, in the figure, we can see that “Counting” is the most pre-requisite skill. The subsequent skills in fractions depend on “Counting”. Calculating areas of simple figures depends on fraction multiplication. In the same way, calculating the volume depends on calculating the area. Hence, from this, we can see that we are able to learn a meaningful DAG using our methodology.

1https://eedi.com/projects/neurips-2022

2The code for our solution can be found at: https://github.com/umass-ml4ed/Neurips-Challenge-22


© 2023 Copyright is held by the author(s). This work is distributed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0) license.