Neural Cognitive Diagnosis for Intelligent Education Systems

9
Neural Cognitive Diagnosis for Intelligent Education Systems Fei Wang 1 , Qi Liu 1* , Enhong Chen 1* , Zhenya Huang 1 Yuying Chen 1 , Yu Yin 1 , Zai Huang 1 , Shijin Wang 2 1 Anhui Province Key Lab. of Big Data Analysis and Application, School of Computer Science and Technology, University of Science and Technology of China 2 iFLYTEK Research [email protected], {qiliuql, cheneh}@ustc.edu.cn, {huangzhy, cyy33222, yxonic, huangzai}@mail.ustc.edu.cn, sjwang3@iflytek.com Abstract Cognitive diagnosis is a fundamental issue in intelligent ed- ucation, which aims to discover the proficiency level of stu- dents on specific knowledge concepts. Existing approaches usually mine linear interactions of student exercising process by manual-designed function (e.g., logistic function), which is not sufficient for capturing complex relations between stu- dents and exercises. In this paper, we propose a general Neu- ral Cognitive Diagnosis (NeuralCD) framework, which incor- porates neural networks to learn the complex exercising inter- actions, for getting both accurate and interpretable diagnosis results. Specifically, we project students and exercises to fac- tor vectors and leverage multi neural layers for modeling their interactions, where the monotonicity assumption is applied to ensure the interpretability of both factors. Furthermore, we propose two implementations of NeuralCD by specializing the required concepts of each exercise, i.e., the NeuralCDM with traditional Q-matrix and the improved NeuralCDM+ ex- ploring the rich text content. Extensive experimental results on real-world datasets show the effectiveness of NeuralCD framework with both accuracy and interpretability. 1 Introduction Cognitive diagnosis is a necessary and fundamental task in many real-world scenarios such as games (Chen and Joachims 2016), medical diagnosis (Guo et al. 2017), and education. Specifically, in intelligent education sys- tems (Anderson et al. 2014; Burns et al. 2014), cognitive diagnosis aims to discover the states of students in the learn- ing process, such as their proficiencies on specific knowl- edge concepts (Liu et al. 2018). Figure 1 shows a toy exam- ple of cognitive diagnosis. Generally, students usually first choose to practice a set of exercises (e.g., e 1 , ··· ,e 4 ) and leave their responses (e.g., right or wrong). Then, our goal is to infer their actual knowledge states on the corresponding concepts (e.g., Trigonometric Function). In practice, these diagnostic reports are necessary as they are the basis of fur- ther services, such as exercise recommendation and targeted training (Kuh et al. 2011). * Corresponding Authors. Copyright c 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. In the literature, massive efforts have been devoted for cognitive diagnosis, such as Deterministic Inputs, Noisy- And gate model (DINA) (De La Torre 2009), Item Response Theory (IRT) (Embretson and Reise 2013), Multidimen- sional IRT (MIRT) (Reckase 2009) and Matrix Factoriza- tion (MF) (Koren, Bell, and Volinsky 2009). Despite achiev- ing some effectiveness, these works rely on handcrafted in- teraction functions that just combine the multiplication of student’s and exercise’s trait features linearly, such as lo- gistic function (Embretson and Reise 2013) or inner prod- uct (Koren, Bell, and Volinsky 2009), which may not be sufficient for capturing the complex relationship between students and exercises (DiBello, Roussos, and Stout 2006). Besides, the design of specific interaction functions is also labor-intensive since it usually requires professional exper- tise. Therefore, it is urgent to find an automatic way to learn the complex interactions for cognitive diagnosis instead of manually designing them. In this paper, we address this issue in a principled way of proposing a Neural Cognitive Diagnosis (NeuralCD) frame- work by incorporating neural networks to model complex non-linear interactions. Although the capability of neural networks to approximate continuous functions has been proved in many domains, such as natural language process- ing (Zhang et al. 2018) and recommender systems (Song et al. 2019), it is still highly nontrivial to adapt to cognitive di- agnosis due to the following domain challenges. First, the black-box nature of neural networks makes them difficult to get explainable diagnosis results. That is to say, it is diffi- cult to explicitly realize how much a student has mastered a certain knowledge concept (e.g., Equation). Second, tradi- tional models are designed manually with non-neural func- tions, which makes it hard for them to leverage exercise text content. However, with neural network, it is worthy of find- ing ways to explore the rich information contained in exer- cise text content for cognitive diagnosis. To address these challenges, we propose a NeuralCD framework to approximate interactions between students and exercises, yet preserving the explainability. We first project students and exercises to factor vectors and lever- age multi-layers for modeling the complex interactions of student answering exercises. To ensure the interpretability

Transcript of Neural Cognitive Diagnosis for Intelligent Education Systems

Page 1: Neural Cognitive Diagnosis for Intelligent Education Systems

Neural Cognitive Diagnosis for Intelligent Education Systems

Fei Wang1, Qi Liu1*, Enhong Chen1∗, Zhenya Huang1

Yuying Chen1, Yu Yin1, Zai Huang1, Shijin Wang2

1Anhui Province Key Lab. of Big Data Analysis and Application,School of Computer Science and Technology, University of Science and Technology of China

2iFLYTEK [email protected], {qiliuql, cheneh}@ustc.edu.cn,

{huangzhy, cyy33222, yxonic, huangzai}@mail.ustc.edu.cn, [email protected]

Abstract

Cognitive diagnosis is a fundamental issue in intelligent ed-ucation, which aims to discover the proficiency level of stu-dents on specific knowledge concepts. Existing approachesusually mine linear interactions of student exercising processby manual-designed function (e.g., logistic function), whichis not sufficient for capturing complex relations between stu-dents and exercises. In this paper, we propose a general Neu-ral Cognitive Diagnosis (NeuralCD) framework, which incor-porates neural networks to learn the complex exercising inter-actions, for getting both accurate and interpretable diagnosisresults. Specifically, we project students and exercises to fac-tor vectors and leverage multi neural layers for modeling theirinteractions, where the monotonicity assumption is applied toensure the interpretability of both factors. Furthermore, wepropose two implementations of NeuralCD by specializingthe required concepts of each exercise, i.e., the NeuralCDMwith traditional Q-matrix and the improved NeuralCDM+ ex-ploring the rich text content. Extensive experimental resultson real-world datasets show the effectiveness of NeuralCDframework with both accuracy and interpretability.

1 IntroductionCognitive diagnosis is a necessary and fundamental taskin many real-world scenarios such as games (Chen andJoachims 2016), medical diagnosis (Guo et al. 2017),and education. Specifically, in intelligent education sys-tems (Anderson et al. 2014; Burns et al. 2014), cognitivediagnosis aims to discover the states of students in the learn-ing process, such as their proficiencies on specific knowl-edge concepts (Liu et al. 2018). Figure 1 shows a toy exam-ple of cognitive diagnosis. Generally, students usually firstchoose to practice a set of exercises (e.g., e1, · · · , e4) andleave their responses (e.g., right or wrong). Then, our goal isto infer their actual knowledge states on the correspondingconcepts (e.g., Trigonometric Function). In practice, thesediagnostic reports are necessary as they are the basis of fur-ther services, such as exercise recommendation and targetedtraining (Kuh et al. 2011).

∗Corresponding Authors.Copyright c© 2020, Association for the Advancement of ArtificialIntelligence (www.aaai.org). All rights reserved.

In the literature, massive efforts have been devoted forcognitive diagnosis, such as Deterministic Inputs, Noisy-And gate model (DINA) (De La Torre 2009), Item ResponseTheory (IRT) (Embretson and Reise 2013), Multidimen-sional IRT (MIRT) (Reckase 2009) and Matrix Factoriza-tion (MF) (Koren, Bell, and Volinsky 2009). Despite achiev-ing some effectiveness, these works rely on handcrafted in-teraction functions that just combine the multiplication ofstudent’s and exercise’s trait features linearly, such as lo-gistic function (Embretson and Reise 2013) or inner prod-uct (Koren, Bell, and Volinsky 2009), which may not besufficient for capturing the complex relationship betweenstudents and exercises (DiBello, Roussos, and Stout 2006).Besides, the design of specific interaction functions is alsolabor-intensive since it usually requires professional exper-tise. Therefore, it is urgent to find an automatic way to learnthe complex interactions for cognitive diagnosis instead ofmanually designing them.

In this paper, we address this issue in a principled way ofproposing a Neural Cognitive Diagnosis (NeuralCD) frame-work by incorporating neural networks to model complexnon-linear interactions. Although the capability of neuralnetworks to approximate continuous functions has beenproved in many domains, such as natural language process-ing (Zhang et al. 2018) and recommender systems (Song etal. 2019), it is still highly nontrivial to adapt to cognitive di-agnosis due to the following domain challenges. First, theblack-box nature of neural networks makes them difficult toget explainable diagnosis results. That is to say, it is diffi-cult to explicitly realize how much a student has mastereda certain knowledge concept (e.g., Equation). Second, tradi-tional models are designed manually with non-neural func-tions, which makes it hard for them to leverage exercise textcontent. However, with neural network, it is worthy of find-ing ways to explore the rich information contained in exer-cise text content for cognitive diagnosis.

To address these challenges, we propose a NeuralCDframework to approximate interactions between studentsand exercises, yet preserving the explainability. We firstproject students and exercises to factor vectors and lever-age multi-layers for modeling the complex interactions ofstudent answering exercises. To ensure the interpretability

Page 2: Neural Cognitive Diagnosis for Intelligent Education Systems

Trigonometric

Function

Coordinates

EquationAbsolute

Value

Statistics

Response Logs

Diagnostic Report

Practice Cognitive

Diagnosis

Proficiency

00.5

1.0

Exercise

2e

1e

3e

4e

Knowledge Concepts

Trigonometric Function

Statistics

Coordinates, Equation

Absolute Value, Equation

Exercise

Response

1e 2e 3e 4e

Figure 1: A toy example of cognitive diagnosis. The student choose some exercises for practice and leave the response logs.With cognitive diagnosis methods, we get the diagnostic report containing the student’s proficiency on each knowledge concept.

of both factors, we apply the monotonicity assumption tak-ing from educational property (Reckase 2009) on the multi-layers. Then, we propose two implementations on the ba-sis of the general framework, i.e., NeuralCDM and Neural-CDM+. In NeuralCDM, we simply extract exercise factorvectors from traditional Q-matrix (an example is shown inFigure 6) and achieve the monotonicity property with pos-itive full connection layers, which shows feasibility of theframework. While in NeuralCDM+, we demonstrate howinformation from exercise text can be explored with neu-ral network to extend the framework. Particularly, our Neu-ralCD is a general framework since it can cover many tradi-tional models such as MF, IRT and MIRT. Finally, we con-duct extensive experiments on real-world datasets, and theresults show the effectiveness of NeuralCD framework withboth accuracy and interpretability guarantee.

Our code of NeuralCDM is available athttps://github.com/bigdata-ustc/NeuralCD.

2 Related WorkIn this section, we briefly review the related works from thefollowing three aspects.Cognitive Diagnosis. Existing works about student cogni-tive diagnosis mainly came from educational psychologyarea. DINA (De La Torre 2009; von Davier 2014) andIRT (Embretson and Reise 2013) were two of the most typ-ical works, which model the result of a student answer-ing an exercise as the interaction between the trait fea-tures of the student (θ) and the exercise (β). Specifically,in DINA, θ and β were binary, where β came directly fromQ-matrix (a human labeled exercise-knowledge correlationmatrix, an example is showed in Figure 6). Another two ex-ercise factors, i.e. guessing and slipping (parameterized asg and s) are also taken into consideration. The probabilityof student i correctly answering exercise j was modeled asP (rij = 1|θi) = g

1−ηijj (1 − sj)ηij , where ηij =

∏k θ

βjk

ik .On the other hand, in IRT, θ and β were unidimensionaland continuous latent traits, indicating student ability andexercise difficulty respectively. The interaction between thetrait features was modeled in a logistic way, e.g., a sim-ple version is sigmoid(a(θ − β)), where a is the exercisediscrimination parameter. Although extra parameters wereadded in IRT (Fischer 1995; Lord 2012) and latent trait wasextended to multidimensional(MIRT) (Adams, Wilson, and

Wang 1997; Reckase 2009), most of their item responsefunctions were still logistic-like. These traditional modelsdepended on manually designed functions, which was labor-intensive and restricted their scope of applications.Matrix Factorization. Recently, some researches from datamining perspective have demonstrated the feasibility of MFfor cognitive diagnosis. Student and exercise correspond touser and item in matrix factorization (MF). For instance,Toscher et al. (2010) improved SVD (Singular Value De-composition) methods to factor the score matrix and getstudents and exercises’ latent trait vectors. Thai-Nghe etal. (2010) applied some recommender system techniques in-cluding matrix factorization in the educational context, andcompared it with traditional regression methods. Besides,Thai-Nghe et al. (2015) proposed a multi-relational factor-ization approach for student modeling in the intelligent tu-toring systems. Despite their effectiveness in predicting stu-dents’ scores on exercises, the latent trait vectors in MF isnot interpretable for cognitive diagnosis, i.e. there is no clearcorrespondence between elements in trait vectors and spe-cific knowledge concepts.Artificial Neural Network. Techniques using artificial neu-ral network have reached state-of-the-art in many areas,e.g., speech recognition (Chan et al. 2016), text classifica-tion (Zhang, Zhao, and LeCun 2015) and image caption-ing (Wang, Chen, and Hu 2019). There are also some ed-ucational applications such as question difficulty predic-tion (Huang et al. 2017), code education (Wu et al. 2019),formula image transcribing (Yin et al. 2018) and studentperformance prediction (Huang et al. 2019). However, us-ing neural network for cognitive diagnosis is nontrivial asit performs poorly in parameter interpretation due to its in-herent traits. To the best of our knowledge, deep knowledgetracing (DKT) (Piech et al. 2015) was the first attempt tomodel student learning process using recurrent neural net-work. However, DKT aims to predict students’ scores, anddoes not make a distinction between an exercise and theknowledge concepts it contains, thus it’s unsuitable for cog-nitive diagnosis. Few works with neural network have highinterpretability for student cognitive diagnosis. Towards thisend, in this paper we propose a neural cognitive diagnosis(NeuralCD) framework which borrows concepts from ed-ucational psychology and combine them with interactionfunctions learned from data. NeuralCD could achieve bothhigh accuracy and interpretation with neural network. Be-

Page 3: Neural Cognitive Diagnosis for Intelligent Education Systems

sides, the framework is general that can cover many traditionmodels, and at the same time easy for extension.

3 Neural Cognitive DiagnosisWe first formally introduce cognitive diagnosis task. Thenwe describe the details of NeuralCD framework. After that,we design a specific diagnostic network NeuralCDM withtraditional Q-matrix to show the feasibility of the frame-work, and an improved NeuralCDM+ by incorporating exer-cise text content for better performance. Finally, we demon-strate the generality of NeuralCD framework by showing itsclose relationship with some traditional models.

3.1 Task OverviewSuppose there are N Students, M Exercises and K Knowl-edge concepts at a learning system, which can be repre-sented as S = {s1, s2, . . . , sN}, E = {e1, e2, . . . , eM}and Kn = {k1, k2, . . . , kK} respectively. Each student willchoose some exercises for practice, and the response logs Rare denoted as set of triplet (s, e, r) where s ∈ S, e ∈ E andr is the score (transferred to percentage) that student s goton exercise e. In addition, we have Q-matrix (usually labeledby experts) Q = {Qij}M×K , where Qij = 1 if exercise eirelates to knowledge concept kj and Qij = 0 otherwise.

Problem Definition Given students’ response logs R andthe Q-matrix Q, the goal of our cognitive diagnosis task is tomine students’ proficiency on knowledge concepts throughthe student performance prediction process.

3.2 Neural Cognitive Diagnosis FrameworkGenerally, for a cognitive diagnostic system, there are threeelements need to be considered: student factors, exercisefactors and the interaction function among them (DiBello,Roussos, and Stout 2006). In this paper, we propose a gen-eral NeuralCD framework to address them by using multi-layer neural network modeling, which is shown in Figure 2.Specifically, for each response log, we use one-hot vectorsof the corresponding student and exercise as input and obtainthe diagnostic factors of the student and exercise. Then theinteractive layers learn the interaction function among thefactors and output the probability of correctly answering theexercise. After training, we get students’ proficiency vectorsas diagnostic results. Details are introduced as bellow.

Student Factors. Student factors characterize the traits ofstudents, which would affect the students’ response to exer-cises. As our goal is to mine students’ proficiency on knowl-edge concepts, we do not use the latent trait vectors as in IRTand MIRT, which is not explainable enough to guide stu-dents’ self-assessment. Instead, we design the student fac-tors as explainable vectors similar to DINA, but has a majordifference that they are continuous. Specifically, We use avector F s to characterize a student, namely proficiency vec-tor. Each entry of F s is continuous ([0,1]), which indicatesthe student’s proficiency on a knowledge concept. For ex-ample, F s = [0.9, 0.2] indicates a high mastery on the firstknowledge concept but low mastery on the second. F s is gotthrough the parameter estimation process.

Student One-hot Exercise One-hot

y

Layer 1

Layer n

...

1 0 0 0 0 0 0 0 1 0

Diagnostic Factors

Input Layer

Interactive Layers

Proficiency Vector F s

Output Layer

Monotonicity

Assumption

Student Factors

Exercise Factors

Interaction Function

Other Exercise Factors F other

Knowledge Relevancy Vector F kn

Figure 2: Structure of NeuralCD framework.

Exercise Factors. Exercise factors denote the factors thatcharacterize the traits of exercises. We divide exercise fac-tors into two categories. The first indicates the relationshipbetween exercises and knowledge concepts, which is funda-mental as we need it to make each entry of F s correspondto a specific knowledge concept for our diagnosis goal. Wecall it knowledge relevancy vector and denote it as F kn. F knhas the same dimension as F s, with the ith entry indicat-ing the relevancy between the exercise and the knowledgeconcept ki. Each entry of F kn is non-negative. F kn is pre-viously given (e.g., obtained from Q-matrix). Other factorsare of the second type and are optional. Factors from IRTand DINA such as knowledge difficulty, exercise difficultyand discrimination can be incorporated if reasonable.

Interaction Function. We use artificial neural networkto obtain the interaction function for the following rea-sons. First, the neural network has been proven to be ca-pable of approximating any continuous function (Hornik,Stinchcombe, and White 1989). The strong fitting abilityof neural network makes it competent for capturing rela-tionships among student and exercise factors. Second, withneural network, the interaction function can be learned fromdata with few assumptions (that behind traditional models).This makes NeuralCD more general and can be applied inbroad areas. Third, the framework can be highly extendablewith neural network. For instance, extra information suchas exercise texts can be integrated in with neural network(We will discuss its extendability in the following subsec-tions.). Mathematically, we formulate the output of Neu-ralCD framework as:

y = ϕn(. . . ϕ1(Fs, F kn, F other, θf )), (1)

where ϕi denotes the mapping function of the ith MLPlayer; F other denotes factors other than F s and F kn (e.g.,difficulty); and θf denotes model parameters of all the inter-active layers.

However, due to some intrinsic characteristics, neuralnetworks usually have poor performance on interpreta-tion (Samek et al. 2016). Fortunately, we find that the mono-tonicity assumption, which is used in some IRT and MIRTmodels (Reckase 2009), can be utilized to ensure the inter-pretation of student and exercise factors. Monotonicity as-sumption is general and reasonable in almost all circum-

Page 4: Neural Cognitive Diagnosis for Intelligent Education Systems

stance, thus it has little influence on the generality of Neu-ralCD framework. The assumption is defined as follows:

Monotonicity Assumption The probability of correct re-sponse to the exercise is monotonically increasing at any di-mension of the student’s knowledge proficiency.

This assumption should be converted as a property of theinteraction function. Intuitively, we assume student s to an-swer exercise e correctly. During training, the optimizationalgorithm should increase the student’s proficiency if themodel output a wrong prediction (i.e., a value below 0.5).The increment of each knowledge proficiency is otherwisecontrolled by F kn.

After introducing the structure of NeuralCD framework,we will next show some specific implementations. We firstdesign a diagnostic model based on NeuralCD with ex-tra exercise factors (i.e., knowledge difficulty and exer-cise discrimination)(§3.3), and further show its extendabil-ity by incorporating text information (§3.4) and generalityby demonstrating how it covers traditional models (§3.5).

3.3 Neural Cognitive Diagnosis ModelHere we introduce a specific neural cognitive diagnosismodel (NeuralCDM) under NeuralCD framework. Figure 3illustrates the structure of NeuralCDM.Student Factors. In NeuralCDM, each student is repre-sented with a knowledge proficiency vector. The student fac-tor F s aforementioned is hs here, and hs is obtained bymultiplying the student’s one-hot representation vector xswith a trainable matrix A. That is,

hs = sigmoid(xs ×A), (2)

in which hs ∈ (0, 1)1×K ,xs ∈ {0, 1}1×N ,A ∈ RN×K .Exercise Factors. As for each exercise, the aforementionedexercise factor F kn is Qe here, which directly comes fromthe pre-given Q-matrix:

Qe = xe ×Q, (3)

whereQe ∈ {0, 1}1×K , xe ∈ {0, 1}1×M is the one-hot rep-resentation of the exercise. In order to make a more precisediagnosis, we adopt other two exercise factors: knowledgedifficulty hdiff and exercise discrimination hdisc. hdiff ∈(0, 1)1×K , indicates the difficulty of each knowledge con-cept examined by the exercise, which is extended from exer-cise difficulty used in IRT. hdisc ∈ (0, 1), used in some IRTand MIRT models, indicates the capability of the exercise todifferentiate between those students whose knowledge mas-tery is high from those with low knowledge mastery. Theycan be obtained by:

hdiff = sigmoid(xe ×B),B ∈ RM×K (4)

hdisc = sigmoid(xe ×D),D ∈ RM×1 (5)

where B and D are trainable matrices.Interaction Function. The first layer of the interaction lay-ers is inspired by MIRT models. We formulate it as:

x = Qe ◦ (hs − hdiff )× hdisc, (6)

where ◦ is element-wise product. Following are two full con-nection layers and an output layer:

f1 = φ(W1 × xT + b1), (7)f2 = φ(W2 × f1 + b2), (8)y = φ(W3 × f2 + b3), (9)

where φ is the activation function. Here we use Sigmoid.Different methods can be used to satisfy the monotonicity

assumption. We adopt a simple strategy: restrict each ele-ment of W1,W2,W3 to be positive. It can be easily provedthat ∂y

∂hsi

is positive for each entry hsi in hs. Thus monotonic-ity assumption is always satisfied during training.

The loss function of NeuralCDM is cross entropy betweenoutput y and true label r:

lossCDM = −∑i

(ri log yi + (1− ri) log(1− yi)). (10)

After training, the value of hs is what we get as diagnosisresult, which denotes the student’s knowledge proficiency.

3.4 NeuralCD Extension with Text InformationIn NeuralCDM and some traditional methods (e.g., DINA),Q-matrix is the source of information about the exerciseknowledge concept. However, manually-labeled Q-matrixmay be deficient because of inevitable errors and subjec-tive bias (Liu, Xu, and Ying 2012; DiBello, Roussos, andStout 2006). On the other hand, exercise texts have beenproved to be highly related to some exercise features (e.g.,difficulty, relevant knowledge concepts) (Su et al. 2018;Huang et al. 2017), thus it can be leveraged to refine the Q-matrix. For example, in Q-matrix, maybe only ’Equation’ islabeled for an equation solving exercise. However, we maydiscover that ’Division’ is also required due to the existenceof ’÷’ in the text. Traditional cognitive models didn’t lever-age text content due to the limitation of their handcraft non-neural interaction functions. However, with neural network,we are able to incorporate text information into our frame-work. We denote the extended model as NeuralCDM+, andpresent its structure in Figure 4.

Specifically, we first pre-train a CNN (convolutional neu-ral network) to predict knowledge concepts related to theinput exercise. CNN has advantage of extracting local in-formation in text processing, thus it’s able to capture im-portant words from texts (e.g., words that are highly rela-tive to certain knowledge concepts). The network takes con-catenated word2vec embedding of words in texts as input,and output the relevancy of each predefined knowledge con-cept (that has occurred in data) to the exercise. Human-labeled Q-matrix is used as label for training. We defineV ki = {Vij1 , Vij2 , . . . , Vijk} as the set of top-k knowledgeconcepts of exercise ei outputted by the CNN.

Then we combine V ki with Q-matrix. Although thereare defects in human-labeled Q-matrix, it still has highconfidence. Thus we consider knowledge concepts labeledby Q-matrix are more relative than concepts in {kj |kj ∈V ki and Qij = 0}. To achieve this, we adopt a pairwise

Page 5: Neural Cognitive Diagnosis for Intelligent Education Systems

-( ) hdisc×

Knowledge

Proficiency

Knowledge

Difficulty

Student

One-hotExercise

One-hot

×A ×B

Knowledge

RelevancyExercise

Discrimination

×D×Q

y

Positive Full Connection

Positive Full Connection

1 0 0 0 0 0 0 0 1 0

Positive Full Connection

hdiffQe h

s

Figure 3: Neural cognitive diagnosis model. The color oforange, blue and yellow indicate student factors, exercisefactors and interacion function respectively.

-( ) hdisc×

Knowledge

Proficiency

Knowledge

Difficulty

Student

One-hot

Exercise

One-hot

×A ×B

Exercise

Discrimination

×D

y

Positive Full Connection

Positive Full Connection

1 0 0 0 0 0 0 0 1 0

Positive Full ConnectionCNN

Exercise

Q-matrix

Partial

Order

Knowledge

RelevancyId

Text

Q-matrix Refining NeuralCDM

10

10

00

01

10

Top-k

QId

0.3

00.9

0.8

0

hs

hdiff

k1 k2 k3 k4 k5

e1 0 1 0 0 0

e2 1 0 1 0 0

e3 0 0 0 1 0

e4 0 0 1 1 0

Figure 4: Extended neural cognitive diagnosis model. Theknowledge relevancy vectors are replaced with vectors inQ-matrix that is refined by leveraging exercise texts.

Bayesian method as follows. For convenience, we definepartial order >+

i as:

a >+i b, if Qia = 1 and Qib = 0 and b ∈ V ki , (11)

and define the partial order relationship set as DV ={(i, a, b)|a >+

i b, i = 1, 2, . . . ,M}. Following traditionalBayesian treatment, we assume Q follows a zero meanGaussian prior with standard deviation σ of each dimen-sion. To give Q-matrix labels higher confidence, we definep(a >+

i b|Qi) with a pairwise logistic-like function:

p(a >+i b|Qi) =

1

1 + e−λ(Qia−Qib). (12)

The parameter λ controls the discrimination of relevancevalues between labeled and unlabeled knowledge concepts.The log posterior distribution over DV on Q is finally for-mulated as:

ln p(Q|DV ) = ln∏

(i,a,b)∈DV

p(a >+i b|Qi)p(Qi)

=M∑i=1

K∑a=1

K∑b=1

I(a >+i b) ln

1

1 + e−λ(Qia−Qib)

+ C −M∑i=1

K∑j=1

Q2ij

2σ2,

(13)

where C is a constant that can be ignored during optimiza-tion. Before using Q in NeuralCDM, we need to restrict itselements to the range (0, 1), and set elements of concepts un-labeled or not predicted to 0. Thus, Sigmoid(Q)◦M is usedto replace Q in NeuralCDM, where M ∈ {0, 1}M×K is amask matrix, and Mij = 1 if j ∈ V ki or Qij = 1;Mij = 0

otherwise. Q is trained together with the cognitive diagnos-tic model, thus the loss function is:

loss = − ln p(Q|DV ) + lossCDM . (14)

3.5 Generality of NeuralCDIn this subsection we show that NeuralCD is a generalframework which can cover many traditional cognitive di-agnostic models. Using Eq. (6) as the first layer, we now

show the close relationship between NeuralCD and tradi-tional models, including MF, IRT and MIRT.MF. Qe and hs can be seen as exercise and student latenttrait vectors respectively in MF. By setting hdiff ≡ 0 andhdisc ≡ 1, the output of the first layer is x = Qe ◦hs. Thenin order to work like MF (i.e., y = Qe · hs), all the rest oflayers need to do is to sum up the values of each entry inx, which is easy to achieve. Monotonicity assumption is notapplied in MF approaches.IRT. Take the typical formation of IRT y = Sigmoid((hs −hdiff ) × hdisc) as example. Set Qe ≡ 1, and let hs andhdiff be unidimensional, the output of the first layer isx = (hs − hdiff ) × hdisc, followed by a Sigmoid ac-tivation function. Monotonicity assumption is achieved bylimiting hdisc to be positive. Other variations of IRT (e.g.,y′ = C + (1− C)y where C is guessing parameter) can berealized with a few changes.MIRT. One direct extension from IRT to MIRT is touse multidimensional latent trait vectors of exercises andstudent. Here we take the typical formation proposedin (Adams, Wilson, and Wang 1997) as example:

y =eQe·hs−de

1 + eQe·hs−de. (15)

Let hdisc ≡ 1, the output of the first layer given byEq. (6) is x = Qe ◦ (hs − hdiff ). By Setting W1 =[1 1 · · · 1] , b1 = 0 and φ(x) = x in Eq. (7), we havef1 = Qe · hs − de (where de = Qe · hdiff ). All therest of the layers need to do is to approximate the functiong(f1) = 1−Sigmoid(f1), which can be easily achieved withtwo more layers. Monotonicity assumption can be realizedif each entry ofQe is restricted to be positive.

3.6 DiscussionWe have introduced the details of NeuralCD framework andshowed special cases of it. It’s necessary to point out that thestudent’s proficiency vector F s and exercise’s knowledgerelevancy vector F kn are basic factors needed in NeuralCDframework. Additional factors such as exercise discrimina-tion can be integrated into if reasonable. The formation of

Page 6: Neural Cognitive Diagnosis for Intelligent Education Systems

the first interactive layer is not limited, but it’s better to con-tain the term F s ◦ F kn to ensure that each dimension ofF s corresponds to a specific knowledge concept. The posi-tive full connection is only one of the strategies that imple-ment monotonicity assumption. More sophisticated networkstructures can be designed as the interaction layers. For ex-ample, recurrent neural network may be used to capture thetime characteristics of the student’s learning process.

4 ExperimentsWe first compare our NeuralCD models with some baselineson the student performance prediction task. Then we makesome interpretation assessments of the models.

4.1 Dataset DescriptionWe use two real-world datasets in the experiments, i.e., Mathand ASSIST. Math dataset supplied by iFLYTEK Co., Ltd.is collected from the widely-used online learning systemZhixue1, which contains mathematical exercises and logsof high school examinations. ASSIST (ASSISTments 2009-2010 ”skill builder”) is an open dataset collected by the AS-SISTments online tutoring systems (Feng, Heffernan, andKoedinger 2009), which only provides student response logsand knowledge concepts2. We choose the public correctedversion that eliminates the duplicated data issue proposedby previous work (Xiong et al. 2016). Table 1 summarizesbasic statistics of the datasets.

We filter out students with less than 30 and 15 responselogs for Math and ASSIST respectively to guarantee thateach student has enough data for diagnosis. Therefore fordataset Math, we got 2,507 exercises with 497 knowledgeconcepts for diagnostic network, and the remaining exer-cises with knowledge concepts not appearing in logs areused for the Q-matrix refining part of NeuralCDM+. We per-form a 80%/20% train/test split of each student’s responselog. As for ASSIST, we divide the response logs in the sameway with Math, but NeuralCDM+ is not evaluated on thisdataset as exercise text is not provided. All models are eval-uated with 5-fold cross validation.

Students’ knowledge proficiencies are stable in Math asthe dataset is composed of logs from examinations. How-ever, a student’s proficiency on a knowledge concept maychange as he will be continually given exercises of that con-cept until meeting certain criterion (e.g., answering 3 rele-vant exercises correctly in a row). To analyze whether staticmodels (e.g., NeuralCD models and static traditional mod-els) are suitable to apply on ASSIST, we compare two met-rics between Math and ASSIST. The first metric is the av-erage amount of logs that each student toke for each knowl-edge concept:

AVG#log =

∑Ni

∑Kj Log(i, j)∑N

i

∑Kj I(Log(i, j) > 0)

, (16)

where Log(i, j) is the amount of exercises student si an-swered that related to knowledge concept kj . Further, an-

1https://www.zhixue.com2https://sites.google.com/site/assistmentsdata/home/assistment-

2009-2010-data/skill-builder-data-2009-2010

Table 1: Dataset summary.Dataset Math ASSIST

#Students 10,268 4,163#Exercises 917,495 17,746#Knowledge concepts 1,488 123#Response logs 864,722 324,572#Knowledge concepts per exercise 1.53 1.19AVG#log 2.28 8.05STD#log>1 0.305 0.316

other metric is the mean standard deviation of scores rij thatLog(i, j) > 1 as:

STD#log>1 = meansi∈S

( meankj∈Kn,

Log(i,j)>1

(stdij)), (17)

where stdij is the standard deviation of scores that student sigot for exercises related to knowledge concept kj . As the re-sults showed in Table 1, although ASSIST has a much largerAVG#log than Math, their STD#log>1 are close. Therefore,it is reasonable to assume that the knowledge states of stu-dents in ASSIST are also stable, and our static NeuralCDmodels and baselines are applicable for both dataset. Therewill be more discussions in Model Interpretation.

4.2 Experimental SetupThe dimensions of the full connection layers (Eq. (7) ∼ (9))are 512, 256, 1 respectively, and Sigmoid is used as acti-vation function for all of the layers. We set hyperparame-ters λ = 0.1 (Eq. (12)) and σ = 1 ( Eq. (13)). For k intop-k knowledge concepts selecting, we use the value thatmake the predicting network reach 0.85 recall. That is, inour experiment, k = 20. We initialize the parameters withXavier initialization (Glorot and Bengio 2010), which fillthe weights with random values sampled from N (0, std2),

where std =√

2nin+nout

. nin is the number of neuronsfeeding into the weights, and nout is the number of neuronsthe results is fed to.

The CNN architecture we use in NeuralCDM+ contains3 convolutional layers followed by a full connection outputlayer. MaxPooling are used after 1st and 3rd convolutionallayers. The channels of convolutional layers are 400, 200,100, and kernel sizes are set to 3, 4, 5 respectively. We adoptReLu activation function for convolution layers and Sigmoidfor the output layer. Multi-label binary cross entropy is usedas loss function for training the CNN.

To evaluate the performance of our NeuralCD models, wecompare them with previous approaches, i.e., DINA, IRT,MIRT and PMF. All models are implemented by PyTorchusing Python, and all experiments are run on a Linux serverwith four 2.0GHz Intel Xeon E5-2620 CPUs and a TeslaK20m GPU.

4.3 Experimental ResultsStudent Performance Prediction The performance of acognitive diagnosis model is difficult to evaluate as we can’t

Page 7: Neural Cognitive Diagnosis for Intelligent Education Systems

Table 2: Experimental results on student performance prediction.Math ASSIST

Model Accuracy RMSE AUC Accuracy RMSE AUC

DINA 0.593±.001 0.487±.001 0.686±.001 0.650±.001 0.467±.001 0.676±.002IRT 0.782±.002 0.387±.001 0.795±.001 0.674±.002 0.464±.002 0.685±.001MIRT 0.793±.001 0.378±.002 0.813±.002 0.701±.002 0.461±.001 0.719±.001PMF 0.763±.001 0.407±.001 0.792±.002 0.661±.002 0.476±.001 0.732±.001NeuralCDM 0.792±.002 0.378±.001 0.820±.001 0.719±.008 0.439±.002 0.749±.001NeuralCDM+ 0.804±.001 0.371±.002 0.835±.002 - - -

NeuralCDM-Qmatrix

DINA Random

NeuralCDM-Monotonicity

NeuralCDM+ NeuralCDM

Math ASSIST0.2

0.4

0.6

0.8

0.93

DOA

Figure 5: DOA results of models. In NeuralCD models,there is clear correspondence between entries in hs andknowledge concepts, thus their diagnosis results have highDOA. Removing Q-matrix or monotonicity assumptionwould reduce the performance.

Exercise 1

Exercise 2

Exercise 3

Number Line

Solving Inequalities

Add Whole Numbers

AbsoluteValue

Ordering Fractions

StudentResponse

1 1 0 0 0

0 0 1 1 0

0 0 0 0 1

0

0.2

0.4

0.6

0.8

1

Kno

wle

dge P

rofic

ienc

y0

0.2

0.4

0.6

0.8

1

Kno

wle

dge D

iffic

ulty

StudentExercise 1Exercise 2Exercise 3

Figure 6: Diagnosis example of a student in ASSIST. Theupper part is the Q-matrix of 3 exercises and correspondingresponse logs. The lower part shows the diagnosed student’sknowledge proficiencies (bars) and knowledge difficultiesof each exercise (points).

obtain the true knowledge proficiency of students. As di-agnostic result is usually acquired through predicting stu-dents’ performance in most works, performance on theseprediction tasks can indirectly evaluate the model fromone aspect (Liu et al. 2018). Considering that all the ex-ercises we used in our data are objective exercises, weuse evaluation metrics from both classification aspect andregression aspect, including accuracy, RMSE (root meansquare error) (Pei et al. 2018) and AUC (area under thecurve) (Bradley 1997).

Table 2 shows the experimental results of all modelson student performance prediction task. The error bars af-ter ’±’ is the standard deviations of 5 evaluation runs foreach model. From the table, we can observe that NeuralCDmodels outperform almost all the other baselines on bothdatasets, indicating the effectiveness of our framework. Inaddition, the better performance of NeuralCDM+ over Neu-ralCDM proves that the Q-matrix refining method is effec-tive, and also demonstrates the importance of fine estimatedknowledge relevancy vectors for cognitive diagnosis.

Model Interpretation To assess the interpretability ofNeuralCD framework (i.e., whether the diagnostic result isreasonable), we further conduct several experiments.

Intuitively, if student a has a better mastery on knowledge

concept k than student b, then a is more likely to answerexercises related to k correctly than b (Chen et al. 2017).We adopt Degree of Agreement (DOA) (Pirotte et al. 2007)as the evaluation metric of this kind of ranking performance.For knowledge concept k, DOA(k) is formulated as:

DOA(k) =1

Z

N∑a=1

N∑b=1

δ(F sak, Fsbk)

M∑j=1

IjkJ(j, a, b) ∧ δ(raj , rbj)

J(j, a, b), (18)

where Z =∑Na=1

∑Nb=1 δ(F

sak, F

sbk). F

sak is the proficiency

of student a on knowledge concept k. δ(x, y) = 1 if x > yand δ(x, y) = 0 otherwise. Ijk = 1 if exercise j containsknowledge concept k and Ijk = 0 otherwise. J(j, a, b) = 1if both student a and b did exercise j and J(j, a, b) = 0otherwise. We average DOA(k) on all knowledge conceptsto evaluate the quality of diagnostic result (i.e., knowledgeproficiency acquired by models).

Among traditional models, we only compare with DINA,since for IRT, MIRT and PMF, there are no clear corre-spondence between their latent features and knowledge con-cepts. Besides, we conduct experiments on two reducedNeuralCDM models. In the first reduced model (denotedas NeuralCDM-Qmatrix), knowledge relevancy vectors areestimated during unsupervised training instead of gettingfrom Q-matrix. While in another reduced model (denoted

Page 8: Neural Cognitive Diagnosis for Intelligent Education Systems

as NeuralCDM-Monotonocity), monotonicity assumption isremoved by eliminating the positive restriction on the fullconnection layers. These two reduced models are used todemonstrate the importance of fine-estimated knowledgerelevancy vector and monotonicity assumption respectively.Furthermore, we conduct an extra experiment in which stu-dents’ knowledge proficiencies are randomly estimated, andcompute the DOA for comparison.

Figure 5 presents the experimental results. From the fig-ure we can observe that DOAs of NeuralCDM and Neural-CDM+ are significantly higher than baselines, which provesthat knowledge proficiencies diagnosed by them are reason-able. The DOAs of NeuralCDM-Qmatrix and NeuralCDM-Monotonicity are much lower than NeuralCDM, which indi-cates that both information from Q-matrix and monotonicityassumption are important for getting interpretable diagno-sis results (knowledge proficiency vectors). DOA of DINAis slightly higher than Random due to the use of Q-matrix.Besides, NeuralCDM performs much better on Math thanon ASSIST. This is mainly due to the contradictions in logs,i.e., a student may answer some exercises containing knowl-edge concept kj correctly while others containing kj wrong(reasons may be the change of knowledge proficiency, orother knowledge concepts contained by the exercises). Asshowed in Table 1, ASSIST has much larger AVG#log andslightly higher STD#log>1 than Math dataset, which makesmore contradictions in logs. Longer logs with more contra-dictions would decrease DOA.

Case Study. Here we present an example of a student’sdiagnostic result of NeuralCDM on dataset ASSIST in Fig-ure 6. The upper part of Figure 6 shows the Q-matrix of threeexercises on five knowledge concepts and the response of astudent to the exercises. The bars in the underneath subfig-ure represent the student’s proficiency on each knowledgeconcept. The lines with different colors and markers rep-resent the knowledge difficulties of the three exercises (forclarity, we only present difficulties of relevant knowledgeconcepts for each exercise). We can observe from the figurethat the student is more likely to response correctly whenhis proficiency satisfies the requirement of the exercise. Forexample, exercise 3 requires the mastery of ’Ordering Frac-tion’ and corresponding difficulty is 0.35. The student’s pro-ficiency on ’Ordering Fraction’ is 0.60, which is higher thanrequired, thus he answered it correctly. Both knowledge dif-ficulty (hdiff ) and knowledge proficiency (hs) in Neural-CDM are explainable as expected.

4.4 Discussion.From the above experiments, we can observe that NeuralCDmodels provide both accurate and interpretable results forcognitive diagnosis.

There still some directions for future studies. First, wemay make our effort to design a more efficient model forknowledge concept prediction, which would promote theperformance of NeuralCDM+. Second, the positive restric-tion on neural network weights may limit the approximateability, thus we would like to explore more flexible meth-ods to satisfy the monotonicity assumption. Third, since

students’ knowledge statuses change in many online self-learning circumstances, we would like to extend NeuralCDfor dynamic cognitive diagnosis.

5 ConclusionIn this paper, we proposed a neural cognitive diagnosticframework, NeuralCD framework, for students’ cognitivediagnosis. Specifically, we first discussed fundamental stu-dent and exercise factors in the framework, and placed amonotonicity assumption on the framework to ensure its in-terpretability. Then, we implemented a specific model Neu-ralCDM under the framework to show its feasibility, andfurther extended NeuralCDM by incorporating exercise textto refine Q-matrix. Extended experimental results on real-world datasets showed the effectiveness of our models withboth accuracy and interpretability. We also showed that Neu-ralCD could be seen as the generalization of some traditionalcognitive diagnostic models (e.g., MIRT). The structure ofthe diagnostic network in our work is designed intuitively.However, with the high flexibility and potential of neuralnetwork, we hope this work could lead to further studies.

6 AcknowledgmentsThis research was partially supported by grants from theNational Natural Science Foundation of China (Grants No.61922073, 61672483, U1605251, 61727809), the ScienceFoundation of Ministry of Education of China & China Mo-bile (No. MCM20170507), and the Iflytek joint researchprogram.

ReferencesAdams, R. J.; Wilson, M.; and Wang, W.-c. 1997. The mul-tidimensional random coefficients multinomial logit model.Applied psychological measurement 21(1):1–23.Anderson, A.; Huttenlocher, D.; Kleinberg, J.; and Leskovec,J. 2014. Engaging with massive online courses. In Proceed-ings of the 23rd international conference on World wide web,687–698. ACM.Bradley, A. P. 1997. The use of the area under the roc curve inthe evaluation of machine learning algorithms. Pattern recog-nition 30(7):1145–1159.Burns, H.; Luckhardt, C. A.; Parlett, J. W.; and Redfield, C. L.2014. Intelligent tutoring systems: Evolutions in design. Psy-chology Press.Chan, W.; Jaitly, N.; Le, Q.; and Vinyals, O. 2016. Lis-ten, attend and spell: A neural network for large vocabularyconversational speech recognition. In 2016 IEEE Interna-tional Conference on Acoustics, Speech and Signal Process-ing (ICASSP), 4960–4964. IEEE.Chen, S., and Joachims, T. 2016. Predicting matchups andpreferences in context. In Proceedings of the 22nd ACMSIGKDD International Conference on Knowledge Discoveryand Data Mining, 775–784. ACM.Chen, Y.; Liu, Q.; Huang, Z.; Wu, L.; Chen, E.; Wu, R.; Su,Y.; and Hu, G. 2017. Tracking knowledge proficiency ofstudents with educational priors. In Proceedings of the 2017

Page 9: Neural Cognitive Diagnosis for Intelligent Education Systems

ACM on Conference on Information and Knowledge Manage-ment, 989–998. ACM.De La Torre, J. 2009. Dina model and parameter estimation:A didactic. Journal of educational and behavioral statistics34(1):115–130.DiBello, L. V.; Roussos, L. A.; and Stout, W. 2006. 31a re-view of cognitively diagnostic assessment and a summary ofpsychometric models. Handbook of statistics 26:979–1030.Embretson, S. E., and Reise, S. P. 2013. Item response theory.Psychology Press.Feng, M.; Heffernan, N.; and Koedinger, K. 2009. Addressingthe assessment challenge with an online system that tutorsas it assesses. User Modeling and User-Adapted Interaction19(3):243–266.Fischer, G. H. 1995. Derivations of the rasch model. In Raschmodels. Springer. 15–38.Glorot, X., and Bengio, Y. 2010. Understanding the difficultyof training deep feedforward neural networks. In Proceedingsof the thirteenth international conference on artificial intelli-gence and statistics, 249–256.Guo, X.; Li, R.; Yu, Q.; and Haake, A. R. 2017. Modelingphysicians’ utterances to explore diagnostic decision-making.In IJCAI, 3700–3706.Hornik, K.; Stinchcombe, M.; and White, H. 1989. Multilayerfeedforward networks are universal approximators. Neuralnetworks 2(5):359–366.Huang, Z.; Liu, Q.; Chen, E.; Zhao, H.; Gao, M.; Wei, S.;Su, Y.; and Hu, G. 2017. Question difficulty prediction forreading problems in standard tests. In AAAI, 1352–1359.Huang, Z.; Yin, Y.; Chen, E.; Xiong, H.; Su, Y.; Hu, G.; et al.2019. Ekt: Exercise-aware knowledge tracing for student per-formance prediction. IEEE Transactions on Knowledge andData Engineering.Koren, Y.; Bell, R.; and Volinsky, C. 2009. Matrix fac-torization techniques for recommender systems. Computer42(8):30–37.Kuh, G. D.; Kinzie, J.; Buckley, J. A.; Bridges, B. K.; andHayek, J. C. 2011. Piecing together the student successpuzzle: research, propositions, and recommendations: ASHEHigher Education Report, volume 116. John Wiley & Sons.Liu, Q.; Wu, R.; Chen, E.; Xu, G.; Su, Y.; Chen, Z.; and Hu,G. 2018. Fuzzy cognitive diagnosis for modelling examineeperformance. ACM Transactions on Intelligent Systems andTechnology (TIST) 9(4):48.Liu, J.; Xu, G.; and Ying, Z. 2012. Data-driven learning of q-matrix. Applied psychological measurement 36(7):548–564.Lord, F. M. 2012. Applications of item response theory topractical testing problems. Routledge.Pei, H.; Yang, B.; Liu, J.; and Dong, L. 2018. Group sparsebayesian learning for active surveillance on epidemic dynam-ics. In Thirty-Second AAAI Conference on Artificial Intelli-gence.Piech, C.; Bassen, J.; Huang, J.; Ganguli, S.; Sahami, M.;Guibas, L. J.; and Sohl-Dickstein, J. 2015. Deep knowl-edge tracing. In Advances in Neural Information ProcessingSystems, 505–513.

Pirotte, A.; Renders, J.-M.; Saerens, M.; et al. 2007. Random-walk computation of similarities between nodes of a graphwith application to collaborative recommendation. IEEETransactions on Knowledge & Data Engineering 19(3):355–369.Reckase, M. D. 2009. Multidimensional item response the-ory models. In Multidimensional Item Response Theory.Springer. 79–112.Samek, W.; Binder, A.; Montavon, G.; Lapuschkin, S.; andMuller, K.-R. 2016. Evaluating the visualization of what adeep neural network has learned. IEEE transactions on neu-ral networks and learning systems 28(11):2660–2673.Song, K.; Ji, M.; Park, S.; and Moon, I.-C. 2019. Hierarchi-cal context enabled recurrent neural network for recommen-dation. arXiv preprint arXiv:1904.12674.Su, Y.; Liu, Q.; Liu, Q.; Huang, Z.; Yin, Y.; Chen, E.; Ding,C.; Wei, S.; and Hu, G. 2018. Exercise-enhanced sequen-tial modeling for student performance prediction. In Thirty-Second AAAI Conference on Artificial Intelligence.Thai-Nghe, N., and Schmidt-Thieme, L. 2015. Multi-relational factorization models for student modeling in intel-ligent tutoring systems. In Knowledge and Systems Engineer-ing (KSE), 2015 Seventh International Conference on, 61–66.IEEE.Thai-Nghe, N.; Drumond, L.; Krohn-Grimberghe, A.; andSchmidt-Thieme, L. 2010. Recommender system for pre-dicting student performance. Procedia Computer Science1(2):2811–2819.Toscher, A., and Jahrer, M. 2010. Collaborative filtering ap-plied to educational data mining. KDD cup.von Davier, M. 2014. The dina model as a constrained gen-eral diagnostic model: Two variants of a model equivalency.British Journal of Mathematical and Statistical Psychology67(1):49–71.Wang, W.; Chen, Z.; and Hu, H. 2019. Hierarchical attentionnetwork for image captioning. In Proceedings of the AAAIConference on Artificial Intelligence, volume 33, 8957–8964.Wu, M.; Mosse, M.; Goodman, N.; and Piech, C. 2019. Zeroshot learning for code education: Rubric sampling with deeplearning inference. In Proceedings of the AAAI Conferenceon Artificial Intelligence, volume 33, 782–790.Xiong, X.; Zhao, S.; Van Inwegen, E. G.; and Beck, J. E.2016. Going deeper with deep knowledge tracing. Interna-tional Educational Data Mining Society.Yin, Y.; Huang, Z.; Chen, E.; Liu, Q.; Zhang, F.; Xie, X.; andHu, G. 2018. Transcribing content from structural imageswith spotlight mechanism. In Proceedings of the 24th ACMSIGKDD International Conference on Knowledge Discovery& Data Mining, 2643–2652. ACM.Zhang, M.; Wang, W.; Liu, X.; Gao, J.; and He, Y. 2018.Navigating with graph representations for fast and scalabledecoding of neural language models. In Advances in NeuralInformation Processing Systems, 6308–6319.Zhang, X.; Zhao, J.; and LeCun, Y. 2015. Character-levelconvolutional networks for text classification. In Advances inneural information processing systems, 649–657.