论文标题
BI-CLKT:基于学习的基于学习的知识跟踪
Bi-CLKT: Bi-Graph Contrastive Learning based Knowledge Tracing
论文作者
论文摘要
知识追踪(KT)的目的是估计学生如何根据他们对相关练习的历史学习来掌握一个概念。知识追踪的好处是,可以更好地组织和调整学生的学习计划,并在必要时进行干预。随着深度学习的最新兴起,深度知识追踪(DKT)利用了经常性的神经网络(RNN)来完成这项任务。其他作品试图引入图形神经网络(GNN)并相应地重新定义任务以实现重大改进。但是,这些努力至少遭受了以下缺点之一:1)他们过分关注节点的细节,而不是对高级语义信息; 2)他们努力建立节点的空间关联和复杂结构; 3)它们仅代表概念或练习,而无需整合它们。受到自我监督学习的最新进展的启发,我们提出了一个基于Biaph的基于学习的知识跟踪(BI-CLKT)来解决这些局限性。具体而言,我们根据“锻炼到运动”(E2E)关系子图设计了两层对比学习方案。它涉及子图的节点级对比度学习,以获得练习的歧视性表示,以及图形对比度学习以获得概念的歧视性表示。此外,我们设计了一个联合对比损失,以获得更好的表示形式,从而更好地预测性能。此外,我们使用RNN和记忆增强的神经网络作为比较的预测层,探索了两个不同的变体,以分别获得练习和概念的更好表示。在四个现实世界数据集上进行的广泛实验表明,所提出的BI-CLKT及其变体的表现优于其他基线模型。
The goal of Knowledge Tracing (KT) is to estimate how well students have mastered a concept based on their historical learning of related exercises. The benefit of knowledge tracing is that students' learning plans can be better organised and adjusted, and interventions can be made when necessary. With the recent rise of deep learning, Deep Knowledge Tracing (DKT) has utilised Recurrent Neural Networks (RNNs) to accomplish this task with some success. Other works have attempted to introduce Graph Neural Networks (GNNs) and redefine the task accordingly to achieve significant improvements. However, these efforts suffer from at least one of the following drawbacks: 1) they pay too much attention to details of the nodes rather than to high-level semantic information; 2) they struggle to effectively establish spatial associations and complex structures of the nodes; and 3) they represent either concepts or exercises only, without integrating them. Inspired by recent advances in self-supervised learning, we propose a Bi-Graph Contrastive Learning based Knowledge Tracing (Bi-CLKT) to address these limitations. Specifically, we design a two-layer contrastive learning scheme based on an "exercise-to-exercise" (E2E) relational subgraph. It involves node-level contrastive learning of subgraphs to obtain discriminative representations of exercises, and graph-level contrastive learning to obtain discriminative representations of concepts. Moreover, we designed a joint contrastive loss to obtain better representations and hence better prediction performance. Also, we explored two different variants, using RNN and memory-augmented neural networks as the prediction layer for comparison to obtain better representations of exercises and concepts respectively. Extensive experiments on four real-world datasets show that the proposed Bi-CLKT and its variants outperform other baseline models.