Abstract: Knowledge tracing (KT) refers to the problem of predicting a learner’s future performance based on their past performance in education. Recently, attention-based sequence modeling methods achieve impressive predictive performance. However, existing solutions merely consider one single sequence modeling method, which might fail to capture the comprehensive state of knowledge across long sequences. In this paper, we propose Dual Sequence Modeling for Knowledge Tracing (DSMKT). DSMKT aims to enhance the modeling of a learner’s long-term profile by collaborating two sequence modeling methods, i.e., the masked self-attention mechanism and the gated recurrent unit. To further exploit the synergy between two sequence models, we adopt the idea of online knowledge distillation and adaptively combine two branches to form a stronger teacher model, which in turn provides predictions as extra supervision for better modeling ability. Extensive experiments on four real-world benchmark datasets show that DSMKT performs excellently in predicting future learner responses.
External IDs:dblp:journals/dase/NingGLTZZZ25
Loading