CL$^2$GEC: A Multi-Discipline Benchmark for Continual Learning in Chinese Literature Grammatical Error Correction

ACL ARR 2026 January Submission8928 Authors

06 Jan 2026 (modified: 20 Mar 2026)ACL ARR 2026 January SubmissionEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Chinese Grammatical Error Correction;Benchmark Evaluation;Continual Learning;Large Language Models
Abstract: The growing demand for automated writing assistance in scientific domains highlights the need for robust Chinese Grammatical Error Correction (CGEC) systems that can adapt across disciplines. However, existing CGEC research lacks dedicated benchmarks for academic writing and overlooks continual learning as a solution to handle domain-specific variation. To fill this gap, we introduce CL2 GEC, a Continual Learning benchmark for Chinese Literature Grammatical Error Correction, designed to evaluate adaptive CGEC across multiple academic fields. Our benchmark includes 10,000 human-annotated sentences spanning 10 disciplines, each exhibiting distinct linguistic styles and error patterns. We evaluate large language models under sequential tuning, parameter-efficient adaptation, and representative continual learning strategies, using both standard GEC metrics and continual learning metrics adapted to task-level variation. Experimental results show that regularization-based continual learning methods, such as OGD and GEM, outperform replay-based and sequential approaches in both grammatical accuracy and knowledge retention. These findings underscore the feasibility and importance of integrating continual learning into CGEC and position our benchmark as a foundation for future research on adaptive scientific writing assistance.
Paper Type: Long
Research Area: Resources and Evaluation
Research Area Keywords: Grammatical Error Correction, Continual Learning for NLP, Educational applications, Benchmarking
Contribution Types: Approaches to low-resource settings, Data resources, Data analysis
Languages Studied: Chinese
Submission Number: 8928
Loading