Journal of Information Resources Management ›› 2024, Vol. 14 ›› Issue (6): 143-155.doi: 10.13365/j.jirm.2024.06.143

Previous Articles     Next Articles

Construction of Chinese Classical-Modern Translation Model Based on Pre-trained Language Model

Wu Mengcheng1,2,3 Liu Chang1,2,3 Meng Kai4 Wang Dongbo1,2,3   

  1. 1.College of Information Management of Nanjing Agricultural University,Nanjing,210095; 
    2.Research Center for Humanities and Social computing of Nanjing Agricultural University,Nanjing,210095; 
    3.Research Center for Correlation of Domain Knowledge of Nanjing Agricultural University,Nanjing,210095; 
    4.School of Marxism of Nanjing Agricultural University, Nanjing,210095
  • Online:2024-11-26 Published:2024-12-20
  • About author:Wu Mengcheng, Ph.D. candidate, research direction: digital humanities and machine translation; Liu Chang, Ph.D. candidate, research direction: digital humanities and large language model; Meng Kai, master's supervisor and associate professor, research direction: ancient Chinese philosophy and digital humanities; Wang Dongbo(corresponding author), doctoral supervisor and professor, research direction: intelligent information processing in ancient books,Email:db.wang@njau.edu.cn.
  • Supported by:
    This paper was supported by the Major Project of the National Social Science Fund of China, titled "Construction and Application Research of Cross-language Knowledge Base of Chinese Ancient Classics"(21&ZD331).

Abstract: This study aims to construct and validate a Chinese ancient-modern translation model based on pre-trained language models, providing strong technical support for the research of ancient Chinese and the inheritance and dissemination of cultural heritage. The study selected a total of 300,000 pairs of meticulously processed parallel corpora from the "Twenty-Four Histories" as the experimental dataset and developed a new translation model—Siku-Trans. This model innovatively combines Siku-RoBERTa(as the encoder) and Siku-GPT(as the decoder), designed specifically for translating ancient Chinese, to build an efficient encoder-decoder architecture. To comprehensively evaluate the performance of the Siku-Trans model, the study introduced three models as control groups: OpenNMT, SikuGPT, and SikuBERT_UNILM. Through comparative analysis of the performance of each model in ancient Chinese translation tasks, we found that Siku-Trans exhibits significant advantages in terms of translation accuracy and fluency. These results not only highlight the effectiveness of combining Siku-RoBERTa with Siku-GPT as a training strategy but also provide important references and insights for in-depth research and practical applications in the field of ancient Chinese translation.

Key words: Language model, Machine translation, Ancient Chinese translation, Siku-RoBERTa, Siku-GPT, Siku-Trans

CLC Number: