Chinese Journal of Intelligent Science and Technology ›› 2022, Vol. 4 ›› Issue (1): 97-108.doi: 10.11959/j.issn.2096-6652.202212

• Special Topic: Crowd Intelligence • Previous Articles     Next Articles

Exploration of the continual learning ability that supports the application ecological evolution of the large-scale pretraining Peng Cheng series open source models

Yue YU1,2, Xin LIU1, Fangqing JIANG1, Han ZHANG1, Hui WANG1, Wei ZENG3   

  1. 1 Open Source Institution, Network Intelligence Department, Peng Cheng Laboratory, Shenzhen 518055, China
    2 National University of Defense Technology, Changsha 410073, China
    3 Peking University, Beijing 100091, China
  • Revised:2022-01-13 Online:2022-03-15 Published:2022-03-01
  • Supported by:
    Research on Chinese Technological Open Source Strategy Under New Situation(GHZX2020ZCQ013)

Abstract:

Large-scale pre-training models have achieved great success in the field of natural language processing by using large-scale corpora and pre-training tasks.With the gradual development of large models, the continual learning ability of large models has become a new research focus.The continual learning technology of the Peng Cheng series large models, the exploration of practice and the still facing challenges were mainly introduced, including the Peng Cheng series continual learning technology through task expansion, data increment and knowledge reasoning, Peng Cheng PANGU multi-task continual learning and the practical exploration of the continual learning ability of the Peng Cheng TONGYAN open source large model, the vocabulary update, semantic mapping and knowledge conflicts that the large model faces in the process of continual learning.

Key words: Peng Cheng series large model, continual learning, Peng Cheng PANGU, Peng Cheng TONGYAN, open source large model

CLC Number: 

No Suggested Reading articles found!