Review of Enhancement Research for Closed-Source Large Language Model

With the rapid development of large language models in the field of natural language processing, performance enhancement of closed-source large language models represented by the GPT family has become a challenge. Due to the inaccessibility of parameter weights inside the models, traditional trainin...

Full description

Saved in:
Bibliographic Details
Main Author: LIU Hualing, ZHANG Zilong, PENG Hongshuai
Format: Article
Language:zho
Published: Journal of Computer Engineering and Applications Beijing Co., Ltd., Science Press 2025-05-01
Series:Jisuanji kexue yu tansuo
Subjects:
Online Access:http://fcst.ceaj.org/fileup/1673-9418/PDF/2407021.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:With the rapid development of large language models in the field of natural language processing, performance enhancement of closed-source large language models represented by the GPT family has become a challenge. Due to the inaccessibility of parameter weights inside the models, traditional training methods, such as fine-tuning techniques, are difficult to be applied to closed-source large language models, which makes it difficult for further optimization on these models. Meanwhile, closed-source large language models have been widely used in downstream real-world tasks, and thus it is important to investigate how to enhance the performance of closed-source large language models. This paper focuses on the enhancement of closed-source large language models, analyzes three techniques, namely prompt engineering, retrieval augmented generation, and agent, and further subdivides them according to the technical characteristics and modular architectures of the different methods. The core idea, main method and application effect of each technology are introduced in detail, and the superiority and limitation of different augmentation methods in terms of reasoning ability, generation credibility and task adaptability are studied. In addition, this paper also discusses the combined application of these three techniques, combining with specific cases to emphasize the great potential of the combined techniques in enhancing model performance. Finally, this paper summarizes the research status and problems of the existing techniques, and looks forward to the future development of enhancement techniques for closed-source large language models.
ISSN:1673-9418