农业图书情报学报

• •    

微调大模型在专利文本挖掘中的应用效果研究

吕璐成1,2, 周健1,2, 孙文君1,2, 赵亚娟1,2, 韩涛1,2   

  1. 1. 中国科学院文献情报中心,北京 100190
    2. 中国科学院大学 经济与管理学院信息资源管理系,北京 100190
  • 收稿日期:2025-11-25 出版日期:2026-01-23
  • 作者简介:

    吕璐成,男,博士,副研究员,硕士生导师,研究方向为专利情报分析、技术挖掘

    周健,男,博士,研究方向为专利文本挖掘

    孙文君,女,博士生,研究方向为专利情报分析

    赵亚娟,女,博士,博士生导师,研究员,研究方向为知识产权情报

    韩涛,男,博士,研究员,硕士生导师,研究方向为智能情报

  • 基金资助:
    国家自然科学基金青年科学基金项目“技术距离视角下的技术融合模式、特征及预测研究”(72304268)

Performance of Fine-Tuned Large Language Models in Patent Text Mining

LYU Lucheng1,2, ZHOU Jian1,2, SUN Wenjun1,2, ZHAO Yajuan1,2, HAN Tao1,2   

  1. 1. National Science Library, Chinese Academy of Sciences, Beijing 100190
    2. Department of Information Resources Management, School of Economics and Management, University of Chinese Academy of Sciences, Beijing 100190
  • Received:2025-11-25 Online:2026-01-23

摘要:

[目的/意义] 利用大模型开展专利文本挖掘是近年来专利情报分析的热点研究方向,但现有研究多集中于特定任务的大模型应用研究,缺乏对微调大模型在多场景中的应用效果系统评估。 [方法/过程] 以支持本地微调的开源大模型ChatGLM为例,围绕技术术语抽取、专利文本生成和专利自动分类3类专利文本挖掘任务,在统一实验框架下从不同数据量、不同提示词(Prompt)、不同训练轮数、不同前缀长度、不同数据集、单一任务及多任务微调6个角度对比评估应用效果。 [结果/结论] 在专利文本挖掘工作中微调大模型时并不简单地遵循“大力出奇迹”的规律,而是需要结合任务的特点进行训练数据量、训练轮次、前缀长度等参数的研究和最优参数的确定,相关结论为实际专利情报工作中开展大模型微调提供参考启示。

关键词: 专利文本挖掘, 大语言模型, 微调, 专利数据分析, 情报分析方法

Abstract:

[Purpose/Significance] The use of large language models (LLMs) for patent text mining has become a major research topic in recent years. However, existing studies mainly focus on the application of LLMs to specific tasks, and there is a lack of systematic evaluation of the application effects of fine-tuned LLMs across multiple scenarios. To address this problem, this study takes ChatGLM, an open-source LLM that supports local fine-tuning, as an example. We conduct a comparative evaluation of three types of patent text mining tasks-technical term extraction, patent text generation, and automatic patent classification-under a unified experimental framework. The performance of fine-tuned models is compared from six aspects: different training data sizes, different numbers of training epochs, different prompts, different prefix lengths, different datasets, and single-task versus multi-task fine-tuning. [Method/Process] This study was based on an open-source LLM and carried out fine-tuning research for specific patent tasks in order to clarify the impact of different fine-tuning strategies on the performance of LLMs in patent tasks. Considering task adaptability, model size, inference efficiency, and resource consumption, ChatGLM-6B-int4 was selected as the base model, and P-Tuning V2 was adopted as the fine-tuning method. Three categories of patent tasks are included: extraction, generation, and classification. The extraction task is patent keyword extraction. The generation tasks include: 1) innovation point generation; 2) abstract generation based on a given title; 3) rewriting an existing title; 4) rewriting an existing abstract; 5) generating novelty points based on an existing abstract; 6) generating patent advantages based on an existing abstract; and 7) generating patent application scenarios based on an existing abstract. Six experimental comparison dimensions are designed: 1) different training data sizes; 2) different numbers of training epochs; 3) different datasets with the same data size; 4) different prompts under the same task and data; 5) different P-Tuning V2 prefix lengths with the same training data; and 6) single-task fine-tuning versus multi-task fine-tuning. Two type of evaluation metrics were used. For extraction and generation tasks, the BLEU metric based on n-gram string matching was adopted. For classification tasks, accuracy, recall, and F1 score were used. [Results/Conclusions] Based on the fine-tuning results, several conclusions were obtained. First, a larger training data size does not always lead to better performance. Second, the appropriate number of training epochs depends on the data size. Third, under the same data distribution, different data subsets have limited influence on performance. Fourth, under the same task and dataset, different prompts have little impact on model performance. Fifth, the optimal prefix length is closely related to the training data size. Sixth, for a specific task, single-task fine-tuning performs better than multi-task fine-tuning. These conclusions provide reference and guidance for fine-tuning LLMs in practical patent information work.

Key words: patent text mining, large language model (LLM), fine-tuning, patent data analysis, information analysis method

中图分类号:  G306

引用本文

吕璐成, 周健, 孙文君, 赵亚娟, 韩涛. 微调大模型在专利文本挖掘中的应用效果研究[J/OL]. 农业图书情报学报. https://doi.org/10.13998/j.cnki.issn1002-1248.25-0672.

LYU Lucheng, ZHOU Jian, SUN Wenjun, ZHAO Yajuan, HAN Tao. Performance of Fine-Tuned Large Language Models in Patent Text Mining[J/OL]. Journal of library and information science in agriculture. https://doi.org/10.13998/j.cnki.issn1002-1248.25-0672.