中文    English

Journal of library and information science in agriculture

   

Performance of Fine-Tuned Large Language Models in Patent Text Mining

LYU Lucheng1,2, ZHOU Jian1,2, SUN Wenjun1,2, ZHAO Yajuan1,2, HAN Tao1,2   

  1. 1. National Science Library, Chinese Academy of Sciences, Beijing 100190
    2. Department of Information Resources Management, School of Economics and Management, University of Chinese Academy of Sciences, Beijing 100190
  • Received:2025-11-25 Online:2026-01-23

Abstract:

[Purpose/Significance] The use of large language models (LLMs) for patent text mining has become a major research topic in recent years. However, existing studies mainly focus on the application of LLMs to specific tasks, and there is a lack of systematic evaluation of the application effects of fine-tuned LLMs across multiple scenarios. To address this problem, this study takes ChatGLM, an open-source LLM that supports local fine-tuning, as an example. We conduct a comparative evaluation of three types of patent text mining tasks-technical term extraction, patent text generation, and automatic patent classification-under a unified experimental framework. The performance of fine-tuned models is compared from six aspects: different training data sizes, different numbers of training epochs, different prompts, different prefix lengths, different datasets, and single-task versus multi-task fine-tuning. [Method/Process] This study was based on an open-source LLM and carried out fine-tuning research for specific patent tasks in order to clarify the impact of different fine-tuning strategies on the performance of LLMs in patent tasks. Considering task adaptability, model size, inference efficiency, and resource consumption, ChatGLM-6B-int4 was selected as the base model, and P-Tuning V2 was adopted as the fine-tuning method. Three categories of patent tasks are included: extraction, generation, and classification. The extraction task is patent keyword extraction. The generation tasks include: 1) innovation point generation; 2) abstract generation based on a given title; 3) rewriting an existing title; 4) rewriting an existing abstract; 5) generating novelty points based on an existing abstract; 6) generating patent advantages based on an existing abstract; and 7) generating patent application scenarios based on an existing abstract. Six experimental comparison dimensions are designed: 1) different training data sizes; 2) different numbers of training epochs; 3) different datasets with the same data size; 4) different prompts under the same task and data; 5) different P-Tuning V2 prefix lengths with the same training data; and 6) single-task fine-tuning versus multi-task fine-tuning. Two type of evaluation metrics were used. For extraction and generation tasks, the BLEU metric based on n-gram string matching was adopted. For classification tasks, accuracy, recall, and F1 score were used. [Results/Conclusions] Based on the fine-tuning results, several conclusions were obtained. First, a larger training data size does not always lead to better performance. Second, the appropriate number of training epochs depends on the data size. Third, under the same data distribution, different data subsets have limited influence on performance. Fourth, under the same task and dataset, different prompts have little impact on model performance. Fifth, the optimal prefix length is closely related to the training data size. Sixth, for a specific task, single-task fine-tuning performs better than multi-task fine-tuning. These conclusions provide reference and guidance for fine-tuning LLMs in practical patent information work.

Key words: patent text mining, large language model (LLM), fine-tuning, patent data analysis, information analysis method

CLC Number: 

  • G306

Fig.1

Research framework"

Table 1

Evaluation sub-tasks and prompt strategies for three types of tasks"

3类任务 子任务 数据 Prompt
抽取 关键词抽取 999条人工智能领域的专利,包含<标题>及<摘要>,每条专利经过专家标注其<关键词>和<创新点>,以及经过专家<改写的标题>和<改写的摘要>,已基于摘要从新颖点,优势和用途3个方面对专利进行总结,分别表示为<从新颖点角度改写的摘要>,<从优势角度改写的摘要>和<从用途角度改写的摘要>

Instruction:请阅读以下专利的标题和摘要。<标题><摘要>。请给出以上专利的关键词表

Response:该专利的关键词表为:<关键词表>

生成 创新点生成

Instruction:请阅读以下专利的标题和摘要。<标题><摘要>。请给出以上专利的创新点

Response:该专利的创新点为:<创新点>

基于标题生成摘要

Instruction:请根据以下专利的标题生成摘要。<标题>

Response:生成的摘要为:<摘要>

改写已有的标题

Instruction:请改写以下专利的标题。<标题>

Response:改写后的标题为:<改写后的标题>

改写已有的摘要

Instruction:请改写以下专利的摘要。<摘要>

Response:改写后的摘要为:<改写后的摘要>

基于摘要生成专利的新颖点

Instruction:请从新颖点角度改写以下专利的摘要。<摘要>

Response:从新颖点角度改写后的摘要为:<从新颖点角度改写的摘要>

基于摘要生成专利优势

Instruction:请从优势角度改写以下专利的摘要。<摘要>

Response:从优势角度改写后的摘要为:<从优势角度改写的摘要>

基于摘要生成专利用途

Instruction:请从用途角度改写以下专利的摘要。<摘要>

Response:从用途角度改写后的摘要为:<从用途角度改写的摘要>

分类 专利分类 80 000条专利数据,按照IPC部类分类A~H各10 000条

Instruction:请阅读以下专利的标题和摘要。<标题><摘要>。

Response:该专利的标签为:<IPC部类>

Fig.2

Single task: Keyword extraction task"

Fig.3

Single task: Keyword extraction task - Results of different training epochs"

Fig.4

Single task: Keyword extraction task - Results of different prefix lengths"

Fig.5

Single task: Keyword extraction task - Results of different datasets"

Fig.6

Keyword extraction task - Comparison between single-task and multi-task"

Fig.7

Single task: Keyword extraction task - Results of different prompts"

Fig.8

Single task: Innovation point generation task"

Fig.9

Single task: Innovation point generation task - Results of different training epochs"

Fig.10

Single task: Innovation point generation task - Results of different prefix lengths"

Fig.11

Innovation point generation task - Comparison between single-task and multi-task"

Fig.12

Changes of different tasks with training epochs in multi-task fine-tuning"

Fig.13

Fine-tuning results in different prefix lengths in multi-task fine-tuning"

Fig.14

Comparison results of classification tasks under different conditions"

[1]
张旭, 雷孝平. 国内外专利情报分析方法、技术与应用研究进展[M]//情报学进展. 北京: 国防工业出版社, 2012, 9: 287-321.
Zhang Xu, Lei Xiaoping. Research progress on patent intelligence analysis methods, technologies and applications at home and abroad[M]//Progress in Information Science. Beijing: National Defense Industry Press, 2012, 9: 287-321.
[2]
吕璐成, 罗文馨, 许景龙, 等. 专利情报方法、工具、应用研究进展及新技术应用趋势[J]. 情报学进展, 2020, 13: 235-278.
Lucheng Lyu, Luo Wenxin, Xu Jinglong, et al. Research progress of patent intelligence methods, tools, applications and application trends of new technologies[J]. Progress in Information Science, 2020, 13: 235-278.
[3]
胡雅敏, 吴晓燕, 陈方. 基于机器学习的技术术语识别研究综述[J]. 数据分析与知识发现, 2022, 6(2): 7-17.
Hu Yamin, Wu Xiaoyan, Chen Fang. Review of technology term recognition studies based on machine learning[J]. Data Analysis and Knowledge Discovery, 2022, 6(2): 7-17.
[4]
吕璐成, 韩涛, 周健, 等. 基于深度学习的中文专利自动分类方法研究[J]. 图书情报工作, 2020, 64(10): 75-85.
Lucheng Lyu, Han Tao, Zhou Jian, et al. Research on the method of Chinese patent automatic classification based on deep learning[J]. Library and Information Service, 2020, 64(10): 75-85.
[5]
马俊, 吕璐成, 赵亚娟, 等. 基于预训练语言模型的中文专利自动分类研究[J]. 中华医学图书情报杂志, 2022, 31(11): 20-28.
Ma Jun, Lucheng Lyu, Zhao Yajuan, et al. Research on automatic classification of Chinese patents based on pre-trained language models[J]. Chinese Journal of Medical Library and Information Science, 2022, 31(11): 20-28.
[6]
孙蒙鸽, 韩涛, 王燕鹏, 等. GPT技术变革对基础科学研究的影响分析[J]. 中国科学院院刊, 2023, 38(8): 1212-1224.
Sun Mengge, Han Tao, Wang Yanpeng, et al. Impact analysis of GPT technology revolution on fundamental scientific research[J]. Bulletin of Chinese Academy Of Sciences, 2023, 38(8): 1212-1224.
[7]
张智雄, 于改红, 刘熠, 等. ChatGPT对文献情报工作的影响[J]. 数据分析与知识发现, 2023, 7(3): 36-42.
Zhang Zhixiong, Yu Gaihong, Liu Yi, et al. The influence of chat GPT on library & information services[J]. Data Analysis and Knowledge Discovery, 2023, 7(3): 36-42.
[8]
张铭洁, 赵瑞雪. ChatGPT驱动的智慧图书馆情感感知与服务优化[J]. 农业图书情报学报, 2024, 36(12): 74-88.
Zhang Mingjie, Zhao Ruixue. Emotion perception and serviceoptimization in ChatGPT-driven smart libraries[J]. Journal of Library and Information Science in Agriculture, 2024, 36(12): 74-88.
[9]
Bai Z L, Zhang R J, Chen L Q, et al. PatentGPT: A large language model for intellectual property[PP/OL]. V5. arXiv (2024-06-05)[2025-09-15].
[10]
Ren R T, Ma J, Luo J X. Large language model for patent concept generation[J]. Advanced Engineering Informatics, 2025, 65: 103301.
[11]
Lee J S. InstructPatentGPT: Training patent language models to follow instructions with human feedback[J]. Artificial Intelligence and Law, 2025, 33(3): 739-782.
[12]
Liu C, Fomin N I, Xiao S T, et al. How ChatGPT is shaping next-generation patent solutions[J]. Buildings, 2025, 15(13): 2273.
[13]
Pelaez S, Verma G, Ribeiro B, et al. Large-scale text analysis using generative language models: A case study in discovering public value expressions in AI patents[J]. Quantitative Science Studies, 2024, 5(1): 153-169.
[14]
Kosonocky C W, Wilke C O, Marcotte E M, et al. Mining patents with large language models elucidates the chemical function landscape[J]. Digital Discovery, 2024, 3(6): 1150-1159.
[15]
Zhang W, Wang Q G, Kong X T, et al. Fine-tuning large language models for chemical text mining[J]. Chemical Science, 2024, 15(27): 10600-10611.
[16]
Liu X, Ji K X, Fu Y C, et al. P-tuning: Prompt tuning can be comparable to fine-tuning across scales and tasks[C]//Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). Stroudsburg, PA, USA: ACL, 2022: 61-68.
[17]
Hu E J, Shen Y L, Wallis P, et al. LoRA: Low-rank adaptation of large language models[PP/OL]. V2. arXiv (2021-10-16)[2025-09-15].
[18]
李鑫鑫, 马雨萌, 鞠孜涵, 王敬. 基于大语言模型的科技政策评论方面级情感分析研究——以新能源汽车产业为例[J]. 农业图书情报学报, 2025, 37(10): 53-66.
Li Xinxin, Ma Yumeng, Ju Zihan, et al. Aspect-level sentiment analysis of science and technology policy reviews based on large language models: A case study of the new energy vehicle industry[J]. Journal of Library and Information Science in Agriculture, 2025, 37(10): 53-66.
[19]
赵建飞, 陈挺, 王小梅, 等. 基于大语言模型知识自蒸馏的无标注专利关键信息抽取[J]. 数据分析与知识发现, 2024, 8(8): 133-143.
Zhao Jianfei, Chen Ting, Wang Xiaomei, et al. Extracting key information from unlabeled patents based on knowledge self-distillation of large language model[J]. Data Analysis and Knowledge Discovery, 2024, 8(8): 133-143.
[20]
时宗彬, 朱丽雅, 乐小虬. 基于本地大语言模型和提示工程的材料信息抽取方法研究[J]. 数据分析与知识发现, 2024, 8(7): 23-31.
Shi Zongbin, Zhu Liya, Le Xiaohong. Material information extraction based on local large language model and prompt engineering[J]. Data Analysis and Knowledge Discovery, 2024, 8(7): 23-31.
[21]
Prottasha N J, Mahmud A, Sobuj M S I, et al. Parameter-efficient fine-tuning of large language models using semantic knowledge tuning[J]. Scientific Reports, 2024, 14: 30667.
[1] JIANG Peng, REN Yan, ZHU Beiling. Exploration and Practice of Classification Indexing Combined with Large Language Models [J]. Journal of library and information science in agriculture, 2024, 36(5): 32-42.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!