农业图书情报学报 ›› 2026, Vol. 38 ›› Issue (3): 76-87.doi: 10.13998/j.cnki.issn1002-1248.25-0594

• 研究论文 • 上一篇    下一篇

面向学术评价的成果数据分析智能体构建研究

邓启平1, 柯佳秀1, 甘鹏2, 周松2   

  1. 1. 电子科技大学 图书馆,成都 611731
    2. 重庆维普智图数据科技有限公司,重庆 401123
  • 收稿日期:2025-11-02 出版日期:2026-03-05 发布日期:2026-03-30
  • 作者简介:

    邓启平(1990- ),男,硕士,馆员,研究方向为情报分析方法与技术、科学计量与评价

    柯佳秀(1991- ),女,硕士,馆员,研究方向为学科情报分析、科学计量与评价

    甘鹏(1997- ),男,学士,大数据开发工程师,研究方向为大数据分析

    周松(1982- ),男,学士,助理工程师,研究方向为学术数据管理、学科情报服务

  • 基金资助:
    四川省社会科学重点研究基地——四川学术成果分析与应用研究中心资助项目“多源异构数据视角下的图书馆智能化精准科研评价服务研究”(SCAA25-B02)

Construction of an Intelligent Agent for Academic Output Data Analysis Oriented to Academic Evaluation

DENG Qiping1, KE Jiaxiu1, GAN Peng2, ZHOU Song2   

  1. 1. Library, University of Electronic Science and Technology of China, Chengdu 611731
    2. Chongqing VIPSMART Data Technology Co. , Ltd. , Chongqing 401123
  • Received:2025-11-02 Online:2026-03-05 Published:2026-03-30

摘要:

[目的/意义] 利用大模型开展智能化成果数据分析,有助于高校图书馆提升学术评价服务的效能。 [方法/过程] 基于构建的学术评价智能体理论框架,结合图书馆的学术评价服务实践,从评价的对象、指标等维度梳理成果数据需求,并制定元数据方案融合题录数据、标引数据及评价数据,通过集成大模型和ReAct推理框架,构建本地化成果数据分析智能体,以自然语言作为输入,利用Text2SQL技术将用户意图转为PostgreSQL查询语句,结合多轮自我修正机制获取准确数据,最终以可视化图表和结构化报告输出分析结果。 [结果/结论] 以3万余篇结构化的学术论文作为测试数据,并构建多维度的问答测试集进行效果验证和对比分析,结果显示智能体准确回答了全部20个提问,表明其能有效处理多维复杂的学术评价任务,且在回答的准确率和稳定性方面优于现有通用数据分析智能体。

关键词: 大模型, 智能体, 成果数据, 学术评价

Abstract:

[Purpose/Significance] University libraries require efficient, data-driven academic evaluation to support management decisions. Traditional manual methods are slow, subjective, and untimely. While large language models (LLMs) offer automation potential, existing applications in this domain are limited, often focusing on auxiliary tasks and raising data security concerns with cloud-based processing. This study addresses these gaps by proposing a localized, intelligent agent for secure and interactive analysis of academic output. [Method/Process] A four-layer theoretical framework based on the DIKW model was established to guide the agent's design from data integration to wisdom generation. Grounded on the practical experience of academic evaluation services in libraries, this study systematically identified data requirements from dimensions of academic evaluation objects (institution, school, discipline, and researcher) and metrics (output, collaboration, impact, and quality), and formulated a metadata scheme to integrate bibliographic data, indexing data and evaluation data into a single structured table for research papers. A localized agent was implemented using open-source tools: Chainlit for the conversational interface, LangChain with the Kimi-K2-0905-Preview LLM as the core, and the ReAct framework to enable an iterative "Thought-Action-Observation" loop for complex reasoning and self-correction. The agent employs Text-to-SQL technology to translate natural language queries into executable PostgreSQL statements. Comprehensive prompt engineering was conducted to guide the LLM in accurate SQL generation, handling challenges such as data deduplication, multi-value fields, and entity disambiguation. This enables dynamic intent interpretation, multi-step data retrieval and validation, and output generation combining visualizations and structured reports. [Results/Conclusions] The agent was evaluated using a test dataset of over 30 000 structured academic papers and a multi-dimensional set of 20 test queries covering various evaluation scenarios and complex composite questions. The agent achieved a 100% final accuracy rate. The initial query accuracy was 85%, with errors primarily related to recognizing informal entity names (e.g., abbreviations). All errors were autonomously corrected within one ReAct iteration, demonstrating effective self-repair. Comparative analysis against two general-purpose data analysis agents showed the proposed agent's superior accuracy and stability, particularly in handling entity disambiguation and complex multi-turn tasks. The study confirms that the locally-deployed intelligent agent provides an effective, secure, and interactive solution for academic output analysis, successfully bridging natural language queries with precise data retrieval. Limitations include the evaluation's primary focus on data retrieval accuracy rather than narrative quality, and a test scope limited to core academic evaluation queries. Future work will expand the agent's capabilities to support diverse research outputs (e.g., patents and monographs), enhance visualization integration, and enable customizable report template generation.

Key words: large language model, intelligent agent, academic output data, academic evaluation

中图分类号:  G353

引用本文

邓启平, 柯佳秀, 甘鹏, 周松. 面向学术评价的成果数据分析智能体构建研究[J]. 农业图书情报学报, 2026, 38(3): 76-87.

DENG Qiping, KE Jiaxiu, GAN Peng, ZHOU Song. Construction of an Intelligent Agent for Academic Output Data Analysis Oriented to Academic Evaluation[J]. Journal of library and information science in agriculture, 2026, 38(3): 76-87.