Conference Agenda (All times are shown in Eastern Daylight Time)

Overview and details of the sessions of this conference. Please select a date or location to show only sessions at that day or location. Please select a single session for detailed view (with abstracts and downloads if available).

 
Only Sessions at Location/Venue 
Only Sessions at Date / Time 
 
 
Session Overview
Session
Virtual Paper Session 9: Science, AI, and scholarly publishing
Time:
Friday, 12/Dec/2025:
12:00pm - 1:30pm

Virtual location: Virtual


Show help for 'Increase or decrease the abstract text size'
Presentations
12:00pm - 12:30pm

Knowledge breadth and depth measurement of Large Language Models (LLMs)

X. Peng, Q. Lu, K. Liu

Wuhan University, China

This study constructed an integrated framework for measuring LLMs’ knowledge breadth (able to cover multiple fields) and depth (able to handle complex problems) and perform calculations based on evidence. The current research evidence on measuring LLMs was obtained through a systematic review. Then, an integrated framework based on the matching of capability and task was constructed, where the generation capability and knowledge management task reflect the knowledge breadth, while other capabilities and tasks reflect the knowledge depth. Knowledge breadth and depth were measured through Coverage and Revealed Technological Advantage. The results showed that 36 matches of capability and task were selected from representative papers, and the knowledge breadth and depth panorama of LLMs was displayed. Furthermore, there is no absolute advantage or disadvantage in LLMs. This study clarifies the boundaries of the LLM’s measurement, and the framework further ensures the diversity of benchmark models and avoids redundancy and misalignment biases.



12:30pm - 1:00pm

Chinese Large Language Models Evaluation in the Field of Scientific and Technical Information

L. Xiaosong, L. Zenghua, Z. Keran, L. Yifei, T. Shanhong, G. Qiang, Z. Yingxiao, G. Guotong

Center for Information Research of Academy of Military Science, People's Republic of China

The field of scientific and technical information (STI) is characterized by limited open-source training data, strong timeliness requirements, and high demands for professional expertise, necessitating the combination of general capabilities and domain-specific capabilities of large language models (LLMs). Based on common STI research tasks, this study establishes benchmark datasets for evaluating LLMs in the STI field, comprising basic knowledge ability, dynamic research ability, and thematic research ability. A total of 1,557 objective and subjective questions were selected to evaluate the performance of eight LLMs developed by commercial organizations, research institutions, and universities in the STI field. The results indicate that LLMs perform well in terms of STI domain knowledge but still exhibit significant gaps in dynamic and thematic information research. There is a need to actively explore and promote the integration, adaptation, and application of LLM technologies in the STI field to provide robust support for high-quality and high-efficiency STI services.



 
Contact and Legal Notice · Contact Address:
Privacy Statement · Conference: ASIS&T 2025
Conference Software: ConfTool Pro 2.6.154+TC
© 2001–2025 by Dr. H. Weinreich, Hamburg, Germany