注释
计算机科学
过程(计算)
情态动词
情报检索
模式
数据挖掘
数据提取
人工智能
程序设计语言
化学
高分子化学
社会科学
梅德林
社会学
政治学
法学
作者
Yi Tang,Chia-Ming Chang,Xi Yang
标识
DOI:10.1145/3640543.3645174
摘要
The document contains substantial unannotated data, necessitating extensive manual labeling efforts. To address this issue, we introduce PDFChatAnnotator, a human-LLM collaborative tool to collect multi-modal data from PDF catalogs. Initially, PDFChatAnnotator automatically employs our proposed multi-modal binding rules to link related data from different modalities and harnesses the information extraction capabilities of large language models (LLMs) to extract specific information from text descriptions. Furthermore, the tool empowers users to guide and refine the LLM's annotations. During the annotation process, users can influence the LLM through multiple rounds of communication and example establishment via the provided interfaces. To assess the effectiveness of PDFChatAnnotator's techniques, we conducted a technical evaluation using three catalogs with typical layouts as experimental data. The results showed that all accuracy rates for multi-modal binding exceeded 90%, and both the proposed "example establishment" and "interactive adjustment of requirements" contributed to enhanced accuracy rates.
科研通智能强力驱动
Strongly Powered by AbleSci AI