注释
杠杆(统计)
计算机科学
领域(数学分析)
自然语言处理
可靠性(半导体)
人工智能
情报检索
数据科学
数学分析
功率(物理)
物理
数学
量子力学
作者
Xinru Wang,Hannah Kim,Sajjadur Rahman,Kushan Mitra,Zhengjie Miao
标识
DOI:10.1145/3613904.3641960
摘要
Large language models (LLMs) have shown remarkable performance across various natural language processing (NLP) tasks, indicating their significant potential as data annotators. Although LLM-generated annotations are more cost-effective and efficient to obtain, they are often erroneous for complex or domain-specific tasks and may introduce bias when compared to human annotations. Therefore, instead of completely replacing human annotators with LLMs, we need to leverage the strengths of both LLMs and humans to ensure the accuracy and reliability of annotations. This paper presents a multi-step human-LLM collaborative approach where (1) LLMs generate labels and provide explanations, (2) a verifier assesses the quality of LLM-generated labels, and (3) human annotators re-annotate a subset of labels with lower verification scores. To facilitate human-LLM collaboration, we make use of LLM's ability to rationalize its decisions. LLM-generated explanations can provide additional information to the verifier model as well as help humans better understand LLM labels. We demonstrate that our verifier is able to identify potentially incorrect LLM labels for human re-annotation. Furthermore, we investigate the impact of presenting LLM labels and explanations on human re-annotation through crowdsourced studies.
科研通智能强力驱动
Strongly Powered by AbleSci AI