脆弱性(计算)
计算机科学
软件
计算机安全
安全编码
软件安全保证
集合(抽象数据类型)
软件工程
信息安全
程序设计语言
保安服务
作者
Moumita Das Purba,Arpita Ghosh,Benjamin J. Radford,Bill Chu
标识
DOI:10.1109/issrew60843.2023.00058
摘要
Software development is among the first demonstrations of using Large Language Models (LLMs) to enhance human productivity. Such a co-pilot paradigm envisions LLM working side-by-side with human developers to assist in programming tasks. Ensuring the security of software products is a critical factor for the success of such a paradigm. There have been various anecdotal reports on the success of using LLMs to detect vulnerabilities in programs. This paper reports a set of experiments applying four well-known LLMs to two widely referenced public datasets to evaluate the performance of LLMs in detecting software vulnerabilities. Our results show a significant performance gap between these LLMs and those from popular static analysis tools, primarily due to their high false positive rates. However, LLMs show great promise in identifying subtle patterns commonly associated with software vulnerabilities. This observation suggests a possible path forward by combining LLMs and other program analysis techniques to achieve better software vulnerability detection.
科研通智能强力驱动
Strongly Powered by AbleSci AI