模糊测试
Python(编程语言)
计算机科学
人工智能
语法
语义学(计算机科学)
自然语言处理
程序设计语言
软件
作者
Yinlin Deng,Chunqiu Steven Xia,Chenyuan Yang,Shizhuo Dylan Zhang,Shujing Yang,Lingming Zhang
标识
DOI:10.1145/3597503.3623343
摘要
Bugs in Deep Learning (DL) libraries may affect almost all downstream DL applications, and it is crucial to ensure the quality of such systems. It is challenging to generate valid input programs for fuzzing DL libraries, since the input programs need to satisfy both the syntax/semantics of the supported languages (e.g., Python) and the tensor/operator constraints for constructing valid computational graphs. Recently, the TitanFuzz work demonstrates that modern Large Language Models (LLMs) can be directly leveraged to implicitly learn all the language and DL computation constraints to generate valid programs for fuzzing DL libraries (and beyond). However, LLMs tend to generate ordinary programs following similar patterns/tokens with typical programs seen in their massive pre-training corpora (e.g., GitHub), while fuzzing favors unusual inputs that cover edge cases or are unlikely to be manually produced.
科研通智能强力驱动
Strongly Powered by AbleSci AI