模糊测试
正确性
计算机科学
稳健性(进化)
操作员(生物学)
推论
人工智能
人工神经网络
深度学习
机器学习
软件
算法
程序设计语言
生物化学
化学
抑制因子
转录因子
基因
作者
Xufan Zhang,Jiawei Liu,Ning Sun,Chunrong Fang,Jia Liu,Jiang Wang,Dong Chai,Zhenyu Chen
标识
DOI:10.1109/tr.2021.3107165
摘要
Deep learning (DL) libraries reduce the barriers to the DL model construction. In DL libraries, various building blocks are DL operators with different functionality, responsible for processing high-dimensional tensors during training and inference. Thus, the quality of operators could directly impact the quality of models. However, existing DL testing techniques mainly focus on robustness testing of trained neural network models and cannot locate DL operators' defects. The insufficient test input and undetermined test output in operator testing have become challenging for DL library developers. In this article, we propose an approach, namely Duo, which combines fuzzing techniques and differential testing techniques to generate input and evaluate corresponding output. It implements mutation-based fuzzing to produce tensor inputs by employing nine mutation operators derived from genetic algorithms and differential testing to evaluate outputs' correctness from multiple operator instances. Duo is implemented in a tool and used to evaluate seven operators from TensorFlow, PyTorch, MNN, and MXNet in an experiment. The result shows that Duo can expose defects of DL operators and realize multidimension evaluation for DL operators from different DL libraries.
科研通智能强力驱动
Strongly Powered by AbleSci AI