Deep Neural Networks (DNNs) have a wide range of applications, and software employing them must be thoroughly tested, especially in safety-critical domains. However, traditional software test coverage metrics cannot be applied directly to DNNs. In this paper, inspired by the MC/DC coverage criterion, we propose four novel test criteria that are tailored to structural features of DNNs and their semantics. We validate the criteria by demonstrating that the generated test inputs, guided by our coverage criteria, are able to capture undesirable behaviours in DNNs. Test cases are generated using both a symbolic approach and a gradient-based heuristic. Our experiments are conducted on state-of-the-art DNNs, obtained using the MNIST and ImageNet datasets.
Youcheng SunXiaowei HuangDaniel KroeningJames J. SharpMatthew Q. HillRob Ashmore
Zenan LiXiaoxing MaChang XuChun Cao
Dongcheng LiW. Eric WongHu LiuMan ZhaoZizhao Chen
Abdesselam LakehalIoannis Parissis