Recently developed adversarial attacks on neural networks have become more aggressive and dangerous, because of which Artificial Intelligence (AI) models are no longer sufficiently robust against them. It is important to have a set of effective and reliable methods to detect malicious attacks to ensure the security of AI models. Such standardized methods can also serve as a reference for researchers to develop robust models and new kinds of attacks. This study proposes a method to assess the robustness of AI models. Six commonly used image classification CNN models were evaluated when subjected to 13 types of adversarial attacks. The robustness of the models is calculated unbiased and can be used as a reference for further improvement. It is distinguished from prior related works that our algorithm is attack-agnostic and is applicable to neural network model.
Yiyi TaoYixian ShenHang ZhangYanxin ShenLun WangChuanqi ShiShaoshuai Du
Dimitrios Christos AsimopoulosPanagiotis Radoglou‐GrammatikisΘωμάς ΛάγκαςVasileios ArgyriouIoannis D. MoscholiosJorgen CaniGeorgios Th. PapadopoulosEvangelos MarkakisPanagiotis Sarigiannidis
Gean Trindade PereiraAndré C. P. L. F. de Carvalho