JOURNAL ARTICLE

Robustness Against Adversarial Attacks Via Learning Confined Adversarial Polytopes

Abstract

Deep neural networks (DNNs) could be deceived by generating human-imperceptible perturbations of clean samples. Therefore, enhancing the robustness of DNNs against adversarial attacks is a crucial task. In this paper, we aim to train robust DNNs by limiting the set of outputs reachable via a norm-bounded perturbation added to a clean sample. We refer to this set as adversarial polytope, and each clean sample has a respective adversarial polytope. Indeed, if the respective polytopes for all the samples are compact such that they do not intersect the decision boundaries of the DNN, then the DNN is robust against adversarial samples. Hence, the inner-working of our algorithm is based on learning confined adversarial polytopes (CAP). By conducting a thorough set of experiments, we demonstrate the effectiveness of CAP over existing adversarial robustness methods in improving the robustness of models against state-of-the-art attacks including AutoAttack.

Keywords:
Adversarial system Polytope Robustness (evolution) Bounded function Deep neural networks Computer science Norm (philosophy) Artificial intelligence Limiting Deep learning Mathematical optimization Mathematics Combinatorics Engineering

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
31
Refs
0.03
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Adversarial Robustness in Machine Learning
Physical Sciences →  Computer Science →  Artificial Intelligence
Integrated Circuits and Semiconductor Failure Analysis
Physical Sciences →  Engineering →  Electrical and Electronic Engineering
Physical Unclonable Functions (PUFs) and Hardware Security
Physical Sciences →  Computer Science →  Hardware and Architecture
© 2026 ScienceGate Book Chapters — All rights reserved.