JOURNAL ARTICLE

Bolstering Adversarial Robustness with Latent Disparity Regularization

Abstract

Recent research has revealed that neural networks and other machine learning models are vulnerable to adversarial attacks that aim to subvert their predictions' integrity or privacy by adding a small calculated perturbation to inputs. Further, the adversary can significantly degrade the performance of the model. The number and severity of attacks continues to grow. However, a dearth of techniques robustly defends machine learning models in a computationally inexpensive way. Against this background, we propose an adversarially robust training procedure and objective function for arbitrary neural network architectures. Robustness of neural networks against adversarial attacks on integrity is achieved by augmentation of a novel regularization term. This regularizer penalizes the discrepancy between the representations induced in hidden layers by benign and adversarial data. We benchmark our regularization approach on the Fashion-Mnist and Cifar-10 datasets. Our model is benchmarked against three state-of-the-art defense methods, namely: (i) regularization to the largest eigenvalue in the Fisher information matrix of the activity of the terminal layer, (ii) a higher-level representation guided denoising autoencoder (trained with adversarial examples), and (iii) training an otherwise undefended model on data distorted by additive white Gaussian noise. Our experiments show that the proposed regularizer provides significant improvements in adversarial robustness over both an undefended baseline model as well as the same model defended with other techniques. This result is observed over several adversarial budgets with only a small (but seemingly unavoidable) decline in benign test accuracy.

Keywords:
MNIST database Adversarial system Computer science Artificial intelligence Robustness (evolution) Machine learning Regularization (linguistics) Artificial neural network Deep neural networks Pattern recognition (psychology) Algorithm

Metrics

3
Cited By
0.42
FWCI (Field Weighted Citation Impact)
46
Refs
0.69
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Adversarial Robustness in Machine Learning
Physical Sciences →  Computer Science →  Artificial Intelligence
Anomaly Detection Techniques and Applications
Physical Sciences →  Computer Science →  Artificial Intelligence
Advanced Neural Network Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

JOURNAL ARTICLE

Consistency Regularization for Adversarial Robustness

Jihoon TackSihyun YuJongheon JeongMinseon KimSung Ju HwangJinwoo Shin

Journal:   Proceedings of the AAAI Conference on Artificial Intelligence Year: 2022 Vol: 36 (8)Pages: 8414-8422
JOURNAL ARTICLE

Adversarial Robustness Via Fisher-Rao Regularization

Marine PicotFrancisco MessinaMalik BoudiafFabrice LabeauIsmail Ben AyedPablo Piantanida

Journal:   IEEE Transactions on Pattern Analysis and Machine Intelligence Year: 2022 Vol: 45 (3)Pages: 2698-2710
© 2026 ScienceGate Book Chapters — All rights reserved.