Adversarial Sample Generation and Training using Neural Network

The neural network classifier is known to be susceptible to adversarial attacks, where projected gradient descent-like noise is added to the data, causing misclassification. These attacks can be prevented by min-max training, where the neural network is trained to handle adversarial attack data. Alt...

Full description

Saved in:
Bibliographic Details
Published inKorean Institute of Smart Media Vol. 13; no. 10; pp. 43 - 49
Main Author Jung, Ho Yub
Format Journal Article
LanguageEnglish
Published (사)한국스마트미디어학회 31.10.2024
Subjects
Online AccessGet full text
ISSN2287-1322
2288-9671
DOI10.30693/SMJ.2024.13.10.43

Cover

More Information
Summary:The neural network classifier is known to be susceptible to adversarial attacks, where projected gradient descent-like noise is added to the data, causing misclassification. These attacks can be prevented by min-max training, where the neural network is trained to handle adversarial attack data. Although min-max training is very effective, it requires a large amount of training time because each adversarial attack data generation requires several iterations of gradient back-propagation to produce. In this paper, convolutional layers are used to replace the projected gradient descent-based production of adversarial attack data in an attempt to reduce the training time. By replacing the adversarial noise generation with the output of convolutional layers, the training time becomes comparable to that of a simple neural network classifier with a few additional layers. The proposed approach significantly reduced the effects of smaller-scale adversarial attacks, and under certain circumstances, was shown to be as effective as min-max training. However, for severe attacks, the proposed approach was not able to compete with modern min-max-based remedies. KCI Citation Count: 0
ISSN:2287-1322
2288-9671
DOI:10.30693/SMJ.2024.13.10.43