• May 23, 2018 News![CFP] 2018 the annual meeting of IJMLC Editorial Board, ACMLC 2018, will be held in Ho Chi Minh, Vietnam, December 7-9, 2018   [Click]
  • May 23, 2018 News!Good News! All papers from Volume 8, Number 1 have been indexed by Scopus!   [Click]
  • Jun 21, 2018 News!Vol.8, No.3 has been published with online version.   [Click]
Search
General Information
Editor-in-chief
Dr. Lin Huang
Metropolitan State University of Denver, USA
It's my honor to take on the position of editor in chief of IJMLC. We encourage authors to submit papers concerning any branch of machine learning and computing.
IJMLC 2018 Vol.8(3): 203-207 ISSN: 2010-3700
DOI: 10.18178/ijmlc.2018.8.3.688

Gradient Masking Is a Type of Overfitting

Yusuke Yanagita and Masayuki Yamamura
Abstract—Neural networks have recently been attracting attention again as classifiers with high accuracy, so called “deep learning,” which is applied in a wide variety of fields. However, this advanced machine learning algorithms are vulnerable to adversarial perturbations. Although they cannot be recognized by humans, these perturbations deliver a fatal blow to the estimation ability of classifiers. Thus, while humans perceive perturbed examples as being the same as the original natural examples, sophisticated classifiers identify them as completely different examples. Although several defensive measures against such adversarial examples have been suggested, they are known to fail in undesirable phenomena, gradient masking. Gradient masking can neutralize the useful gradient for adversaries, but adversarial perturbations tend to transfer across most models, and these models can be deceived by adversarial examples crafted based on other models, which is called a black-box attack. Therefore, it is necessary to develop training methods to withstand black-box attacks and conduct studies to investigate the weak points of current NN training. This paper argues that no special defensive measures are necessary for NN to fall into gradient masking, and it is sufficient to slightly change the initial learning rate of Adam from the recommended value. Moreover, our experiment implies that gradient masking is a type of overfitting.

Index Terms—Adam, adversarial examples, gradient masking, machine learning, neural network.

The authors are with the Department of Computer Science, School of Computer, Tokyo Institute of Technology, Yokohama, 226-8503, Japan (e-mail: y_yanagita@ali.c.titech.ac.jp, my@c.titech.ac.jp).

[PDF]

Cite: Yusuke Yanagita and Masayuki Yamamura, "Gradient Masking Is a Type of Overfitting," International Journal of Machine Learning and Computing vol. 8, no. 3, pp. 203-207, 2018.

Copyright © 2008-2018. International Journal of Machine Learning and Computing. All rights reserved.
E-mail: ijmlc@ejournal.net