Abstract

We propose a new regularization method based on virtual adversarial loss: a new measure of local smoothness of the conditional label distribution given input. Virtual adversarial loss is defined as the robustness of the conditional label distribution around each input data point against local perturbation. Unlike adversarial training, our method defines the adversarial direction without label information and is hence applicable to semi-supervised learning. Because the directions in which we smooth the model are only "virtually" adversarial, we call our method virtual adversarial training (VAT). The computational cost of VAT is relatively low. For neural networks, the approximated gradient of virtual adversarial loss can be computed with no more than two pairs of forward- and back-propagations. In our experiments, we applied VAT to supervised and semi-supervised learning tasks on multiple benchmark datasets. With a simple enhancement of the algorithm based on the entropy minimization principle, our VAT achieves state-of-the-art performance for semi-supervised learning tasks on SVHN and CIFAR-10.

Keywords

Adversarial systemComputer scienceArtificial intelligenceMachine learningRegularization (linguistics)Entropy (arrow of time)MinificationSupervised learningSemi-supervised learningCross entropyArtificial neural networkPattern recognition (psychology)

Affiliated Institutions

Related Publications

Publication Info

Year
2018
Type
article
Volume
41
Issue
8
Pages
1979-1993
Citations
2696
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

2696
OpenAlex
366
Influential
1695
CrossRef

Cite This

Takeru Miyato, Shin‐ichi Maeda, Masanori Koyama et al. (2018). Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning. IEEE Transactions on Pattern Analysis and Machine Intelligence , 41 (8) , 1979-1993. https://doi.org/10.1109/tpami.2018.2858821

Identifiers

DOI
10.1109/tpami.2018.2858821
PMID
30040630
arXiv
1704.03976

Data Quality

Data completeness: 88%