Finding Patterns in Adversarial Training
Yutong Wang1,2; Fei-Yue Wang1
刊名Chinese Automation Congress
2020
期号页码:gress. 2020: 4130 4134.
关键词Adversarial training Feature map denoising Channel pruning
英文摘要

Adversarial training has become an universally accepted robust method to train networks defending against adversarial attacks. However, feeding networks with adversarial examples is a cumbersome process, and there is no clear explanation for the differences between networks with the same architecture that are learned by adversarial training and learned by natural training method. So in this paper, we focus on the patterns of network learned by adversarial training. For comparison, we visualize the weights and feature maps of networks learned by these two training methods on MNIST, and find some patterns of adversarially trained model that are important for defending against adversarial examples and that are unique to naturally trained model. First, we find that adversarially trained network denoises the noisy data caused by adversarial images and enhances the outlines of semantically informative content in feature maps. Second, we find that adversarial training performs model compression, making some of the channels of a convolutional layer have no activation on input images. Further observation shows that different from naturally trained model, which have different most activation channel corresponding to different input, adversarially trained model often has the same most activation channel regardless of input. This phenomenon also leads to a sparse network reducing 50% number of parameters and 87% computation costs, and the pruned network performs as good as original network after pruning the least activation channels without retraining. But the naturally trained counterpart loses more than 40% accuracy using the same pruning strategy.

语种英语
内容类型期刊论文
源URL[http://ir.ia.ac.cn/handle/173211/44701]  
专题自动化研究所_复杂系统管理与控制国家重点实验室_先进控制与自动化团队
通讯作者Fei-Yue Wang
作者单位1.The State Key Laboratory for Management and Control of Complex Systems, Institute of Automation, Chinese Academy of Sciences
2.University of Chinese Academy of Sciences
推荐引用方式
GB/T 7714
Yutong Wang,Fei-Yue Wang. Finding Patterns in Adversarial Training[J]. Chinese Automation Congress,2020(无):gress. 2020: 4130 4134..
APA Yutong Wang,&Fei-Yue Wang.(2020).Finding Patterns in Adversarial Training.Chinese Automation Congress(无),gress. 2020: 4130 4134..
MLA Yutong Wang,et al."Finding Patterns in Adversarial Training".Chinese Automation Congress .无(2020):gress. 2020: 4130 4134..
个性服务
查看访问统计
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。


©版权所有 ©2017 CSpace - Powered by CSpace