Blogs (1) >>
ICSE 2019
Sat 25 - Fri 31 May 2019 Montreal, QC, Canada

Deep neural networks (DNN) have been shown to be useful in a wide range of applications. However, they are also known to be vulnerable to adversarial samples. By transforming a normal sample with some carefully crafted human non-perceptible perturbations, even highly accurate DNN makes wrong decisions. Multiple defense mechanisms have been proposed which aim to hinder the generation of such adversarial samples. However, a recent work show that most of them are ineffective. In this work, we propose an alternative approach to detect adversarial samples at runtime. Our main observation is that adversarial samples are much more sensitive than normal samples if we impose random mutations on the DNN. We thus first propose a measure of `sensitivity’ and show empirically that normal samples and adversarial samples have distinguishable sensitivity. We then integrate statistical model checking and mutation testing to check whether an input sample is normal or adversarial at runtime by measuring its sensitivity. We evaluated our approach on the MNIST and CIFAR10 dataset. The results show that our approach detects adversarial samples generated by state-of-art attacking methods efficiently and accurately.

Fri 31 May

Displayed time zone: Eastern Time (US & Canada) change

16:00 - 17:20
Testing and Analysis: Domain-Specific ApproachesTechnical Track / Journal-First Papers / Papers at Place du Canada
Chair(s): Gregory Gay University of South Carolina, Chalmers | University of Gothenburg
16:00
20m
Talk
Detecting Incorrect Build RulesArtifacts AvailableACM SIGSOFT Distinguished Paper AwardTechnical Track
Technical Track
Nandor Licker University of Cambridge, Andrew Rice University of Cambridge, UK
Pre-print Media Attached
16:20
20m
Talk
Adversarial Sample Detection for Deep Neural Network through Model Mutation TestingTechnical Track
Technical Track
Jingyi Wang National University of Singapore, Singapore, Guoliang Dong Computer College of Zhejiang University, Jun Sun Singapore Management University, Singapore, Xinyu Wang Zhejiang University, Peixin Zhang Zhejiang University
16:40
10m
Talk
Oracles for Testing Software Timeliness with UncertaintyJournal-First
Journal-First Papers
Chunhui Wang University of Luxembourg, Fabrizio Pastore University of Luxembourg, Lionel Briand SnT Centre/University of Luxembourg
16:50
20m
Talk
Deep Differential Testing of JVM ImplementationsTechnical Track
Technical Track
Yuting Chen Shanghai Jiao Tong University, Ting Su Nanyang Technological University, Singapore, Zhendong Su ETH Zurich
17:10
10m
Talk
Discussion Period
Papers