【论文回顾】The Limitations of Deep Learning in Adversarial Settings

The Limitations of Deep Learning in Adversarial Settings

paper notes:

This paper introduces the background of adversarial examples including adversarial goals, capabilities and then explains how to generate adversarial examples by forward gradient: just take the derivative of network on input features. that is,

another picture helps to understand:

The general algorithm they proposed:

In this paper, saliency map is soptlight. They induced this based on the forward derivative and help us understand the existing of adv examples and they succeed both by increasing or decreasing pixel intensities.

they found decreasing is more less successful because it reduces information entropy and makes harder to extract information by dnn to classify.

In evaluation, they study class pair(source-target) and found that there exists some pairs are harder. they do the hardness measure (measure of quantifying the distance between two classes) and adversarial distance (predictive measure from adversarial saliency maps). At last, they study the human perception of adversarial samples.

Strengths:

1.reducing the distortion (L0: the number of features altered)

2.induce the adversarial saliency map

3.mitigate the adversarial examples: measure hardness and adversarial distance.

Detailed comments, possible improvements, or related ideas:

1.defense is possible by evaluating the regularity of examples. for example, the squared difference between each pair of neighbouring pixels is always higher for adversarial examples than for benign examples.

 

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值