【论文阅读】Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models(2018)

摘要

In recent years(近年来), deep neural network approaches(深度神经网络方法) have been widely adopted for(被广泛应用于) machine learning tasks(机器学习任务), including classification(分类). However, they were shown(被证明) to be vulnerable(容易受到) to adversarial perturbations(对抗性扰动): carefully crafted small perturbations(精心制作的小干扰) can cause misclassification of legitimate images(合法图像的错误分类). We propose Defense-GAN, a new framework(新框架) leveraging(利用) the expressive capability(表达能力) of generative models(生成模型) to defend(防御) deep neural networks against such attacks. Defense-GAN is trained to model the distribution of unperturbed images(无扰动图像的分布). At inference time(在推理时), it finds a close output(接近输出) to a given image(给定图像) which does not contain the adversarial changes(不包含对抗性变化). This output is then fed to the classifier(反馈给分类器). Our proposed method(我们提出的方法) can be used with any classification model(用在任何分类模型) and does not modify the classifier structure or training procedure(不修改分类器结构或训练过程). It can also be used as a defense against any attack(防御任何攻击) as it does not assume(假设) knowledge of the process(过程的知识) for generating the adversarial examples(生成对抗样本). We empirically show (经验表明)that Defense-GAN is consistently effective(一致有效) against different attack methods(不同的攻击方法) and improves on existing defense strategies(改进了现有的防御策略).

方法

在这里插入图片描述

论文链接

Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models

  • 3
    点赞
  • 5
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

Bosenya12

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值