首页> 外国专利> Interpretability-Aware Adversarial Attack and Defense Method for Deep Learnings

Interpretability-Aware Adversarial Attack and Defense Method for Deep Learnings

机译:深度学习的可解释性感知对抗攻击与防御方法

摘要

Embodiments relate to a system, program product, and method to support a convolutional neural network (CNN). A class-specific discriminative image region is localized to interpret a prediction of a CNN and to apply a class activation map (CAM) function to received input data. First and second attacks are generated on the CNN with respect to the received input data. The first attack generates first perturbed data and a corresponding first CAM, and the second attack generates second perturbed data and a corresponding second CAM. An interpretability discrepancy is measured to quantify one or more differences between the first CAM and the second CAM. The measured interpretability discrepancy is applied to the CNN. The application is a response to an inconsistency between the first CAM and the second CAM and functions to strengthen the CNN against an adversarial attack.
机译:实施例涉及一种支持卷积神经网络(CNN)的系统,程序产品和方法。类特定的鉴别图像区域是本地化以解释CNN的预测并将类激活图(CAM)功能应用于接收的输入数据。关于接收的输入数据,在CNN上生成第一和第二攻击。第一攻击产生第一扰动数据和相应的第一凸轮,第二攻击产生第二扰动数据和相应的第二凸轮。测量可解释性差异以量化第一凸轮和第二凸轮之间的一个或多个差异。测量的解释性差异施加到CNN。申请是对第一凸轮和第二凸轮之间的不一致的响应,并且功能以加强CNN免受对抗攻击。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号