首页> 外国专利> DEFENDING MACHINE LEARNING SYSTEMS FROM ADVERSARIAL ATTACKS

DEFENDING MACHINE LEARNING SYSTEMS FROM ADVERSARIAL ATTACKS

机译:防御来自对抗攻击的机器学习系统

摘要

Techniques are disclosed for detecting adversarial attacks. A machine learning (ML) system processes the input into and output of a ML model using an adversarial detection module that does not include a direct external interface. The adversarial detection module includes a detection model that generates a score indicative of whether the input is adversarial using, e.g., a neural fingerprinting technique or a comparison of features extracted by a surrogate ML model to an expected feature distribution for the output of the ML model. In turn, the adversarial score is compared to a predefined threshold for raising an adversarial flag. Appropriate remedial measures, such as notifying a user, may be taken when the adversarial score satisfies the threshold and raises the adversarial flag.
机译:公开了用于检测对抗性攻击的技术。机器学习(ML)系统使用不包括直接外部接口的对手检测模块处理ML模型的输入和输出。对手检测模块包括一种检测模型,其产生指示输入是对越野的分数,例如,神经指纹技术或由代理ML模型提取的特征与ML模型输出输出的预期特征分布的特征的比较。反过来,将对越野的分数与预定义的阈值进行比较,以提高对抗性标志。当对抗性分数满足阈值并提高对抗旗帜时,可以采取适当的补救措施,例如通知使用者。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号