首页> 外文会议>Conference on Uncertainty in Artificial Intelligence >IDK Cascades: Fast Deep Learning by Learning not to Overthink
【24h】

IDK Cascades: Fast Deep Learning by Learning not to Overthink

机译:IDK瀑布:通过学习不要过度思考快速学习

获取原文

摘要

Advances in deep learning have led to substantial increases in prediction accuracy but have been accompanied by increases in the cost of rendering predictions. We conjecture that for a majority of real-world inputs, the recent advances in deep learning have created models that effectively "over-think" on simple inputs. In this paper we revisit the classic question of building model cascades that primarily leverage class asymmetry to reduce cost. We introduce the "I Don't Know" (IDK) prediction cascades framework, a general framework to systematically compose a set of pre-trained models to accelerate inference without a loss in prediction accuracy. We propose two search based methods for constructing cascades as well as a new cost-aware objective within this framework. The proposed IDK cascade framework can be easily adopted in the existing model serving systems without additional model retraining. We evaluate the proposed techniques on a range of benchmarks to demonstrate the effectiveness of the proposed framework.
机译:深度学习的进步导致预测准确性的大幅增加,但伴随着呈现预测成本的增加。我们推测,为大多数现实投入投入,深度学习的最近进步已经创造了有效地“过度思考”的模型。在本文中,我们重新审视建筑模型级联的经典问题,主要利用阶级不对称以降低成本。我们介绍了“我不知道”(IDK)预测级联框架,一般框架来系统地构成一组预先训练的模型,以加速推理而不会以预测精度损失。我们提出了两种基于搜索的方法,用于构建级联和此框架内的新的成本感知目标。在没有额外的模型再培训的情况下,可以在现有的模型服务系统中轻松采用所提出的IDK级联框架。我们评估了一系列基准的建议技术来证明所提出的框架的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号