首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Loss Decomposition for Fast Learning in Large Output Spaces
【24h】

Loss Decomposition for Fast Learning in Large Output Spaces

机译:大输出空间中的快速学习损失分解

获取原文
           

摘要

For problems with large output spaces, evaluation of the loss function and its gradient are expensive, typically taking linear time in the size of the output space. Recently, methods have been developed to speed up learning via efficient data structures for Nearest-Neighbor Search (NNS) or Maximum Inner-Product Search (MIPS). However, the performance of such data structures typically degrades in high dimensions. In this work, we propose a novel technique to reduce the intractable high dimensional search problem to several much more tractable lower dimensional ones via dual decomposition of the loss function. At the same time, we demonstrate guaranteed convergence to the original loss via a greedy message passing procedure. In our experiments on multiclass and multilabel classification with hundreds of thousands of classes, as well as training skip-gram word embeddings with a vocabulary size of half a million, our technique consistently improves the accuracy of search-based gradient approximation methods and outperforms sampling-based gradient approximation methods by a large margin.
机译:对于输出空间较大的问题,损失函数及其梯度的评估很昂贵,通常会在输出空间的大小上花费线性时间。近来,已经开发了通过用于最近邻居搜索(NNS)或最大内部产品搜索(MIPS)的有效数据结构来加快学习速度的方法。但是,此类数据结构的性能通常会在高维度上下降。在这项工作中,我们提出了一种新技术,通过损失函数的双重分解,将难处理的高维搜索问题减少为几个更易处理的低维搜索问题。同时,我们通过贪婪的消息传递过程证明了对原始损失的有保证的收敛。在我们针对数十万个类别的多类别和多标签分类进行的实验以及训练词汇量为一百万的跳过克单词嵌入的过程中,我们的技术不断提高了基于搜索的梯度近似方法的准确性,并且胜过了采样基于梯度的近似方法很大。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号