首页> 外文会议>European Conference on Information Retrieval >Exploring Classic and Neural Lexical Translation Models for Information Retrieval: Interpretability, Effectiveness,and Efficiency Benefits
【24h】

Exploring Classic and Neural Lexical Translation Models for Information Retrieval: Interpretability, Effectiveness,and Efficiency Benefits

机译:探索信息检索的经典和神经词汇翻译模型:可解释性,有效性和效率效益

获取原文

摘要

We study the utility of the lexical translation model (IBM Model 1) for English text retrieval, in particular, its neural variants that are trained end-to-end. We use the neural Modell as an aggregator layer applied to context-free or contextualized query/document embeddings. This new approach to design a neural ranking system has benefits for effectiveness, efficiency, and interpretability. Specifically, we show that adding an interpretable neural Model 1 layer on top of BERT-based contextualized embeddings (1) does not decrease accuracy and/or efficiency; and (2) may overcome the limitation on the maximum sequence length of existing BERT models. The context-free neural Model 1 is less effective than a BERT-based ranking model, but it can run efficiently on a CPU (without expensive index-time precomputation or query-time operations on large tensors). Using Model 1 we produced best neural and non-neural runs on the MS MARCO document ranking leaderboard in late 2020.
机译:我们研究了词汇翻译模型(IBM Model 1)进行英语文本检索的效用,特别是其培训结束到底的神经变体。 我们使用神经模式作为应用于无内容或上下文化查询/文档嵌入的聚合器层。 这种设计神经排名系统的新方法具有有效性,效率和可解释性的益处。 具体地,我们表明在基于BERT的上下文化嵌入物(1)顶部添加可解释的神经模型1层不会降低精度和/或效率; (2)可以克服现有BERT模型的最大序列长度的限制。 无与伦比的神经模型1比基于BERT的排名模型更低,但它可以有效地在CPU上运行(没有昂贵的指数 - 时间预先计算或大张力上的查询时间操作)。 使用型号1我们在2020年底,我们在MS Marco文件排名排行榜上产生了最佳的神经和非神经运行。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号