【24h】

CA-RNN: Using Context-Aligned Recurrent Neural Networks for Modeling Sentence Similarity

机译:CA-RNN:使用上下文对齐的经常性神经网络来建模句子相似性

获取原文

摘要

The recurrent neural networks (RNNs) have shown good performance for sentence similarity modeling in recent years. Most RNNs focus on modeling the hidden states based on the current sentence, while the context information from the other sentence is not well investigated during the hidden state generation. In this paper, we propose a context-aligned RNN (CA-RNN) model, which incorporates the contextual information of the aligned words in a sentence pair for the inner hidden state generation. Specifically, we first perform word alignment detection to identify the aligned words in the two sentences. Then, we present a context alignment gating mechanism and embed it into our model to automatically absorb the aligned words' context for the hidden state update. Experiments on three benchmark datasets, namely TREC-QA and WikiQA for answer selection and MSRP for paraphrase identification, show the great advantages of our proposed model. In particular, we achieve the new state-of-the-art performance on TREC-QA and WikiQA. Furthermore, our model is comparable to if not better than the recent neural network based approaches on MSRP.
机译:近年来,经常性的神经网络(RNNS)对句子相似性建模具有良好的性能。大多数RNNS专注于基于当前句子建模隐藏状态,而在隐藏状态生成期间没有很好地研究来自其他句子的上下文信息。在本文中,我们提出了一种上下文 - 对齐的RNN(CA-RNN)模型,其包括用于内部隐藏状态生成的句子对中的对齐词的上下文信息。具体地,我们首先执行词对准检测以识别两句话中的对齐字。然后,我们介绍了一个上下文对齐选通机制,并将其嵌入到我们的模型中以自动吸收隐藏状态更新的对齐字的上下文。关于三个基准数据集的实验,即TREC-QA和WikiQA用于答案选择和MSRP进行解释识别,表现出我们所提出的模型的巨大优势。特别是,我们在TREC-QA和WikiQA上实现了新的最先进的表现。此外,我们的模型可以与最近的基于神经网络的MSRP方法更好。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号