...
首页> 外文期刊>Neurocomputing >Attention augmentation with multi-residual in bidirectional LSTM
【24h】

Attention augmentation with multi-residual in bidirectional LSTM

机译:双向LSTM中具有多残差的注意力增强

获取原文
获取原文并翻译 | 示例
           

摘要

Recurrent neural networks (RNNs) have been proven to be efficient in processing sequential data. However, the traditional RNNs have suffered from the gradient diminishing problem until the advent of Long Short-Term Memory (LSTM). However, LSTM is weak in capturing long-time dependency in sequential data due to the inadequacy of memory capacity in LSTM cells. To address this challenge, we propose an Attention-augmentation Bidirectional Multi-residual Recurrent Neural Network (ABMRNN) to overcome the deficiency. We propose an algorithm which integrates both past and future information at every time step with omniscient attention model. The multi-residual mechanism has also been leveraged in the proposed model targeting the pattern of the relationship between current time step and further distant time steps instead of only one previous time step. The results of experiments show that our model outperforms the traditional statistical classifiers and other existing RNN architectures. (C) 2020 Elsevier B.V. All rights reserved.
机译:事实证明,递归神经网络(RNN)可有效处理顺序数据。但是,传统的RNN一直存在梯度递减问题,直到长短期记忆(LSTM)出现为止。但是,由于LSTM单元中的存储容量不足,LSTM在捕获顺序数据中的长期依赖性方面很弱。为了解决这一挑战,我们提出了一种注意力增强双向多残差递归神经网络(ABMRNN)来克服这一缺陷。我们提出了一种算法,该算法利用全知的注意力模型将每个时间步骤的过去和将来的信息进行整合。在针对当前时间步长与其他较远的时间步长之间的关系模式的拟议模型中,也已利用了多残差机制,而不仅仅是一个先前的时间步长。实验结果表明,我们的模型优于传统的统计分类器和其他现有的RNN体系结构。 (C)2020 Elsevier B.V.保留所有权利。

著录项

  • 来源
    《Neurocomputing》 |2020年第14期|340-347|共8页
  • 作者

  • 作者单位

    Texas A&M Univ Dept Elect & Comp Engn College Stn TX 77840 USA|Chongqing Univ Posts & Telecommun Coll Comp Sci & Technol Chongqing 400065 Peoples R China;

    Southern Methodist Univ Dept Elect Engn Dallas TX 75205 USA;

    Texas A&M Univ Dept Elect & Comp Engn College Stn TX 77840 USA;

    Texas A&M Univ Dept Comp Sci College Stn TX 77840 USA;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Long Short-Term Memory; Attention augmentation; Natural language processing; Multi-residual network;

    机译:长短期记忆;注意力增强;自然语言处理;多残留网络;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号