首页> 外文期刊>ACM transactions on Asian and low-resource language information processing >Learning Syllables Using Conv-LSTM Model for Swahili Word Representation and Part-of-speech Tagging
【24h】

Learning Syllables Using Conv-LSTM Model for Swahili Word Representation and Part-of-speech Tagging

机译:使用Conv-LSTM模型进行斯瓦希里语字表示和词语标记的学习音节

获取原文
获取原文并翻译 | 示例
           

摘要

The need to capture intra-word information in natural language processing (NLP) tasks has inspired research in learning various word representations at word, character, or morpheme levels, but little attention has been given to syllables from a syllabic alphabet. Motivated by the success of compositional models in morphological languages, we present a Convolutional-long short term memory (Conv-LSTM) model for constructing Swahili word representation vectors from syllables. The unified architecture addresses the word agglutination and polysemous nature of Swahili by extracting high-level syllable features using a convolutional neural network (CNN) and then composes quality word embeddings with a long short term memory (LSTM). The word embeddings are then validated using a syllable-aware language model (31.267) and a part-of-speech (POS) tagging task (98.78), both yielding very competitive results to the state-of-art models in their respective domains. We further validate the language model using Xhosa and Shona, which are syllabic-based languages. The novelty of the study is in its capability to construct quality word embeddings from syllables using a hybrid model that does not use max-over-pool common in CNN and then the exploitation of these embeddings in POS tagging. Therefore, the study plays a crucial role in the processing of agglutinative and syllabic-based languages by contributing quality word embeddings from syllable embeddings, a robust Conv-LSTM model that learns syllables for not only language modeling and POS tagging, but also for other downstream NLP tasks.
机译:需要在自然语言处理中捕获词内信息(NLP)任务的灵感研究了在语言,角色或语素级别学习各种单词表示的研究,但是从音节字母表中都有很少的关注。通过形态语言成分模型的成功,我们提出了一种卷积的长短期记忆(CONC-LSTM)模型,用于构建音节的斯瓦希里语字表示向量。统一架构通过使用卷积神经网络(CNN)提取高级音节特征来解决斯瓦希里语的单词凹陷和多仪性质,然后用​​长短短期内存(LSTM)归组质量字嵌入品。然后使用音节感知语言模型(31.267)和一个语音(POS)标记任务(98.78)进行验证单词eMbeddings,两者都在其各自的域中对最先进的模型产生非常竞争力的结果。我们进一步使用Xhosa和Shona验证了语言模型,这些模型是基于Syllabic的语言。该研究的新颖性是它的能力,它可以使用在CNN中不使用MAX-过池的混合模型来构建质量词嵌入的音色,然后在POS标记中利用这些嵌入的嵌入式。因此,该研究在从音节嵌入品中贡献质量单词嵌入来处理agGlutinative和Syslabic的语言中的一个至关重要的作用,这是一个强大的conv-lstm模型,它不仅为语言建模和POS标记而学习音节,而且还用于其他下游NLP任务。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号