首页> 外文会议>IEEE International Conference on Computational Science and Engineering >Sentiment Analysis of Code-Mixed Roman Urdu-English Social Media Text using Deep Learning Approaches
【24h】

Sentiment Analysis of Code-Mixed Roman Urdu-English Social Media Text using Deep Learning Approaches

机译:使用深度学习方法的码混罗马乌尔都语 - 英语社交媒体文本的情感分析

获取原文

摘要

Sentiment analysis is the computational study of attitudes, opinions, and sentiments towards certain issues, products, individuals, and organizations. Companies and customers are making decisions by seeking opinions from social media text. Sentiment analysis is getting intelligent with the advancement of artificial intelligence and natural language processing. With a stunning increase in the use of social media, a huge volume of text available on these platforms is in imperfect and informal languages like Roman Urdu mixed with the English language. Present sentiment analysis techniques do not perform precisely on these code-mixed imperfect, informal, and poorly resourced languages. A promising solution is the use of deep learning models on these code-mixed Roman Urdu and English text. Therefore, the objective of this paper is to perform a sentiment analysis of code-mixed Roman Urdu and English social media text using state-of-the-art deep learning models. Our work is independent of lexical normalization, language dictionary, and code transfer indication. We perform sentiment analysis using Multilingual BERT (mBERT) and XLM-RoBERTa (XLM-R) models. The results reveal that performance of XLM-R model with tuned hyperparameters for code-mixed Roman Urdu and English social media text is better than the mBERT model with F1 score of 71%.
机译:情感分析的态度,观点和对某些问题,产品,个人和组织情绪的计算研究。公司和客户寻求从社交媒体文本的意见作出决定。情感分析越来越智能与人工智能和自然语言处理的进步。随着使用社交媒体的惊人增长,在这些平台上可用的文本量巨大与英文混合的不完善和非正式的语言,如罗马乌尔都语。目前情感分析技术并没有对这些代码混合不完善的,非正式的,资源贫乏的语言精确地执行。一个有希望的解决方案是对这些代码混合罗马乌尔都语和英语文本使用深度学习模型。因此,本文的目的是利用国家的最先进的深学习模型来执行代码混合罗马乌尔都语和英语社交媒体文本的情感分析。我们的工作是独立的词汇规范化,语言词典和代码转移指示。我们执行使用多语种BERT(mBERT)和XLM - 罗伯塔(XLM-R)模型情感分析。结果表明XLM-R模型与代码混合罗马乌尔都语和英语社交媒体的文字是比F1分数的71%的mBERT模型更好地调整超参数,业绩。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号