首页> 外文学位 >Analysis, vocal-tract modeling and automatic detection of vowel nasalization.
【24h】

Analysis, vocal-tract modeling and automatic detection of vowel nasalization.

机译:元音鼻腔化的分析,声道建模和自动检测。

获取原文
获取原文并翻译 | 示例

摘要

The aim of this work is to clearly understand the salient features of nasalization and the sources of acoustic variability in nasalized vowels, and to suggest Acoustic Parameters (APs) for the automatic detection of vowel nasalization based on this knowledge. Possible applications in automatic speech recognition, speech enhancement, speaker recognition and clinical assessment of nasal speech quality have made the detection of vowel nasalization an important problem to study. Although several researchers in the past have found a number of acoustical and perceptual correlates of nasality, automatically extractable APs that work well in a speaker-independent manner are yet to be found. In this study, vocal tract area functions for one American English speaker, recorded using Magnetic Resonance Imaging, were used to simulate and analyze the acoustics of vowel nasalization, and to understand the variability due to velar coupling area, asymmetry of nasal passages, and the paranasal sinuses. Based on this understanding and an extensive survey of past literature, several automatically extractable APs were proposed to distinguish between oral and nasalized vowels. Nine APs with the best discrimination capability were selected from this set through Analysis of Variance. The performance of these APs was tested on several databases with different sampling rates, recording conditions and languages. Accuracies of 96.28%, 77.90% and 69.58% were obtained by using these APs on StoryDB, TIMIT and WS96/97 databases, respectively, in a Support Vector Machine classifier framework. To my knowledge, these results are the best anyone has achieved on this task. These APs were also tested in a cross-language task to distinguish between oral and nasalized vowels in Hindi. An overall accuracy of 63.72% was obtained on this task. Further, the accuracy for phonemically nasalized vowels, 73.40%, was found to be much higher than the accuracy of 53.48% for coarticulatorily nasalized vowels. This result suggests not only that the same APs can be used to capture both phonemic and coarticulatory nasalization, but also that the duration of nasalization is much longer when vowels are phonemically nasalized. This language and category independence is very encouraging since it shows that these APs are really capturing relevant information.
机译:这项工作的目的是清楚地了解鼻音化的显着特征和鼻音元音的声学变异性来源,并根据此知识为自动检测元音鼻音化提出声学参数(APs)。在自动语音识别,语音增强,说话人识别和鼻腔语音质量临床评估中的可能应用使得元音鼻腔化的检测成为需要研究的重要问题。尽管过去有几位研究人员发现了许多鼻部的声学和知觉相关性,但尚未找到以说话者无关的方式很好工作的可自动提取的AP。在这项研究中,使用磁共振成像记录的一位美国英语说话者的声道区域功能被用来模拟和分析元音鼻腔化的声学特性,并了解由于膜耦合区域,鼻道不对称以及鼻旁窦。基于这种理解和对过去文献的广泛研究,提出了几种可自动提取的AP,以区分口头和鼻腔元音。通过方差分析从该集合中选择了9个具有最佳区分能力的AP。这些AP的性能已在具有不同采样率,记录条件和语言的几个数据库上进行了测试。在Support Vector Machine分类器框架中,通过在StoryDB,TIMIT和WS96 / 97数据库上使用这些AP分别获得96.28%,77.90%和69.58%的精度。据我所知,这些结果是任何人在这项任务上取得的最好成绩。还对这些AP进行了跨语言测试,以区分北印度语中的口服元音和鼻音元音。此任务的总体准确度为63.72%。此外,发现经语音处理的鼻音元音的准确度为73.40%,远高于针对发音的鼻腔处理的元音的准确度53.48%。该结果不仅表明可以使用相同的AP来捕获音位和共发音的鼻音,而且当元音被音韵化时,鼻音的持续时间要长得多。这种语言和类别独立性非常令人鼓舞,因为它表明这些AP确实在捕获相关信息。

著录项

  • 作者

    Pruthi, Tarun.;

  • 作者单位

    University of Maryland, College Park.;

  • 授予单位 University of Maryland, College Park.;
  • 学科 Language Linguistics.; Speech Communication.; Engineering Electronics and Electrical.
  • 学位 Ph.D.
  • 年度 2007
  • 页码 215 p.
  • 总页数 215
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 语言学;语言学;无线电电子学、电信技术;
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号