...
首页> 外文期刊>Journal of software >Training MEMM with PSO: A Tool for Part-of-Speech Tagging
【24h】

Training MEMM with PSO: A Tool for Part-of-Speech Tagging

机译:使用PSO训练MEMM:用于词性标记的工具

获取原文
获取原文并翻译 | 示例
           

摘要

Maximum Entropy Markov Models (MEMM) can avoid the assumption of independence in traditional Hidden Markov Models (HMM), and thus take advantage of context information in most text mining tasks. Because the convergence rate of the classic generalized iterative' scaling (CIS) algorithm is too low to be tolerated, researchers proposed a lot of improved methods such as IIS, SCGIS and LBFGS for parameters training in MEMM. However these methods sometimes do not satisfy task requirements in efficiency and robustness. This article modifies the traditional Particle Swarm Optimization (PSO) algorithm by using dynamic global mutation probability (DGMP) to solve the local optimum and infinite loops problems and use the modified PSO in MEMM for estimating the parameters. We introduce the MEMM trained by modified PSO into Chinese Part-of-Speech (POS) tagging, analysis the experimental results and find it has higher convergence rate and accuracy than traditional MEMM.
机译:最大熵马尔可夫模型(MEMM)可以避免传统隐马尔可夫模型(HMM)中的独立性假设,因此可以在大多数文本挖掘任务中利用上下文信息。由于经典的广义迭代缩放算法(CIS)的收敛速度太低而无法容忍,因此研究人员提出了许多改进的方法,如IIS,SCGIS和LBFGS,用于在MEMM中进行参数训练。但是,这些方法有时不能满足效率和鲁棒性方面的任务要求。本文通过使用动态全局变异概率(DGMP)来解决局部最优和无限循环问题,并在MEMM中使用经过修改的PSO来估计参数,从而对传统的粒子群优化(PSO)算法进行了修改。将经过改进的PSO训练的MEMM引入中文词性(POS)标签中,对实验结果进行分析,发现它具有比传统MEMM更高的收敛速度和准确性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号