首页> 外文期刊>IEEE Transactions on Circuits and Systems for Video Technology >Model-aided coding: a new approach to incorporate facial animationinto motion-compensated video coding
【24h】

Model-aided coding: a new approach to incorporate facial animationinto motion-compensated video coding

机译:模型辅助编码:一种将面部动画合并到运动补偿视频编码中的新方法

获取原文
获取原文并翻译 | 示例
           

摘要

We show that traditional waveform coding and 3-D model-based coding are not competing alternatives, but should be combined to support and complement each other. Both approaches are combined such that the generality of waveform coding and the efficiency of 3-D model-based coding are available where needed. The combination is achieved by providing the block-based video coder with a second reference frame for prediction, which is synthesized by the model-based coder. The model-based coder uses a parameterized 3-D head model, specifying the shape and color of a person. We therefore restrict our investigations to typical videotelephony scenarios that show head-and-shoulder scenes. Motion and deformation of the 3-D head model constitute facial expressions which are represented by facial animation parameters (FAPs) based on the MPEG-4 standard. An intensity gradient-based approach that exploits the 3-D model information is used to estimate the FAPs, as well as illumination parameters, that describe changes of the brightness in the scene. Model failures and objects that are not known at the decoder are handled by standard block-based motion-compensated prediction, which is not restricted to a special scene content, but results in lower coding efficiency. A Lagrangian approach is employed to determine the most efficient prediction for each block from either the synthesized model frame or the previous decoded frame. Experiments on five video sequences show that bit rate savings of about 35% are achieved at equal average peak signal-to-noise ratio (PSNR) when comparing the model-aided codec to TMN-10, the state-of-the-art test model of the M.263 standard. This corresponds to a gain of 2-3 dB in PSNR when encoding at the same average bit rate
机译:我们表明,传统的波形编码和基于3D模型的编码不是竞争性的选择,而应结合起来相互支持和互补。两种方法相结合,以便在需要时可以使用波形编码的通用性和基于3-D模型的编码效率。通过向基于块的视频编码器提供用于预测的第二参考帧来实现该组合,该第二参考帧由基于模型的编码器合成。基于模型的编码器使用参数化的3-D头部模型,用于指定人的形状和颜色。因此,我们的调查仅限于显示头肩场景的典型电视电话场景。 3-D头部模型的运动和变形构成了基于MPEG-4标准的由面部动画参数(FAP)表示的面部表情。利用基于3-D模型信息的基于强度梯度的方法来估计FAP以及描述场景中亮度变化的照明参数。解码器未知的模型故障和对象由基于标准块的运动补偿预测处理,该预测不限于特殊的场景内容,但会导致编码效率降低。拉格朗日方法用于根据合成模型帧或先前解码帧为每个块确定最有效的预测。在五个视频序列上进行的实验表明,在将模型辅助编解码器与最新技术测试TMN-10进行比较时,在平均峰值信噪比(PSNR)相等的情况下,可以节省约35%的比特率M.263标准的模型。当以相同的平均比特率编码时,这对应于PSNR的2-3 dB增益

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号