发明名称 Method for generating realistic facial animation directly from speech utilizing hidden markov models
摘要 <p>A system for learning a mapping between time-varying signals is used to drive facial animation directly from speech, without laborious voice track analysis. The system learns dynamical models of facial and vocal action from observations of a face and the facial gestures made while speaking. Instead of depending on heuristic intermediate representations such as phonemes or visemes, the system trains hidden Markov models to obtain its own optimal representation of vocal and facial action. An entropy-minimizing training technique using an entropic prior ensures that these models contain sufficient dynamical information to synthesize realistic facial motion to accompany new vocal performances. In addition, they can make optimal use of context to handle ambiguity and relatively long-lasting facial co-articulation effects. The output of the system is a sequence of facial control parameters suitable for driving a variety of different kinds of animation ranging from warped photorealistic images to 3D cartoon characters. <IMAGE></p>
申请公布号 EP0992933(A2) 申请公布日期 2000.04.12
申请号 EP19990107625 申请日期 1999.04.16
申请人 MITSUBISHI DENKI KABUSHIKI KAISHA 发明人 BRAND, MATTHEW E.
分类号 G10L15/00;G06K9/00;G06K9/62;G06T13/20;G06T13/40;G10L21/06;(IPC1-7):G06K9/00;G06K9/68 主分类号 G10L15/00
代理机构 代理人
主权项
地址