Hi,
I pick up old ones.
Brand, M.E., "Voice Puppetry", ACM SIGGRAPH, ISBN: 0-201-48560-5, pps 21-28 , August 1999
http://www.cs.cmu.edu/~ph/869/papers/Brand-sigg99.pdf
Although this is not Audio-visual one, this paper
(trajectory HMM plus annealing!) is also nice.
http://www.mrl.nyu.edu/publications/style-machines/brand-hertzmann.pdf
Regards,
Junichi Yamagishi
CSTR
On 3 Jul 2009, at 11:14, Heiga ZEN (Byung Ha CHUN) wrote:
Hi,
Girish Malkarnenkar wrote:
I am trying to synthesise visual speech via HTS by replacing the MGC files created with facial parameters. While I would appreciate it if someone who has done a similar thing before could share his experience,
Bailly, G., O. Govokhina, G. Breton, F. Elisei and C. Savariaux (2008). The trainable trajectory formation model TD-HMM parameterized for the LIPS 2008 challenge. Interspeech, Brisbane, Australia.
http://www.icp.inpg.fr/ICP/publis/synthese/_gb/gb_IS08.pdf
Regards,
Heiga ZEN (Byung Ha CHUN)
--
--------------------------
Heiga ZEN (Byung Ha CHUN)
Speech Technology Group
Cambridge Research Lab
Toshiba Research Europe
phone: +44 1223 436975
______________________________________________________________________
This email has been scanned by the MessageLabs Email Security System.
For more information please visit http://www.messagelabs.com/email ______________________________________________________________________--
The University of Edinburgh is a charitable body, registered in
Scotland, with registration number SC005336.