[mary-users] RES: MaryTTS Viseme data

Jose Carlos de Oliveira oliveirakol at gmail.com
Sun Apr 16 15:58:07 CEST 2017


Sorry!

 

I desired to redirect your message to my development group and made it wrong.

 

By the way, your PhD thesis is very good.

 

Here in Brazil we are working with MaryTTS since 2016 because and other TTS´s for few years.

The reason is that we developed a software called RYBENA to help Brazilian with have visual or hearth problems. You can see it in (Brazilian) site www.rybena.com.br 

For years we are using a very expansive TTS software for (Brazilian Portuguese) Developed by CPqD. 

So we are developing our own TTS using MaryTTS, because we found other with didn’t gave good voice sintetisys.

 

Regards, 

 

Oliveira/Jose Carlos de

Pessoal

Brasilia - DF - Lago Norte

SHIN CA05 CJM02 AP309

   cel:+5561-99311.9226

 

De: idoor [mailto:idoorlab88 at gmail.com] 
Enviada em: domingo, 16 de abril de 2017 10:28
Para: Jose Carlos de Oliveira
Cc: Joan Pere Sanchez; mary-users at dfki.de
Assunto: Re: [mary-users] MaryTTS Viseme data

 

Sorry, I do not know what is "Pode ajudar", is it Spanish?

 

Best regards

 

On Sun, Apr 16, 2017 at 9:21 AM, Jose Carlos de Oliveira <oliveirakol at gmail.com> wrote:

Pode ajudar

 

 

Oliveira/Jose Carlos de

Pessoal

Brasilia - DF - Lago Norte

SHIN CA05 CJM02 AP309

   cel:+5561-99311.9226 <tel:+55%2061%2099311-9226> 

 

De: mary-users-bounces at dfki.de [mailto:mary-users-bounces at dfki.de] Em nome de Joan Pere Sanchez
Enviada em: sábado, 15 de abril de 2017 15:05
Para: idoor Du
Cc: mary-users at dfki.de
Assunto: Re: [mary-users] MaryTTS Viseme data

 

Hi Dave,

This task is the main goal of my PhD thesis. I'm doing lip-sync from the input text over the time duration estimation done while the speech is generated. You can develop your own strategy for lip/mouth synchronization, but often this is an avatar (or interface -I'm using a talking head too-) dependent task. So, if you are using an avatar, it depends if you can use blend shapes to mix by interpolation from the initial pose to the next one. Most of MPEG-4 systems are able to do that automatically.

On one hand, you have each phoneme and their start and finish time. On the other hand, you can adjust a set of visemes for each basic expression (no more than 15 are needed) and then choose the sequence corresponding to each word you are generating. It's the more efficient and simple way to have an effective lip synchronization.

Don't hesitate to contact me if you want more info or refs about.

Bes regards,

 

2017-04-15 18:27 GMT+02:00 idoor Du <idoorlab88 at gmail.com>:

Hi all,

 

I am new to MaryTTS, tried to call its API via:

 

AudioInputStream audio = mary.generateAudio("testing");

 

Now I want to animate mouth/lip shapes at runtime based on the audio sound, how to achieve that? are there any viseme data associated with the audio?

 

Thanks in advance.

 

Dave


_______________________________________________
Mary-users mailing list
Mary-users at dfki.de
http://www.dfki.de/mailman/cgi-bin/listinfo/mary-users




-- 

Joan Pere Sànchez Pellicer

kaiserjp at gmail.com

www.chamaleon.net
+34 625 012 741 <tel:+34%20625%2001%2027%2041> 

 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://www.dfki.de/pipermail/mary-users/attachments/20170416/12345a12/attachment-0001.htm 


More information about the Mary-users mailing list