[hts-users:00769] Re: Questions on training and flat pitch pattern
- Subject: [hts-users:00769] Re: Questions on training and flat pitch pattern
- From: "Alexander Gutkin" <alexander.gutkin@xxxxxxxxx>
- Date: Tue, 7 Aug 2007 22:22:34 +0100
- Cc: "Junichi Yamagishi" <jyamagis@xxxxxxxxxxxx>
- Delivered-to: hts-users@xxxxxxxxxxxxxxx
- Dkim-signature: a=rsa-sha1; c=relaxed/relaxed; d=gmail.com; s=beta; h=domainkey-signature:received:received:message-id:date:from:to:subject:cc:in-reply-to:mime-version:content-type:content-transfer-encoding:content-disposition:references; b=DhPQNV9Z6bRk752YWMg3p/eVeBalQ6TrAB7b08Vmj0cohCqzWdGTAUiQJgiSQb8pSZzf0UXXWf5KqF6TQezdCdvrYMPIXMmh7EZvQLykk/dI0t5SgQZBcaVlnoIADjBSmusylhy1d7dZOuv7s0XO2uadoNEDvppXCHvqdV3g4Wg=
- Domainkey-signature: a=rsa-sha1; c=nofws; d=gmail.com; s=beta; h=received:message-id:date:from:to:subject:cc:in-reply-to:mime-version:content-type:content-transfer-encoding:content-disposition:references; b=WI1MjPVe0TFUvglGlFvw0mNDu1ahZmYLM0R/wBo4hjs/7Jp6qQEShnmy36TUMoOxPrfBCVWX0jPXV3CD4atZTOUnupdj2NSnFc4hS/inTEd3/xgjheQ9FqbCmudmqXIl5OTCs6gZK0mEDjo+a6Mz0RXgP24fe++ryoz66K42ysQ=
On 8/7/07, Nickolay V. Shmyrev <nshmyrev@xxxxxxxxx> wrote:
> В Втр, 07/08/2007 в 16:59 +0100, Junichi Yamagishi пишет:
> > Hi,
> >
> > On 2007/08/07, at 16:45, Heiga ZEN (Byung Ha CHUN) wrote:
> >
> > > Lee Sillon wrote (2007/08/08 0:10):
> > >
> > >> 1. If the training corpus is larger (for example,10,000
> > >> sentences), the training process will crash because of memory
> > >> consuming(about 2GB). Could it be solved?
> > >
> > > I guess tree-based clustering (HHEd) consumes huge memory for
> > > larger training data. You can use low-memory implementation of tree-
> > > based clustering by specifying -r option.
> >
> > I've trained HMMs for HMM-based speech synthesis from 23,000 sentences.
>
> Oh, so much, how many hours is it? I don't believe it had manual
> segmentation :)
> Did you noticed any real improvement with so big database, for example
> Alan says that 500-1000 are enough.
>
I assume this was used by Junichi was the adaptation work, i.e.
training the speaker-independent models. But, in general, I can't see
anything wrong with training on a large database - as long as you can
afford the resources... Probably there is no perfect recepy for the
optimal database size - it's too speaker and coverage-specific...
- Follow-Ups
-
- [hts-users:00771] Re: Questions on training and flat pitch pattern, Junichi Yamagishi
- References
-
- [hts-users:00764] Questions on training and flat pitch pattern, Lee Sillon
- [hts-users:00765] Re: Questions on training and flat pitch pattern, Heiga ZEN (Byung Ha CHUN)
- [hts-users:00766] Re: Questions on training and flat pitch pattern, Junichi Yamagishi
- [hts-users:00767] Re: Questions on training and flat pitch pattern, Nickolay V. Shmyrev