|
From: Daniel P. <dp...@gm...> - 2015-07-02 18:25:53
|
>
> Back to training on the Blizzard dataset, I was able to dump the iVectors
> for Blizzard's 19-hour subset. Where are they needed, though? Neither
> train_more2.sh nor get_egs2.sh seem to accept dumped iVectors as input.
It's the --online-ivector-dir option.
> Regardless, I ran the train_more2.sh script on Blizzard's data/ and egs/
> folder (generated with get_egs2.sh), and I get the following errors in
> train.*.*.log:
>
> KALDI_ASSERT: at nnet-train-parallel:FormatNnetInput:nnet-update.cc:212,
> failed:
> data[0].input_frames.NumRows() >= num_splice
> [...]
> LOG (nnet-train-parallel:DoBackprop():nnet-update.cc:275) Error doing
> backprop, nnet info is: num-components 17
> num-updatable-components 5
> left-context 7
> right-context 7
> input-dim 140
> output-dim 5816
> parameter-dim 10351000
> [...]
>
> The logs tell me that the left and right contexts were set to 7. However, I
> specified them both to 3 when running get_egs2.sh. The
> egs/info/{left,right}_context files even confirm that they are set to 3. Is
> it possible that train_more2.sh is using the contexts from another
> directory?
The problem is that 3 < 7. The neural net requires a certain amount
of temporal context (7 left and right, here) and if you dump less than
that in the egs it will crash. So you need to set them to 7 when
dumping egs.
Dan
> On Tue, Jun 30, 2015 at 2:07 PM, Daniel Povey <dp...@gm...> wrote:
>>
>> Check the script that generated it; probably the graph directory was
>> in a different location e.g. in tri6 or something like that.
>> Hopefully we would have uploaded that too.
>> We only need to regenerate the graph when the tree changes.
>> Dan
>>
>>
>> On Tue, Jun 30, 2015 at 2:05 PM, Mate Andre <ele...@gm...> wrote:
>> > To ensure that the nnet_a_online model is performing well on the 19-hour
>> > Blizzard dataset and that it is producing correct alignments, I want to
>> > run
>> > the decoding script on the Blizzard data. However, the nnet_a_online
>> > model
>> > on kadi-asr.org doesn't seem to have a graph directory needed for
>> > decoding.
>> > Is there any way I can get a hold of this directory without training the
>> > entire model?
>
>
|