Pretrain decoder with large amount of text data. #1753
Replies: 3 comments
-
We have not tried that. However, people from Google have tried it. Please see the following paper |
Beta Was this translation helpful? Give feedback.
-
I think finding a way to pretrain or jointly train the encoder would be more useful. |
Beta Was this translation helpful? Give feedback.
-
Thank you for these answers. |
Beta Was this translation helpful? Give feedback.
-
Hello, we've been working with K2 for almost a year now and we've developed systems for around fifteen languages thanks to it.
The performance is consistently better than our old TDNN-f trained with Kaldi systems.
The only problem is that K2's RNN-T models are very bad at spelling proper nouns in particular, and this is not surprising since these models only use transcriptions to train the decoder.
Having seen that LODR rescoring is not very effective, my question is, have you ever tried to pretrain the decoder with large amounts of text from newspapers or other sources, and then freezing or finetuning the decoder while training the models?
Beta Was this translation helpful? Give feedback.
All reactions