Professional Documents
Culture Documents
Gaofeng Cheng1,3 , Vijayaditya Peddinti4,5 , Daniel Povey 4,5 , Vimal Manohar 4,5 ,
Sanjeev Khudanpur4,5 ,Yonghong Yan1,2,3
1 Key Laboratory of Speech Acoustics and Content Understanding, Institute of Acoustics,
2 Xinjiang Laboratory of Minority Speech and Language Information Processing, Xinjiang Technical Institute of Physics and Chemistry,
Chinese Academy of Sciences, China
3 University of Chinese Academy of Sciences
4 Center of Language and Speech Processing
5 Human Language Technology Center Of Excellence,
The Johns Hopkins University, Baltimore, MD, USA
{gfcheng.cn}@gmail.com,{vijay.p,vmanoha1,khudanpur}@jhu.edu,yanyonghong@hccl.ioa.ac.cn
In this section we introduce some terminology to describe the output gate input
various forms of dropout that we experimented with.
input
3.1. Per-frame versus per-element dropout Location 4 recurrent
One of the modifications to dropout that we tried is per-frame recurrent forget gate Location 4
dropout. Unlike per-element dropout in which each element of Input gate input
the dropout mask is chosen independently, in per-frame dropout
the entire vector is set to either zero or one. However, this is
done separately for each individual part of the network, e.g. if
dropout is used for multiple layers or gates then each would
input recurrent
have an independently chosen dropout mask.
3.2. Dropout schedules Figure 1: Dropout location of LSTMP based RNNs architec-
We experimented with various dropout schedules, and settled ture with peepholes[2] and LSTM projection[4]. Single memory
on schedules in which the dropout probability p is zero for a block is shown for clarity and simplicity.
time at the start of training and at the end, and rises to a peak
somewhere closer to the start of training. Bear in mind that
in the Kaldi nnet3-based recipes we use here, the number of • Location 4 (dropout on i, f and o): replace equation (1),
epochs is determined in advance (early stopping is not used). (2), (4) with:
(it )
Therefore we are able to specify dropout schedules that are ex- it = σ(Wix xt + Wir rt−1 + Wic ct−1 + bi ) mdrop
pressed relative to the total number of epochs we will use. We t
ft = σ(Wf x xt + Wf r rt−1 + Wf c ct−1 + bf ) mdrop
(f )
express the dropout schedule as a piecewise linear function on (ot )
the interval [0, 1], where f (0) gives the dropout proportion at ot = σ(Wox xt + Wor rt−1 + Woc ct + bo ) mdrop
the start of training and f (1) gives the dropout proportion after • Location 5 (dropout in the recurrence only): replace
seeing all the data. As an example of our notation, the sched- equation (7) with:
ule ’0@0,0.2@0.4,0@1’ means a function f (x) that is linearly (rt )
rt = (Wrm mt ) mdrop
interpolated between the points f (0) = 0, f (0.4) = 0.2, and
f (1) = 0. We write this as ’0, 0.2@0.4, 1’ as a shorthand, be- As alternatives to Location 4, we also tried dropout on all
cause the first point is always for x = 0 and the last is always possible subsets of the i, f , o gates (experiments not shown),
for x = 1. and found that dropping out all 3 was the best. We didn’t try
After experimenting with various schedules, we generally dropping out the output of either of the tanh gates, because it
recommend schedules of the form ’0,0@0.2,p@0.5,0’, where would have the same effect as dropping out the sigmoid gate
p is a constant (e.g. 0.3, 0.5 or 0.7) that is different in differ- that each one is multiplied by. In the rest of the paper, when we
ent models. Because these schedules end at zero, there is no talk about per-frame dropout, we refer to the per-frame dropout
need to do the normal trick of rescaling the parameters before location 4.
using the models for inference. The optimal p for BLSTMPs is
0.15∼0.10, we use 0.10 here. 4. Experimental Setup
3.3. LSTMP with dropout In this paper, HMM-DNN hybrid neural network acoustic mod-
els are used. The nnet3 toolkit in Kaldi speech recognition
Some of the locations in the LSTMP layer where we tried toolkit [9] is used to perform neural network training. The dis-
putting dropout are as follows (see Figure 1 for illustration): tributed neural network training algorithm is described in [10].
• Location 1 (dropout before the projection): replace equa- LF-MMI [6] is the training criterion.
tion (5) with: 40-dimensional Mel-frequency cepstral coefficients
mt = (ot tanh(ct )) mdrop
(mt ) (MFCCs) without cepstral truncation are used as the input into
the neural network [10][11]. These 40-dimensional features are
• Location 2 (dropout on the output of the layer): replace spliced across ±n (n may be 1 or 2) frames of context, then
equation (8) with: appended with a 100-dimensional i-vectors [12] to perform
(yt )
yt = (pt , rt ) mdrop instantaneous adaptation of the neural network [13]. These
i-vectors contain information about the mean offset of the
• Location 3 (dropout on each half of the projection out-
speaker’s data, so cepstral normalization is not necessary.
put): replace equation (6), (7) with:
(pt ) In all cases we use the speed-perturbation method described
pt = (Wpm mt ) mdrop in [14] to augment the data 3-fold. Our AMI setup is similar
t (r )
rt = (Wrm mt ) mdrop to that described in [15]; one notable feature is that alignments
Table 1: Model configuration for TDNN-LSTMPs in SWBD, Table 3: Dropout results on AMI SDM, dropout schedule is
Tedlium and AMI ’0,0@0.20,p@0.5,0@0.75,0’
Figure 2: Diagram of our TDNN-LSTMP configuration. The in- With these configurations we observed relative WER reductions
put context is {-2,-1,0,1,2}, and the splicing indexes for layer2, of between 3% and 7% on a range of LVCSR datasets. Our
layer4 and layer5 are {-1,1}, {-3,0,3}, {-3,0,3}; LSTM time conclusions are a little different from [8], which recommends
delay is 3. dropout only for the forward (non-recurrent) connections, like
our Location 2; but we should point out that [8] did not report
some very minor differences in how the diagonal matrices of any experimental comparisons with alternative ways of doing
the LSTM are trained. The general conclusion that we draw is dropout.
that our recipe works consistently using both LSTM implemen-
tations and a range of datasets. Of course we cannot guarantee 7. References
that this would work in a setup that was very different from [1] S. Hochreiter and J. Schmidhuber, “Long short-term memory,”
ours. For instance, our LSTMs generally operate with a recur- Neural computation, vol. 9, no. 8, pp. 1735–1780, 1997.
rence that spans 3 time steps instead of 1.
[2] F. A. Gers, N. N. Schraudolph, and J. Schmidhuber, “Learning
We have not shown our experiments where we tuned the precise timing with lstm recurrent networks,” Journal of machine
dropout schedule. We did find that it was important to leave the learning research, vol. 3, pp. 115–143, 2002.
[3] A. Graves, A.-r. Mohamed, and G. Hinton, “Speech recognition
with deep recurrent neural networks,” in 2013 IEEE international
conference on acoustics, speech and signal processing. IEEE,
2013, pp. 6645–6649.
[4] H. Sak, A. W. Senior, and F. Beaufays, “Long short-term memory
based recurrent neural network architectures for large vocabulary
speech recognition,” CoRR, vol. abs/1402.1128, pp. 157–166,
2014. [Online]. Available: http://arxiv.org/abs/1402.1128
[5] N. Srivastava, G. E. Hinton, A. Krizhevsky, I. Sutskever, and
R. Salakhutdinov, “Dropout: a simple way to prevent neural net-
works from overfitting.” Journal of Machine Learning Research,
vol. 15, no. 1, pp. 1929–1958, 2014.
[6] D. Povey, V. Peddinti, D. Galvez, P. Ghahrmani, V. Manohar,
X. Na, Y. Wang, and S. Khudanpur, “Purely sequence-trained neu-
ral networks for asr based on lattice-free mmi,” Interspeech, 2016.
[7] S. J. Rennie, P. L. Dognin, X. Cui, and V. Goel, “Annealed dropout
trained maxout networks for improved lvcsr,” in Acoustics, Speech
and Signal Processing (ICASSP), 2015 IEEE International Con-
ference on. IEEE, 2015, pp. 5181–5185.
[8] W. Zaremba, I. Sutskever, and O. Vinyals, “Recurrent neural
network regularization,” CoRR, vol. abs/1409.2329, 2014.
[Online]. Available: http://arxiv.org/abs/1409.2329
[9] D. Povey, A. Ghoshal, G. Boulianne, L. Burget, O. Glembek,
N. Goel, M. Hannemann, P. Motlicek, Y. Qian, P. Schwarz et al.,
“The kaldi speech recognition toolkit,” in IEEE 2011 workshop
on automatic speech recognition and understanding, no. EPFL-
CONF-192584, 2011.
[10] D. Povey, X. Zhang, and S. Khudanpur, “Parallel training
of deep neural networks with natural gradient and parameter
averaging,” CoRR, vol. abs/1410.7455, 2014. [Online]. Available:
http://arxiv.org/abs/1410.7455
[11] V. Peddinti, D. Povey, and S. Khudanpur, “A time delay neural
network architecture for efficient modeling of long temporal con-
texts,” in Proceedings of INTERSPEECH, 2015, pp. 2440–2444.
[12] N. Dehak, P. J. Kenny, R. Dehak, P. Dumouchel, and P. Ouellet,
“Front-end factor analysis for speaker verification,” IEEE Trans-
actions on Audio, Speech, and Language Processing, vol. 19,
no. 4, pp. 788–798, May 2011.
[13] G. Saon, H. Soltau, D. Nahamoo, and M. Picheny, “Speaker adap-
tation of neural network acoustic models using i-vectors,” in 2013
IEEE Workshop on Automatic Speech Recognition and Under-
standing, 2013.
[14] T. Ko, V. Peddinti, D. Povey, and S. Khudanpur, “Audio augmen-
tation for speech recognition,” in INTERSPEECH, 2015.
[15] V. Peddinti, V. Manohar, Y. Wang, D. Povey, and S. Khudan-
pur, “Far-field asr without parallel data,” in Proceedings of In-
terspeech, 2016.
[16] T. N. Sainath, O. Vinyals, A. Senior, and H. Sak, “Convolutional,
long short-term memory, fully connected deep neural networks,”
in 2015 IEEE International Conference on Acoustics, Speech and
Signal Processing (ICASSP), 2015, pp. 4580–4584.
[17] L. Deng and J. Platt, “Ensemble deep learning for
speech recognition,” in 2015 IEEE International Conference
on Acoustics, Speech and Signal Processing (ICASSP),
2014. [Online]. Available: https://www.microsoft.com/en-
us/research/publication/ensemble-deep-learning-for-speech-
recognition/