LSTM TIME AND FREQUENCY RECURRENCE FOR AUTOMATIC SPEECH RECOGNITION

Similar documents
EXPLORING MULTIDIMENSIONAL LSTMS FOR LARGE VOCABULARY ASR

arxiv: v1 [cs.ne] 5 Feb 2014

IMPROVING WIDEBAND SPEECH RECOGNITION USING MIXED-BANDWIDTH TRAINING DATA IN CD-DNN-HMM

Channel Division Multiple Access Based on High UWB Channel Temporal Resolution

BER Performance Analysis of Cognitive Radio Physical Layer over Rayleigh fading Channel

Endpoint Detection using Grid Long Short-Term Memory Networks for Streaming Speech Recognition

Secure Physical Layer Key Generation Schemes: Performance and Information Theoretic Limits

Performance Measures of a UWB Multiple-Access System: DS/CDMA versus TH/PPM

Resource Allocation via Linear Programming for Multi-Source, Multi-Relay Wireless Networks

Learning the Speech Front-end With Raw Waveform CLDNNs

Pulsed RF Signals & Frequency Hoppers Using Real Time Spectrum Analysis

Airborne Ultrasonic Position and Velocity Measurement Using Two Cycles of Linear-Period-Modulated Signal

Minimizing Distribution Cost of Distributed Neural Networks in Wireless Sensor Networks

Improving the Active Power Filter Performance with a Prediction Based Reference Generation

Rateless Codes for the Gaussian Multiple Access Channel

Convolutional Neural Networks for Small-footprint Keyword Spotting

ADAPTIVE ITERATION SCHEME OF TURBO CODE USING HYSTERESIS CONTROL

Fast Hybrid DFT/DCT Architecture for OFDM in Cognitive Radio System

Information Theoretic Radar Waveform Design for Multiple Targets

Rate-Allocation Strategies for Closed-Loop MIMO-OFDM

Satellite Link Layer Performance Using Two Copy SR-ARQ and Its Impact on TCP Traffic

FEATURE COMBINATION AND STACKING OF RECURRENT AND NON-RECURRENT NEURAL NETWORKS FOR LVCSR

THE TRADEOFF BETWEEN DIVERSITY GAIN AND INTERFERENCE SUPPRESSION VIA BEAMFORMING IN

Radial basis function networks for fast contingency ranking

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /GLOCOM.2003.

Audio Effects Emulation with Neural Networks

COMPARATIVE ANALYSIS OF ULTRA WIDEBAND (UWB) IEEE A CHANNEL MODELS FOR nlos PROPAGATION ENVIRONMENTS

Power Control and Transmission Scheduling for Network Utility Maximization in Wireless Networks

A Low Complexity VCS Method for PAPR Reduction in Multicarrier Code Division Multiple Access

Communication Systems

A BAG-OF-FEATURES APPROACH TO ACOUSTIC EVENT DETECTION. Department of Computer Science, TU Dortmund University, Dortmund, Germany

Resource Allocation via Linear Programming for Fractional Cooperation

Recurrent neural networks Modelling sequential data. MLP Lecture 9 Recurrent Networks 1

Joint Optimization of Scheduling and Power Control in Wireless Networks: Multi-Dimensional Modeling and Decomposition

PROPORTIONAL FAIR SCHEDULING OF UPLINK SINGLE-CARRIER FDMA SYSTEMS

An Approach to use Cooperative Car Data in Dynamic OD Matrix

Sparse Beamforming Design for Network MIMO System with Per-Base-Station Backhaul Constraints

Neural Network Acoustic Models for the DARPA RATS Program

Georgia Institute of Technology. simulating the performance of a 32-bit interconnect bus. referenced to non-ideal planes. A transient simulation

Performance of Single User vs. Multiuser Modulation in Wireless Multicarrier (MC) Communications

Joint Beamforming and Power Optimization with Iterative User Clustering for MISO-NOMA Systems

Predicting Eye Fixations using Convolutional Neural Networks

Availability Analysis for Elastic Optical Networks with Multi-path Virtual Concatenation Technique

Generalized constrained energy minimization approach to subpixel target detection for multispectral imagery

Co-channel Interference Suppression Techniques for STBC OFDM System over Doubly Selective Channel

: taking service robots to play soccer

Effect of Estimation Error on Adaptive L-MRC Receiver over Nakagami-m Fading Channels

Cross-layer queuing analysis on multihop relaying networks with adaptive modulation and coding K. Zheng 1 Y. Wang 1 L. Lei 2 W.

STUDY ON AOTF-BASED NEAR-INFRARED SPECTROSCOPY ANALYSIS SYSTEM OF FARM PRODUCE QUALITY

Joint Optimal Power Allocation and Relay Selection with Spatial Diversity in Wireless Relay Networks

arxiv: v2 [cs.cl] 20 Feb 2018

Audio Effects Emulation with Neural Networks

Sparse Channel Estimation Based on Compressed Sensing for Massive MIMO Systems

A Neural Attention Model for Urban Air Quality Inference: Learning the Weights of Monitoring Stations

ARTI: An Adaptive Radio Tomographic Imaging System

Dealing with Link Blockage in mmwave Networks: D2D Relaying or Multi-beam Reflection?

Fusing Noisy Fingerprints with Distance Bounds for Indoor Localization

FOR energy limited data networks, e.g., sensor networks,

Time-domain Techniques in EMI Measuring Receivers. Technical and Standardization Requirements

Neural Network Part 4: Recurrent Neural Networks

Deep Neural Network Architectures for Modulation Classification

SCHEDULING the wireless links and controlling their

Blind Multiuser Detection in Asynchronous DS-CDMA Systems over Nakagami-m Fading Channels

A Novel Method for Doppler and DOD- DOA Jointly Estimation Based on FRFT in Bistatic MIMO Radar System

Copyright 2000 IEEE. IEEE Global Communications Conference (Globecom 2000), November 27 - December 1, 2000, San Francisco, California, USA

Distribution of Path Durations in Mobile Ad-Hoc Networks and Path Selection

A CPW-Fed Printed Monopole Ultra-Wideband Antenna with E-Shaped Notched Band Slot

Google Speech Processing from Mobile to Farfield

AN Ω(D log(n/d)) LOWER BOUND FOR BROADCAST IN RADIO NETWORKS

Fuzzy Model Predictive Control Applied to Piecewise Linear Systems

Recurrent neural networks Modelling sequential data. MLP Lecture 9 / 13 November 2018 Recurrent Neural Networks 1: Modelling sequential data 1

arxiv: v1 [cs.it] 22 Aug 2007

Utility-Proportional Fairness in Wireless Networks

Performance Comparison of Cyclo-stationary Detectors with Matched Filter and Energy Detector M. SAI SINDHURI 1, S. SRI GOWRI 2

Optimal and Suboptimal Finger Selection Algorithms for MMSE Rake Receivers in Impulse Radio Ultra-Wideband Systems 1

Run to Potential: Sweep Coverage in Wireless Sensor Networks

arxiv: v1 [cs.it] 22 Jul 2014

arxiv: v4 [physics.soc-ph] 31 Dec 2013

Recurrent neural networks Modelling sequential data. MLP Lecture 9 Recurrent Neural Networks 1: Modelling sequential data 1

Powerfully simple event analysis software

Path Delay Estimation using Power Supply Transient Signals: A Comparative Study using Fourier and Wavelet Analysis

An Optimization Framework for XOR-Assisted Cooperative Relaying in Cellular Networks

Radar Signal Demixing via Convex Optimization

Multi-stage Amplifiers Prof. Ali M. Niknejad Prof. Rikky Muller

A Comparative Analysis of Image Fusion Techniques for Remote Sensed Images

New Image Restoration Method Based on Multiple Aperture Defocus Images for Microscopic Images

Chapter 15 Other Modifications

Debugging EMI Using a Digital Oscilloscope

FREQUENCY-DOMAIN TURBO EQUALIZATION FOR SINGLE CARRIER MOBILE BROADBAND SYSTEMS. Liang Dong and Yao Zhao

FBMC/OQAM for the Asynchronous Multi-User MIMO Uplink

Energy-efficient Video Streaming from High-speed Trains

Acoustic modelling from the signal domain using CNNs

Analysis, Analysis Practices, and Implications for Modeling and Simulation

Space-Time Focusing Transmission in Ultra-wideband Cooperative Relay Networks

An Efficient Adaptive Filtering for CFA Demosaicking

GRAY CODE FOR GENERATING TREE OF PERMUTATION WITH THREE CYCLES

Iterative Transceiver Design for Opportunistic Interference Alignment in MIMO Interfering Multiple-Access Channels

Marketing tips and templates

Effect of Interfering Users on the Modulation Order and Code Rate for UWB Impulse-Radio Bit-Interleaved Coded M-ary PPM

1860 IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, VOL. 66, NO. 2, FEBRUARY 2017

Transcription:

LSTM TIME AND FREQUENCY RECURRENCE FOR AUTOMATIC SPEECH RECOGNITION Jinyu Li, Abderahman Mohamed, Geoffrey Zweig, and Yifan Gong Microsoft Corporation, One Microsoft Way, Redmond, WA 98052 { jinyi, asamir, gzweig, ygong}@microsoft.com ABSTRACT Long short-term memory (LSTM) recurrent neura networks (RNNs) have recenty shown significant performance improvements over deep feed-forward neura networks (DNNs). A key aspect of these modes is the use of time recurrence, combined with a gating architecture that ameiorates the vanishing gradient probem. Inspired by human spectrogram reading, in this paper we propose an extension to LSTMs that performs the recurrence in frequency as we as in time. This mode first scans the frequency bands to generate a summary of the spectra information, and then uses the output ayer activations as the input to a traditiona time LSTM (T-LSTM). Evauated on a Microsoft short message dictation task, the proposed mode obtained a 3.6% reative word error rate reduction over the T- LSTM. Index Terms LSTM, RNN, time and frequency 1. INTRODUCTION Recenty, significant progress has been made in automatic speech recognition (ASR) thanks to the appication of deep neura networks (DNNs) [1][2][3][4][5][6]. Unike in the 1990s, today s DNN systems often contain tens of miions of parameters and are more powerfu than their counterparts two decades ago [7][8] in modeing speech signas. DNNs, however, ony consider information in a fixed-ength siding window of frames and thus cannot expoit ong-range correations in the signa. Recurrent neura networks (RNNs), on the other hand, can encode sequence history in their interna state, and thus have the potentia to predict phonemes based on a the speech features observed up to the current frame. Unfortunatey, simpe RNNs, depending on the argest eigenvaue of the state-update matrix, may have gradients which either increase or decrease exponentiay over time. Thus, the basic RNN is difficut to train, and in practice can ony mode short-range effects. Long short-term memory (LSTM) RNNs [9][10] were deveoped to overcome these probems. LSTM-RNNs use input, output and forget gates to achieve a network that can maintain state and propagate gradients in a stabe fashion over ong spans of time. These networks have been shown to outperform DNNs on a variety of ASR tasks [11][12][13] [14][15][16]. A previousy proposed LSTMs use a recurrence aong the time axis to mode the tempora patterns of speech signas, and we ca them T-LSTMs in this paper. The main contribution of this paper is the proposa of a two-eve network where the first eve performs recurrence aong the frequency axis, and the second performs time recurrence. We term this the frequency-time LSTM or F-T-LSTM. Our mode is inspired by the way peope read spectrograms. Note that in common practice, og-fiter-bank features are often used as the input to the neura-networkbased acoustic mode [19][20]. In standard systems, the ogfiter-bank features are independent of one-another, i.e. switching the positions of two fiter-banks won t affect the performance of the DNN or LSTM. However, this is not the case when a human reads a spectrogram: a human reies on both patterns that evove on time, and frequency, to predict phonemes. Switching the positions of two fiter-banks wi destroy the frequency-wise patterns. Our mode addresses this phenomenon by expicity modeing the frequency-wise evoution of spectra patterns. Evauated on a Microsoft interna short message dictation task, the proposed F-T- LSTM obtained 3.6% reative word error rate (WER) reduction from the T-LSTM. The rest of the paper is organized as foows. In Section 2, we briefy introduce LSTMs and then we present the proposed mode which combines frequency LSTM and time LSTM in Section 3. We differentiate the proposed method from the convoutiona LSTM DNN (CLDNN) [16] and muti-dimensiona RNN [17][18] in Section 4. Experimenta evauation of the agorithm is provided in Section 5. We summarize our study and draw concusions in Section 6. 2. THE LSTM-RNN An RNN is fundamentay different from the feed-forward DNN in that the RNN does not operate on a fixed window of frames; instead, it maintains a hidden state vector, which is recursivey updated after seeing each time frame. The interna state encodes the history a the way from the beginning of an utterance up to the ast input, and can thus potentiay mode much onger span effects than a fixed-window DNN. In other words, an RNN is a dynamic system and is more genera than the DNN which performs a static input-output transformation. The incusion of interna states enabes RNNs to represent and earn ong-range sequentia dependencies.

However, the simpe RNN suffers from the vanishing/expoding gradient probem [21] when the error signa is back-propagated through time. This probem is we handed in the LSTM-RNNs through the use of the foowing four components: Memory units: these store the tempora state of the network; Input gates: these moduate the input activations into the ces; Output gates: these moduate the output activations of the ces ; Forget gates: these adaptivey reset the ce s memory. Taken together as in Figure 1 beow, these four components are termed a LSTM ce. 3. FREQUENCY-TIME LSTM-RNN In this section, we propose a frequency-time LSTM (F-T- LSTM) which combines frequency LSTM with time LSTM as shown in Figure 2. We first use a frequency LSTM (F- LSTM) to scan the frequency bands so that frequencyevoving information is summarized by the output of the F- LSTM. The formuation of the F-LSTM is the same as that of the T-LSTM except that the index j now stands for frequency steps instead of time steps. Then we can take the outputs from a F-LSTM steps and use them as the input to T-LSTM to do time anaysis in the traditiona way. Time Anaysis Figure 1. Architecture of LSTM-RNNs with one recurrent ayer. Z 1 is a time-deay node. Figure 1 depicts the architecture of an LSTM-RNN with one recurrent ayer. In LSTM-RNNs, in addition to the past hidden-ayer output h t 1, the past memory activation c t 1 is aso an input to the LSTM ce. This mode can be described as: i j = σ(w xi x j + W hi h j 1 + W ci c j 1 + b i ), (1) f j = σ(w xf c j = f j. c j 1 o j = σ(w xo x j + W hf h j 1 + W cf + i j. tanh(w xc x j + W ho h j 1 + W co c j 1 + b f ), (2) x j + W hc h j 1 + b c ), (3) c j + b o ), (4) h j = o j. tanh(c j ), (5) where i j, o j, f j, and c j denote the activation vectors of input gate, output gate, forget gate, and memory ce at the -th ayer and time j, respectivey. h j is the output of the LSTM ces at ayer and time j. W terms denote different weight matrices. For exampe, W xi is the weight matrix from the ce input to the input gate at the -th ayer. b terms are the bias terms (e.g., b i is the bias of input gate at ayer ).. denotes eement wise mutipication. In [13], a LSTM with an additiona projection ayer prior to the output (termed ) was proposed to reduce the computationa compexity of LSTM. A projection ayer is appied to h j as r j = W hr h j And then h j 1 in Eqs (1)--(4) is repaced by r j 1. Frequency Time Frequency Anaysis Figure 2: A frequency- time LSTM-RNN which scans the frequency axis first for frequency anaysis and then scans the time axis for time anaysis. The detaied F-LSTM processing for each time step is described as foows. Divide tota N og-fiter-banks at current time into M overapped chunks and each chunk contains B ogfiter-banks. There are C og-fiter-banks overapped between adjacent chunks. Here we have the reationship M = N C. An extreme case is C=0 B C where there is no overapped og-fiter-bank. In such a case, M = N. B Use the M overapped chunks as the frequency steps of F-LSTM and generate the output of h m, m = 0 M 1.

Merge h m, m = 0 M 1. into a super-vector h which can be considered as a trajectory of frequency patterns at current time. Then use h as the input to a T-LSTM with mutipe ayers. Figure 3 shows an exampe setup of the F-LSTM used in our experiments. The input at each frame consists of a 40 dimensiona vector of og-fiter-bank vaues at the current time t. We divide the 40 og-fiter-bank channes into 33 overapped chunks with each chunk containing 8 og-fiterbanks. This resuts in 7 og-fiter-banks of overap between adjacent chunks (C=7). Therefore, the first F-LSTM ce takes eight inputs: the og-fiter-banks from 0 to 7, and the second F-LSTM ce takes the og-fiter-banks from 1 to 8, and so on. The m-th F-LSTM ce generates outputs h m, which wi be passed into the m+1-th F-LSTM ce. Finay, h m, m = 0 M 1 (M=33 in this exampe) wi be concatenated as the input to a T-LSTM. 4. RELATION TO PRIOR WORK In this section, we first discuss the difference between our proposed F-T-LSTM and the convoutiona LSTM DNN (CLDNN) [16] which combines CNNs, LSTMs, and DNNs together. The CLDNN first uses a CNN [22][23] to reduce the spectra variation, and then the output of the CNN ayer is fed into a muti-ayer LSTM to earn the tempora patterns. Finay, the output of the ast LSTM ayer is fed into severa fuy connected DNN ayers for the purpose of cassification. The key difference between the proposed F-T-LSTM and the CLDNN is that the F-T-LSTM uses frequency recurrence with the F-LSTM, whereas the CLDNN uses a siding convoutiona window for pattern detection with the CNN. Whie the siding window achieves some invariance through shifting, it is not the same as a fuy recurrent network. The two approaches both aim to achieve invariance to input distortions, but the pattern detectors in the CNN maintain a constant dimensionaity, whie the F-LSTM can perform a genera frequency warping. The proposed F-T-LSTM performs 1-D recurrence over the frequency axis and then performs 1-D recurrence over the time axis. This is different from the concept of mutidimensiona processing which has been proved very successfu in the handwriting recognition tasks [17][18] and outperformed the traditiona handwriting systems that use CNNs [22][23] as the feature extractor. To summarize, the T-F-LSTM works on mutidimensiona space separatey with simpicity whie the mutidimensiona RNN [17][18] works jointy on mutidimensiona space with more powerfu modeing. 5. EXPERIMENTS AND DISCISSIONS In this section, we use a Windows Phone short message dictation task to evauate the proposed method. The training data consists of 60 hours of transcribed US-Engish audio. The test set consists of 3 hours of data from the same Windows Phone task. The audio data is 16k HZ samped, recorded in mobie environments using Windows phones. The vocabuary has around 130k words and the LM has around 6.6M ngrams (up to trigram). A experiments were conducted using the computationa network tookit (CNTK) [24], which aows us to buid and evauate various network structures efficienty without deriving and impementing compicated training agorithms. A the modes were trained to minimize the frame-eve cross-entropy criterion. Figure 3: An exampe setup of F-LSTM. The input to the baseine CD-DNN-HMM system consists of 40-dimensiona og-fiter-bank features. We augment these feature vectors with 5 frames of context on either side (5-1-5). The DNN has 5 hidden ayers, each with 2048 sigmoid units. Both the baseine and LSTM systems use 1812 tied-triphone states or senones. The baseine T- is modeed after that in [13]. It has four T- ayers: each has 1024 hidden units and the

output size of each T-LSTM ayer is reduced to 512 using a inear projection ayer. There is no frame stacking, and the output HMM state abe is deayed by 5 frames as in [13]. When training T-, the backpropagation through time (BPTT) [25] step is 20. We buit the F-T-LSTM with a singe F-LSTM that scans the og-fiter-banks and three T- ayers. The number of parameters of the F-T-LSTM is between the numbers of parameters of the three- and four- ayer s. To generate the input to the F-LSTM, we use the exampe setup in Section 3 by dividing the 40 og-fiter-bank channes into 33 overapped chunks with each chunk containing 8 og-fiterbanks. The F-LSTM has 24 memory ces. In Tabe 1, we compare the WERs of a DNN, T-LSTM, and F-T-LSTM. The T-LSTM is ceary better than the DNN due to its tempora modeing power. With both the frequency and tempora modeing, the F-T-LSTM is better than the 4- ayer T-LSTM, with 3.6% reative WER reduction. Tabe 1: WER comparison of DNN, T-LSTM, and F-T- LSTM Mode WER (%) DNN 21.84 3-ayer T- 20.79 4-ayer T- 20.38 F-LSTM (24 ces)+3-ayer T- 19.64 We investigate the impact of different ce numbers in the F-LSTM in Tabe 2. When the number of ces is very sma, e.g., 8, the power of F-LSTM is very imited with ony a sight improvement over the T-LSTM. However, when the number of ces becomes 24, the F-LSTM shows its advantage because the memory ces are powerfu enough to store the frequency patterns. When we increase the number of ces to 48, there is no further improvement. Tabe 2: Impact of ce numbers in F-LSTM Mode WER (%) F-LSTM (8 ces)+3-ayer T- 20.19 F-LSTM (24 ces)+3-ayer T- 19.64 F-LSTM (48 ces)+3-ayer T- 19.81 In a the aforementioned experiments, we have not stacked mutipe frames of og-fiter-banks as the input to F-T-LSTM. This decision is made based on our previous experience with T-LSTMs, where we found that stacking mutipe frame inputs doesn t have any benefit, and [13] aso doesn t have the frame stacking. In Tabe 3, we compare the setup with and without mutipe-frame stacking. Stacking N frames means that every chunk now has 8*N og-fiter-banks. When stacking 11 frames, we predict the center frame s abe. As shown in Tabe 3, it doesn t provide any benefit to WER by stacking 11 frames as the input to F-LSTM. Tabe 3: Comparison of F-T-LSTM with and without stacking frame inputs Mode Number of Input Frames WER (%) F-LSTM (24 ces)+3-1 19.64 ayer T- F-LSTM (24 ces)+3-11 20.08 ayer T- F-LSTM (48 ces)+3-1 19.81 ayer T- F-LSTM (48 ces)+3- ayer T- 11 20.01 6. CONCLUSIONS AND FUTURE WORK In this paper, we have presented a FT--LSTM architecture that scans both the time and frequency axis to mode the evoving patterns of the spectrogram. The F-T-LSTM first uses an F-LSTM to performs a frequency recurrence that summarizes frequency-wise patterns. This is then fed into a T-LSTM. The proposed F-T-LSTM obtained a 3.6% reative WER reduction from the traditiona T-LSTM on a short message dictation task. We have shown that as ong as the number of memory ces in the F-LSTM is reasonabe, the F- T-LSTM can achieve good performance. We aso evauated the impact of stacking mutipe frames as the input to F- LSTM, and found that it is best to simpy present the frames one at a time Severa research issues wi be addressed in the future to further increase the effectiveness of the agorithm presented in this paper. First, we wi compare the performance of F-T- LSTMs with CLDNNs to better understand their reative advantages. Second, we want to expore architectura variants of the F-T-LSTM. For exampe, we wi examine whether frequency overapping of the input to F-LSTM is necessary. Third, we wi move the input of F-LSTM from og-fiterbanks directy to og-spectrum. There are studies showing that directy working of og-spectrum can be beneficia to DNN [26]. By appying the F-LSTM directy on ogspectrum, we can naturay remove the hand-crafted fiterbanks, and automaticay earn the frequency patterns that benefit the recognizer. Fourth, in [27] it is shown that CNNs can consistenty provide advantages over DNNs in mismatched training-test conditions. It is interesting to see whether the frequency recurrence brought by the F-LSTM can be more hepfu in the mismatched conditions. Last and most importanty, we wi advance our study by proposing a mutidimensiona LSTM with a simpified structure which performs recurrence over the time and frequency axes jointy [28]. We term it the time-frequency LSTM (TF-LSTM). We wi compare TF-LSTM and F-T-LSTM in [28] by using a much arger ASR task. It wi be shown that F-T-LSTM is sti effective on that arger ASR task.

REFERENCES [1] F. Seide, G. Li, and D. Yu, Conversationa speech transcription using context-dependent deep neura networks, in Proc. Interspeech, pp. 437-440, 2011. [2] N. Jaity, P. Nguyen, A. Senior, and V. Vanhoucke, An appication of pretrained deep neura networks to arge vocabuary conversationa speech recognition, in Proc. Interspeech, 2012. [3] T. N. Sainath, B. Kingsbury, B. Ramabhadran, P. Fousek, P. Novak, A.-R. Mohamed, Making deep beief networks effective for arge vocabuary continuous speech recognition, in Proc. ASRU, pp. 30-35, 2011. [4] G. E. Dah, D. Yu, L. Deng, and A. Acero, Large vocabuary continuous speech recognition with context-dependent DBN- HMMs, in Proc. ICASSP, pp. 4688-4691, 2011. [5] A. Mohamed, G. E. Dah, and G. Hinton, Acoustic modeing using deep beief networks, IEEE Trans. Audio Speech and Language Process., vo. 20, no. 1, pp. 14-22, Jan. 2012. [6] L. Deng, J. Li, J.-T. Huang et. a. Recent advances in deep earning for speech research at Microsoft, in Proc. ICASSP, 2013. [7] H. Bourard and N. Morgan. Connectionist speech recognition - A Hybrid approach, Kuwer Academic Press, 1994. [8] S. Renas, N. Morgan, H. Bouard, M. Cohen, and H. Franco. Connectionist probabiity estimators in HMM speech recognition, IEEE Transactions on Speech and Audio Processing, vo. 2, no. 1, pp. 161 174, 1994. [9] S. Hochreiter and J. Schmidhuber, Long short-term memory, Neura Computation, vo. 9, no. 8, pp. 1735 1780, 1997. [10] A. Gers, J. Schmidhuber, and F. Cummins. Learning to forget: Continua prediction with LSTM, Neura Computation, vo. 12, no. 10, pp. 2451-2471, 2000. [11] A. Graves, A. Mohamed, G. Hinton. Speech recognition with deep recurrent neura networks, in Proc. ICASSP, 2013. [12] A. Graves, N. Jaity, A. Mohamed. Hybrid speech recognition with deep bidirectiona LSTM, in Proc. ASRU, 2013. [13] H. Sak, A. Senior, F. Beaufays, "Long short-term memory recurrent neura network architectures for arge scae acoustic modeing," in Proc. Interspeech, 2014. [14] H. Sak, O. Vinyas, G. Heigod, A. Senior, E. McDermott, R. Monga, M. Mao, "Sequence discriminative distributed training of ong short-term memory recurrent neura networks," in Proc. Interspeech, 2014. [15] X. Li and X. Wu, Constructing ong short-term memory based deep recurrent neura networks for arge vocabuary speech recognition, in Proc. ICASSP, 2015. [16] T. N. Sainath, O. Vinyas, A. Senior and H. Sak, "Convoutiona, ong short-term memory, fuy connected deep neura networks," in Proc. ICASSP, 2015. [17] A. Graves, S. Fernández, J. Schmidhuber, Muti-dimensiona recurrent neura networks, in ICANN, pp. 549-558, 2007. [18] A. Graves and J. Schmidhuber, Offine handwriting recognition with mutidimensiona recurrent neura networks, Advances in Neura Information Processing Systems, pp. 545-552, 2009. [19] A. Mohamed, G. Hinton, and G. Penn, Understanding how deep beief networks perform acoustic modeing, in Proc. ICASSP, pp. 4273 4276, 2012. [20] J. Li, D. Yu, J. T. Huang, and Y. Gong. "Improving wideband speech recognition using mixed-bandwidth training data in CD-DNN-HMM," Proc. IEEE Spoken Language Technoogy Workshop, pages 131 136, 2012. [21] Y. Bengio, P. Simard, and P. Frasconi. Learning ong-term dependencies with gradient descent is difficut, IEEE Transactions on Neura Networks, vo. 5, no. 2, pp. 157-166, 1994. [22] T. N. Sainath, A. Mohamed, B. Kingsbury and B. Ramabhadran, "Deep convoutiona neura networks for LVCSR," in Proc. ICASSP, 2013. [23] O. Abde-Hamid, A. Mohamed, H. Jiang, L. Deng, G. Penn, and Dong Yu, Convoutiona neura networks for speech recognition, IEEE/ACM Transactions on Audio, Speech, and Language processing, vo. 22, no. 10, pp. 1533-1545, 2014. [24] D. Yu, A. Eversoe, M. Setzer, et. a., "An introduction to computationa networks and the computationa network tookit," Microsoft Technica Report MSR-TR-2014-112, 2014. [25] Jaeger, H. Tutoria on training recurrent neura networks, covering BPPT, RTRL, EKF and the echo state network approach, GMD Report 159, GMD German Nationa Research Institute for Computer Science, 2002. [26] T. N. Sainath, B. Kingsbury, A. Mohamed and B. Ramabhadran, "Learning fiter banks within a deep neura network framework," in Proc. ASRU, 2013. [27] J.-T. Huang, J. Li, and Y. Gong, An anaysis of convoutiona neura networks for speech recognition, in Proc. ICASSP, 2015. [28] J. Li, A. Mohamed, G. Zweig, and Yifan Gong, Exporing mutidimensiona LSTMs for arge vocabuary ASR, submitted to Proc. ICASSP, 2016.