Neural Network Acoustic Models for the DARPA RATS Program

Similar documents
IMPROVING WIDEBAND SPEECH RECOGNITION USING MIXED-BANDWIDTH TRAINING DATA IN CD-DNN-HMM

IMPROVEMENTS TO THE IBM SPEECH ACTIVITY DETECTION SYSTEM FOR THE DARPA RATS PROGRAM

Convolutional Neural Networks for Small-footprint Keyword Spotting

Using RASTA in task independent TANDEM feature extraction

Acoustic modelling from the signal domain using CNNs

Progress in the BBN Keyword Search System for the DARPA RATS Program

Learning the Speech Front-end With Raw Waveform CLDNNs

The 2010 CMU GALE Speech-to-Text System

Discriminative Training for Automatic Speech Recognition

FEATURE FUSION FOR HIGH-ACCURACY KEYWORD SPOTTING

Acoustic Modeling from Frequency-Domain Representations of Speech

DISTANT speech recognition (DSR) [1] is a challenging

Evaluating robust features on Deep Neural Networks for speech recognition in noisy and channel mismatched conditions

Fusion Strategies for Robust Speech Recognition and Keyword Spotting for Channel- and Noise-Degraded Speech

Endpoint Detection using Grid Long Short-Term Memory Networks for Streaming Speech Recognition

FEATURE COMBINATION AND STACKING OF RECURRENT AND NON-RECURRENT NEURAL NETWORKS FOR LVCSR

Audio Augmentation for Speech Recognition

Auditory motivated front-end for noisy speech using spectro-temporal modulation filtering

Recurrent neural networks Modelling sequential data. MLP Lecture 9 Recurrent Neural Networks 1: Modelling sequential data 1

FEATURE FUSION FOR HIGH-ACCURACY KEYWORD SPOTTING

Recurrent neural networks Modelling sequential data. MLP Lecture 9 / 13 November 2018 Recurrent Neural Networks 1: Modelling sequential data 1

TIME-FREQUENCY CONVOLUTIONAL NETWORKS FOR ROBUST SPEECH RECOGNITION. Vikramjit Mitra, Horacio Franco

Training neural network acoustic models on (multichannel) waveforms

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 8, NOVEMBER

DEEP ORDER STATISTIC NETWORKS. Steven J. Rennie, Vaibhava Goel, and Samuel Thomas

SONG RETRIEVAL SYSTEM USING HIDDEN MARKOV MODELS

arxiv: v2 [cs.cl] 20 Feb 2018

I D I A P. Hierarchical and Parallel Processing of Modulation Spectrum for ASR applications Fabio Valente a and Hynek Hermansky a

An Investigation on the Use of i-vectors for Robust ASR

ACOUSTIC SCENE CLASSIFICATION USING CONVOLUTIONAL NEURAL NETWORKS

Artificial Bandwidth Extension Using Deep Neural Networks for Spectral Envelope Estimation

Hierarchical and parallel processing of auditory and modulation frequencies for automatic speech recognition

MEDIUM-DURATION MODULATION CEPSTRAL FEATURE FOR ROBUST SPEECH RECOGNITION. Vikramjit Mitra, Horacio Franco, Martin Graciarena, Dimitra Vergyri

Reverse Correlation for analyzing MLP Posterior Features in ASR

CP-JKU SUBMISSIONS FOR DCASE-2016: A HYBRID APPROACH USING BINAURAL I-VECTORS AND DEEP CONVOLUTIONAL NEURAL NETWORKS

INTRODUCTION TO DEEP LEARNING. Steve Tjoa June 2013

Introduction to Machine Learning

IMPACT OF DEEP MLP ARCHITECTURE ON DIFFERENT ACOUSTIC MODELING TECHNIQUES FOR UNDER-RESOURCED SPEECH RECOGNITION

Learning to Unlearn and Relearn Speech Signal Processing using Neural Networks: current and future perspectives

Performance study of Text-independent Speaker identification system using MFCC & IMFCC for Telephone and Microphone Speeches

Investigating Modulation Spectrogram Features for Deep Neural Network-based Automatic Speech Recognition

Dimension Reduction of the Modulation Spectrogram for Speaker Verification

Voice Activity Detection

Introduction to Machine Learning

DERIVATION OF TRAPS IN AUDITORY DOMAIN

arxiv: v1 [cs.ne] 5 Feb 2014

Feature Extraction Using 2-D Autoregressive Models For Speaker Recognition

Applications of Music Processing

Automatic Morse Code Recognition Under Low SNR

Signal Analysis Using Autoregressive Models of Amplitude Modulation. Sriram Ganapathy

Mikko Myllymäki and Tuomas Virtanen

AN ANALYSIS OF SPEECH RECOGNITION PERFORMANCE BASED UPON NETWORK LAYERS AND TRANSFER FUNCTIONS

The Automatic Classification Problem. Perceptrons, SVMs, and Friends: Some Discriminative Models for Classification

IMPROVING MICROPHONE ARRAY SPEECH RECOGNITION WITH COCHLEAR IMPLANT-LIKE SPECTRALLY REDUCED SPEECH

Augmenting Short-term Cepstral Features with Long-term Discriminative Features for Speaker Verification of Telephone Data

An Adaptive Multi-Band System for Low Power Voice Command Recognition

Deep learning architectures for music audio classification: a personal (re)view

Dimension Reduction of the Modulation Spectrogram for Speaker Verification

Google Speech Processing from Mobile to Farfield

All for One: Feature Combination for Highly Channel-Degraded Speech Activity Detection

I D I A P. On Factorizing Spectral Dynamics for Robust Speech Recognition R E S E A R C H R E P O R T. Iain McCowan a Hemant Misra a,b

Filterbank Learning for Deep Neural Network Based Polyphonic Sound Event Detection

Experiments with Noise Reduction Neural Networks for Robust Speech Recognition

Improving reverberant speech separation with binaural cues using temporal context and convolutional neural networks

PLP 2 Autoregressive modeling of auditory-like 2-D spectro-temporal patterns

Relative phase information for detecting human speech and spoofed speech

High-speed Noise Cancellation with Microphone Array

Distance Estimation and Localization of Sound Sources in Reverberant Conditions using Deep Neural Networks

Joint recognition and direction-of-arrival estimation of simultaneous meetingroom acoustic events

A New Framework for Supervised Speech Enhancement in the Time Domain

Convolutional Networks for Images, Speech, and. Time-Series. 101 Crawfords Corner Road Operationnelle, Universite de Montreal,

IBM SPSS Neural Networks

Research on Hand Gesture Recognition Using Convolutional Neural Network

Experiments on Deep Learning for Speech Denoising

RASTA-PLP SPEECH ANALYSIS. Aruna Bayya. Phil Kohn y TR December 1991

Cheap, Fast and Good Enough: Speech Transcription with Mechanical Turk. Scott Novotney and Chris Callison-Burch 04/02/10

Mel Spectrum Analysis of Speech Recognition using Single Microphone

arxiv: v2 [cs.sd] 31 Oct 2017

POSSIBLY the most noticeable difference when performing

Speech Synthesis using Mel-Cepstral Coefficient Feature

AN IMPROVED NEURAL NETWORK-BASED DECODER SCHEME FOR SYSTEMATIC CONVOLUTIONAL CODE. A Thesis by. Andrew J. Zerngast

Predicting outcomes of professional DotA 2 matches

SPEECH ENHANCEMENT USING A ROBUST KALMAN FILTER POST-PROCESSOR IN THE MODULATION DOMAIN. Yu Wang and Mike Brookes

Are there alternatives to Sigmoid Hidden Units? MLP Lecture 6 Hidden Units / Initialisation 1

Calibration of Microphone Arrays for Improved Speech Recognition

Robustness (cont.); End-to-end systems

VQ Source Models: Perceptual & Phase Issues

Deep Learning. Dr. Johan Hagelbäck.

Recurrent neural networks Modelling sequential data. MLP Lecture 9 Recurrent Networks 1

THE problem of automating the solving of

Synchronous Overlap and Add of Spectra for Enhancement of Excitation in Artificial Bandwidth Extension of Speech

Modulation Features for Noise Robust Speaker Identification

Comparison of Spectral Analysis Methods for Automatic Speech Recognition

Convolutional Networks for Images, Speech, and. Time-Series. 101 Crawfords Corner Road Operationnelle, Universite de Montreal,

Effective post-processing for single-channel frequency-domain speech enhancement Weifeng Li a

Robust Speaker Identification for Meetings: UPC CLEAR 07 Meeting Room Evaluation System

The Munich 2011 CHiME Challenge Contribution: BLSTM-NMF Speech Enhancement and Recognition for Reverberated Multisource Environments

CNMF-BASED ACOUSTIC FEATURES FOR NOISE-ROBUST ASR

UNSUPERVISED SPEAKER CHANGE DETECTION FOR BROADCAST NEWS SEGMENTATION

Automatic Speech Recognition Adaptation for Various Noise Levels

Transcription:

INTERSPEECH 2013 Neural Network Acoustic Models for the DARPA RATS Program Hagen Soltau, Hong-Kwang Kuo, Lidia Mangu, George Saon, Tomas Beran IBM T. J. Watson Research Center, Yorktown Heights, NY 10598, USA {hsoltau,hkuo,mangu,gsaon,tberan}@us.ibm.com Abstract We present a comparison of acoustic modeling techniques for the DARPA RATS program in the context of spoken term detection (STD) on speech data with severe channel distortions. Our main findings are that both Multi-Layer Perceptrons (MLPs) and Convolutional Neural Networks (CNNs) outperform Gaussian Mixture Models (GMMs) on a very difficult LVCSR task. We discuss pre-training, feature sets and training procedures, as well as weight sharing and shift invariance to increase robustness against channel distortions. We obtained about 20% error rate reduction over our state-of-the-art GMM system. Additionally, we found that CNNs work very well for spoken term detection, as a result of better lattice oracle rates compared to GMMs and MLPs. Index Terms: Multi-Layer Perceptron, Time-Delay Neural Network, Convolutional Neural Network 1. Introduction RATS (Robust Automatic Transcription of Speech) is a DARPA program focusing on speech activity detection, spoken term detection (keyword search), and speaker and language identification in noisy environments. The target languages for STD are Levantine and Farsi. The data collection consists of retransmitting clean data over a noisy channel. The clean audio data has Callhome type characteristics (telephone conversations), while the noisy data was obtained by transmitting the original audio through a pair of sender and receiver. In total, 8 different transmissions were performed by using different sender and receiver combinations. The channels are labeled A to H. The channel distortions vary substantially, the closest to the original data is channel G. In this paper we focus on the acoustic modeling problem for spoken term detection (STD), where the goal is to locate a spoken keyword in audio documents. For this task, 300 hours of acoustic training data is available. From our point of view, STD is essentially LVCSR (large vocabulary continuous speech recognition) plus some form of post-processing of lattices to generate a search-able index. For STD, not only the 1-best of the output matters; lattice quality is even more important. While most LVCSR systems are based on Gaussian Mixture Models (GMM), Seide demonstrated in [1] the great potential of Neural Networks by getting 20-30% improvement over GMMs on Switchboard, a task for which many sites build their best systems for competitive NIST evaluations. Because improvements on this task are important, people started to look at Neural Nets again. The use of Neural Networks for acoustic modeling is actually not new, e.g. Bourlard and Morgan [2] summarize the progress that was already done in the early 1990s. Seide attributes the recent success to three key ingredients: the use of a fully context dependent output layer, input features with large span temporal context, and a large number of hidden layers, initialized with pre-training. The pre-training procedure in [1] consists of representing the Neural Net as a Deep Belief Network (DBN) and training it layer by layer. A DBN [3] is a stack of Restricted Boltzmann Machines (RBM) that constrains the network to be a bipartite graph. Maximum likelihood training of an RBM works by minimizing the divergence between the empirical distribution and an approximation of the distribution of the model (contrastive divergence). This form of pre-training does not make use of target labels. It is therefore also called generative pre-training. On the other hand, for most speech tasks, the training data comes with training labels in form of transcripts. In [4], a discriminative form of pre-training was introduced. Discriminative pre-training is essentially layer-wise back-propagation with only one training iteration per layer. The advantage is that the output targets are now used for pre-training, and not only for fine-tuning. In [4], they found small improvements of discriminative pre-training over generative pre-training. In our experiments, we also observed small reductions in error rate using discriminative pretraining. But regardless of whether pre-training is applied, or what kind, the model is still a Multi Layer Perceptron, trained with back-propagation [5], no matter how shallow or deep the network is. So, for naming conventions, we use the term MLP in the following text. This paper is structured as follows. In Section 2, we begin with a short description of the data and the GMM baseline model. In Section 3, we report on experiments with MLPs, the use of different feature streams and Hessian-free sequence training of MLPs. In Section 4, we discuss weight sharing and shift invariance for Convolutional Neural Networks (CNN) and Time Delay Neural Networks (TDNN). In Section 5, we describe how we used these models for our RATS STD evaluation system and report improvements in terms of word error rates (WERs), lattice oracle word error rates, and STD metrics. 2. GMM Baseline The Levantine and Farsi data for the DARPA RATS program is provided by the LDC (Linguistic Data Consortium) [6, 7]. For both languages, about 300 hours of acoustic training data is available. The clean speech corpus is mostly part of the existing Fisher corpus and contains conversational speech over cellphones. Since the R in RATS stands for robust, the clean speech got distorted by transmitting the original data through eight different radio channels. The clean data contained about 150 h of audio, but only about 65 h was labeled as speech. After retransmission, we obtained about 300 h of noisy data for acoustic model training. Similar to our GALE system [8], we built two GMM acoustic models: a conventional unvowelized model (GMM-U) and a Buckwalter vowelized model (GMM-V). We do not expect that Copyright 2013 ISCA 3092 25-29 August 2013, Lyon, France

Channel A B C D E F G Table 1: Word error rates for RATS GMM acoustic models on Levantine dev-04. Feature space WER 40 SI LDA, +/ 5 frames 18.9% 40 FMLLR, +/ 5 frames 17.7% 40 FMMI, Δ, ΔΔ, ΔΔΔ 16.9% 40 FMMI, Δ, ΔΔ, ΔΔΔ + 40 log-mel 16.3% 40 FMMI, Δ, ΔΔ, ΔΔΔ + 40 log-mel + 40 FDLP 15.9% a morphological analyzer for MSA will do a good job generating vowelizations for Levantine Arabic; however, it did provide a useful source of system diversification for STD. The frontend features for both models are based on VTL-warped PLP features and a context window of 9 frames. We apply speaker based cepstral mean and variance normalization, followed by an LDA transform to reduce the feature dimensionality to 40. The ML training of the acoustic model is interleaved with estimation of a global semi-tied covariance (STC) transform. FMLLR speaker adaptation is applied both in training and testing, while MLLR regression trees are applied only during run-time. The total number of Gaussian components is 120000, distributed over 7000 quinphone context-dependent states. Feature- and model-level discriminative training uses the boosted MMI (bmmi) [9] criterion. The error rates for the GMM system are shown in Table 1. While the error rates seem high, one has to keep in mind that this task is rather difficult and deals with conversational speech over cellphone connections and is re-transmitted over noisy channels. Our GMM uses all of the state-of-the-art techniques, and achieved excellent performance in the 2012 RATS evaluation [10]. 3. MLP Experiments 3.1. Training Recipe The MLP is trained using standard back-propagation and the objective function is Cross-Entropy. We use the sigmoid activation function for the hidden layers and softmax for the output layer. The weights are initialized randomly with a uniform distribution, normalized as proposed in [11]. The training data is split into 50 hour chunks. Each chunk is frame-randomized separately. Each mini-batch is 250 frames. Globally, for the entire training data, we randomize the data at the utterance level. For pre-training, the MLP is grown layer-wise with one pass over the training data for each layer growing phase. After the network is fully grown, we train the network until convergence, requiring typically 10 to 15 iterations. No momentum term is applied. The initial step size is 5e 3, but the initial step size is not overly critical as long as it is not totally out of place. Also, the step size seems to be fairly task independent. After each iteration, we measure the improvement on a 5h held-out set, as it was already suggested in [2]. If the improvement is smaller than 1% relative, the step size gets reduced by a factor of 2. 3.2. Feature Sets Here, we want to share some results when experimenting with different feature sets, that looked promising in the beginning, but did not provide improvements at the end on the RATS task. For the initial experiments, we used a a 50h subset of the English Broadcast News data of the LDC Fisher collection (DARPA EARS RT-04 evaluation). The configuration for all MLPs is the same; the only difference is the input. All MLPs Table 2: 50h EnglishBN MLP models Channel A B C D E F G SI-MLP 86.1 74.3 62.1 62.6 76.3 61.1 43.6 FMLLR-MLP 50.3 72.9 59.4 55.0 75.4 60.1 42.1 FMMI-MLP 50.5 73.0 60.1 56.0 75.1 60.8 42.3 FDLP-MLP 50.9 73.5 59.6 57.0 74.4 59.6 42.5 Table 3: Word error rates for MLPs (Lev dev-04). Useof different feature sets analogous to Table 2. have 5 hidden layers with 1024 units and the output layer has 3000 states. The first two rows in Table 2 demonstrate the value of speaker adaptive features, similar to [12]. The FMLLR transforms were borrowed from the corresponding GMM system. The next row is more interesting. We switched to the best features that GMM systems normally have. We use fmmi features [13] that were originally trained for the GMM. While we did not see improvements when used in a frame splicing context as for the other features, the fmmi features worked well with a Δ, ΔΔ, ΔΔΔ context 1. Next, we added VTL warped log-mel features. The filterbank has 40 filters and we use them together with Δ, ΔΔ, ΔΔΔ context. The log-mel features are mean normalized at the utterance level and reduce the error rate from 16.9% to16.3%. We want to note that the improvements do not come from an increased number of parameters. The MLP has already 8.8 million parameters and the increase in parameters is only 1.1%. In the last row of Table 2 we added another set of features. Since the DARPA RATS program focuses on robust LVCSR, we experimented with Frequency domain linear prediction (FDLP) features [14, 15] which are designed to be noise robust. We added them here only for debugging purposes, but got nevertheless an improvement from 16.3% to 15.9% even on clean data. Table 3 compares MLPs trained on different feature sets on the DARPA RATS task, trained on the same data as the GMM baseline as described in Section 2. For both FMMI-MLP and FDLP-MLP, logmel features are part of the input as described for the EnglishBN setup. For the MLPs we use 6 hidden layers with 2048 units and an output layer with 7000 states. What is noticeable that all MLPs perform about 10% relative better than the GMM baseline on semi-clean data (channel G). Even the MLP with SI LDA features is substantially better (43.6% vs 46.6%) and not much worse than the best speaker adaptive MLP (FMLLR-MLP) with 42.1% 2. The picture changes when looking at the channels with more distortions. SI LDA features 1 frame splicing and delta context have same error rates for FMLLR features 2 The FMMI features are speaker adaptive too, based on FMLLR features 3093

Channel A B C D E F G CE 50.3 72.9 59.4 55.0 75.4 60.1 42.1 smbr 46.2 72.4 57.4 52.1 78.4 56.7 37.8 32 logmel Table 4: WER comparison of FMLLR-MLP models trained under Cross-Entropy or smbr objective functions are not as good as the other feature sets. 3.3. Sequence Training The models described so far are trained with regular Cross- Entropy (CE). In [16], state-level MBR (smbr) was proposed as an objective function in a lattice based framework similar to discriminative training for GMMs. In [17] the training procedure became substantially faster and more practical for large training sets by optimizing the objective function using a Hessian-free optimization approach [18] that allowed us to run the training in parallel on a cluster. We first start with training the MLP with Cross-Entropy and use the models to generate numerator and denominator lattices with a weak (1-gram) language model. The results are shown in Table 4 where we compare Cross- Entropy with smbr for the MLP trained using FMLLR features. On the semi-clean channel G, we get a 10% relative improvement over Cross-Entropy and beat our best GMM baseline by 20%. But even on most of the other channels, the discriminatively trained MLP is substantially better than a discriminatively trained (fmmi+bmmi) GMM system. 4. Weight Sharing and Shift Invariance A regular MLP is fully connected, i.e. each hidden unit has connections to all input units. Rumelhart et.al. discussed in [5] a different type of Neural Network that uses only a subset of inputs in the form of localized receptive fields. That Network was designed to discriminate between the letter T and C and to be invariant to translation. In order to achieve shift invariance, the weight learning was changed such that the weight changes were averaged over the receptive fields. For speech recognition, invariance against small changes in the temporal domain is important. The Time Delay Neural Network (TDNN) [19] uses the concept of weight sharing and shift invariance to beat an HMM baseline on a phoneme classification task. In image recognition, Convolutional Neural Networks (CNN) [20] apply the same concepts to obtain shift invariance in two dimensions. For CNNs, a so called sub-sampling layer is added, that reduces the dimensionality by pooling the outputs of the convolutional layer such that higher layers are more invariant to distortions. In [21], a CNN is used to as a replacement of GMMs in a HMM system. In this work, the CNNs job is to guard against changes in the spectrum, and keep the HMM to deal with temporal invariance. While this work was done on a small scale phone recognition task (TIMIT), [22] uses a CNN HMM setup for LVCSR tasks (Switchboard and Broadcast News). We follow the work of [21, 22] and use CNNs to compute acoustic scores for conventional context dependent 3-state HMMs. Distortions in the temporal domain are handled by the HMM, while distortions in the frequency domain are handled by the CNN. 9x9 11 frame context Figure 1: CNN Layer #0, sliding window 9x9 over input features 24 feature windows 1x3 3 windows in temporal domain Figure 2: CNN Layer #0, non-overlapping window 1x3 over output values 4.1. Network Structure Specifically, our network structure is as follows. Input features are 32-dimensional logmel features 3. The input context is 11 frames. In addition to the logmel features, we use Δ and Δ, Δ features. The input features are mean and variance normalized at speaker level. We also apply VTLN as it was done in [22]. Figure 1 shows the first step: a sliding window operates over the input features: For an input context of 11 frames and a window size of 9, we obtain 3 windows in the temporal domain. And at the logmel level, we get 32 9+1 = 24windows. Altogether, we get 3 24 = 72 windows. Accounting for the Δ and Δ, Δ features, each window has 3 9 9 = 243 features. Each window goes through a regular MLP layer, resulting in 72 output values for each hidden unit. This is followed by a max-pooling layer, where the maximum is taken for the output values over a 1x3 window 4. This operation will help to make the model more invariant against small shifts in the logmel domain and hopefully increase the robustness against the channel distortions we encounter in the RATS data. Since the windows are not overlapping here, the number of windows in the logmel domain gets reduced from 24 to 24/3 =8. The second layer of the network is also a convolutional 3 We normally use a filterbank of 18 for 8kHz data, but increased the number of filters to 32 to get a finer spectrum resolution for the CNN. 4 We do not apply pooling in the temporal domain for now, but consider this in the future. 3094

Layer inputn x outputn #0 243 x 128 #1 1536 x 256 #2 1280 x 2048 #3 2048 x 2048 #4 2048 x 2048 #5 2048 x 2048 #6 2048 x 7000 Table 5: CNN structure and dimensions of layer weights Model CPU GPU MLP 35h 7h CNN 46h 15h Table 7: Cross-Entropy Training Time per iteration 35 30 GMM NN Channel A B C D E F G MLP 46.2 72.4 57.4 52.1 78.4 56.7 37.8 CNN 45.8 70.7 54.8 50.9 72.1 52.5 37.9 Miss probability (%) 25 20 Table 6: WER comparison of GMM, MLP, CNN 15 layer, that uses a sliding window of 3x4 over the outputs of the first layer. In total, the network has 7 layers and the first two layers are convolutional as described above. The network structure is summarized in Table 5. 4.2. Pre-Training As for regular MLPs, discriminative pre-training is applied for the CNN too. The only difference is, that the first two layers are trained together. The first pre-training step estimates the weights for the layers #0,#1,#2,#6. The next step adds another hidden layer and trains #0,#1,#2,#3,#6 and so on. After all layers are added, regular back-propagation continues until convergence. 4.3. Results The CNN is trained on the same RATS training data as the MLP and GMM systems. Frame-level randomization is slightly more complicated for CNNs and requires us to write the features with sufficiently long temporal context. We also apply sequence training [17, 16] for the CNN similar to the MLP models. While the CNN is not better than the MLP for semi-clean data (channel G), we see significant improvements on all other channels. This is in line with our expectations, that the shift invariance in the feature domain helps to make the model more robust. 4.4. Training Time Lastly, we also want to report the training time for Neural Nets. While Neural Nets perform substantially better than GMMs, it also takes much longer to train them. However, with the use of GPU (graphics processing unit) devices, training on 300 hours becomes quite practical. The training time reported in Table 7 is based on training the models described before (6 hidden layers a 2048, 7000 outputs). The CPU is a 12-core Intel machine. The GPU uses a single Kepler110 device. Both machines have sufficient memory. Including pre-training, it takes about 15 passes over the training data. So, for the MLP the total training time is about 4.5 days and 10 days for the CNN. As shown in Table 7, it takes substantially more time to train a CNN than an MLP. The reason is that CNNs require making localized windows of the input. Even with multi-threading, these operations are computationally expensive. And since the 10 0 0.2 0.4 0.6 0.8 1 1.2 1.4 False Alarm probability (%) Figure 3: Farsi STD peformance comparing GMM and Neural Nets operations require memory bandwidth, the speed-up of using GPUs is less (factor of 3) than for MLPs (factor of 5). 5. Neural Nets and STD For the DARPA RATS program, we are not only interested in 1- best error rates but also want to improve our STD system. The STD system does not use only the best path from the recognizer; lattices are used to capture more of the search space. Figure 3 compares STD performance (false alarm vs. miss rate) for our best Farsi GMM and NN models. The GMM system is significantly outperformed by the Neural Net. The reason is that the Neural Net has not only a better 1-best error rate, but the lattice quality is also substantially better with Neural Nets. The goal of the STD DARPA RATS program is to reduce the the false alarm rate for a given miss rate. The miss rate for the 2013 evaluation was set to 20%. At this operating point, the GMM has a false alarm rate of 1.284%, and the Neural Net reduced it to 0.265% - an improvement of almost a factor of 5. 6. Conclusions We presented a study of acoustic modeling techniques for very challenging data. MLPs outperformed our best GMM model by 20%. While the convolutional model did not work better than a regular MLP on semi-clean data, it worked substantially better for the noisy channels and improvements in WER translated to much improved STD performance. 7. Acknowledgements We want to thank Tara Sainath and Brian Kingsbury for useful discussions about convolutional models and sequence training. This work was supported in part by Contract No. D11PC20192 DOI/NBC under the RATS program. The views expressed are those of the author and do not reflect the official policy or position of the Department of Defense or the U.S. Government. 3095

8. References [1] F. Seide, G. Li, and D. Yu, Conversational speech transcription using context-dependent deep neural networks, in Interspeech, 2011. [2] H. Bourlard and N. Morgan, Connectionist Speech Recogniton, A Hybrid Approach. Kluwer, 1994. [3] G. Hinton, S. Osindero, and Y. Teh, A fast learning algorithm for deep belief nets, in Neural Computation, 2006. [4] F. Seide, G. Li, X. Chen, and D. Yu, Feature engineering in context-dependent deep neural networks for conversational speech transcription, in Proc. ASRU, 2011. [5] D. Rumelhart, G. Hinton, and R. Williams, Learning internal representations by error propagation, in Parallel Distributed Processing, 1986. [6] M. Maamouri et al., LDC2006S29, Arabic CTS Levantine QT training data set 5, in Linguistic Data Consortium, 2006. [7] D. Graff, S. Sessa, S. Strassel, and K. Walker, Rats data plan, in Linguistic Data Consortium, Tech. Rep., 2011. [8] H. Soltau, G. Saon, B. Kingsbury, H.-K. Kuo, L. Mangu, D. Povey, and A. Emami, Advances in Arabic Speech Transcription at IBM under the DARPA GALE Program, IEEE TSAP, 2009. [9] D. Povey, D. Kanevsky, B. Kingsbury, B. Ramabhadran, G. Saon, and K. Visweswariah, Boosted MMI for model and featurespace discriminative training, in Proc. ICASSP, vol. II., 2008, pp. 4057 4060. [10] L. Mangu, H. Soltau, H.-K. Kuo, B. Kingsbury, and G. Saon, Exploiting diversity for spoken term detection, in Proc. ICASSP, 2013. [11] X. Glorot and Y. Bengio, Understanding the difficulty of training deep feedforward neural networks, in Proc. AISTATS, 2010, pp. 249 256. [12] T. Sainath, B. Kingsbury, B. Ramabhadran, P. Fousek, P. Novak, and A. Mohamed, Making deep belief networks effective for large vocabulary continuous speech recognition, in Proc. ASRU, 2011, submitted. [13] D. Povey, B. Kingsbury, L. Mangu, G. Saon, H. Soltau, and G. Zweig, fmpe: Discriminatively trained features for speech recognition, in ICASSP-2005, 2005. [14] M. Athineos and D. Ellis, Autoregressive modelling of temporal envelopes, IEEE Transactions on Signal Processing, vol. 55, no. 11, 2007. [15] S. Ganapathy, Signal analysis using autoregressive models of amplitude modulaton, Ph.D. dissertation, Johns Hopkins University, 2012. [16] B. Kingsbury, Lattice-based optimization of sequence classification criteria for neural-network acoustic modeling, in Proc. ICASSP, 2009. [17] B. Kingsbury, T. N. Sainath, and H. Soltau, Scalable minimum Bayes risk training of neural network acoustic models using distributed Hessian-free optimization, in Proc. Interspeech, 2012. [18] J. Martens, Deep learning via Hessian-free optimization, in Proc. ICML, 2010. [19] A. Waibel, T. Hanazawa, G. Hinton, K. Shikano, and K. Lang, Phoneme recognition: Neural networks vs hidden Markov models, in Proc. ICASSP, 1988. [20] Y. L. Cun, B. Boser, J. Denker, D. Henderson, R. Howard, W. Hubbard, and L. Jackel, Handwritten digit recognition with a back-propagation network, in Proc. NIPS, 1990. [21] O. Abdel-Hamid, A. Mohamed, H. Jiang, and G. Penn, Applying convolutional neural network concepts to hybrid NN-HMM model for speech recognition, in Proc. ICASSP, 2012. [22] T. Sainath, A. Mohamed, b. Kingsbury, and B. Ramabhadran, Deep convolutional neural networkds for LVCSR, in Proc. Interspeech, 2013. 3096