30
Building A Highly Accurate Mandarin Speech Recognizer with Language-Independent Technologies and Language-Dependent Modules EDICS: SPE-LVCR Mei-Yuh Hwang * [email protected] Microsoft Corporation, Redmond, WA (O)425-705-3387 (Fax)425-936-7329 Gang Peng, Mari Ostendorf Wen Wang {gpeng,mo}@ee.washington.edu [email protected] University of Washington, Seattle, WA SRI International, Menlo Park, CA Arlo Faria Aaron. Heidel [email protected] [email protected] ICSI, UC Berkeley, CA National Taiwan University, Taipei, Taiwan May 11, 2008 Abstract We describe in detail a system for highly accurate large-vocabulary Mandarin automatic speech recognition (ASR), a collaborative suc- 1

Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

  • Upload
    others

  • View
    8

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

Building A Highly Accurate Mandarin Speech

Recognizer with Language-Independent

Technologies and Language-Dependent Modules

EDICS: SPE-LVCR

Mei-Yuh Hwang∗

[email protected]

Microsoft Corporation, Redmond, WA

(O)425-705-3387 (Fax)425-936-7329

Gang Peng, Mari Ostendorf Wen Wang

{gpeng,mo}@ee.washington.edu [email protected]

University of Washington, Seattle, WA SRI International, Menlo Park, CA

Arlo Faria Aaron. Heidel

[email protected] [email protected]

ICSI, UC Berkeley, CA National Taiwan University, Taipei, Taiwan

May 11, 2008

Abstract

We describe in detail a system for highly accurate large-vocabulary

Mandarin automatic speech recognition (ASR), a collaborative suc-

1

Page 2: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

cess among four research organizations. The prevailing hidden Markov

model (HMM) based technologies are essentially language independent

and constitute the backbone of our system. These include minimum-

phone-error (MPE) discriminative training and maximum-likelihood

linear regression (MLLR) adaptation, among others. Additionally,

careful considerations are taken into account for Mandarin-specific is-

sues including lexical word segmentation, tone modeling, pronunciation

designs, and automatic acoustic segmentation.

Our system comprises two acoustic models (AMs) with significant

differences but similar accuracy performance for the purposes of cross

adaptation. The outputs of the two sub-systems are then rescored

by an adapted n-gram language model (LM). Final confusion network

combination (CNC) [1] yielded 9.1% character error rate (CER) on the

DARPA GALE (http://www.darpa.mil/ipto/programs/gale/) 2007 of-

ficial evaluation, the best Mandarin ASR system in that year.

1 Introduction

With China’s rapid economic growth and one of the most widely spoken lan-

guages in the world, various purposes demand a highly accurate Mandarin

automatic speech recognizer. Under the DARPA GALE Project, this paper

seeks to achieve such a goal on broadcast news (BN) and broadcast conver-

sational (BC) speech. We will prove that the core technologies developed for

English ASR are applicable to a new language such as Mandarin. However,

to achieve the best performance, one needs to add language-dependent com-

ponents, which for Mandarin includes extraction of tone-related features,

lexical word segmentation, phonetic pronunciation, and optimization of au-

2

Page 3: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

tomatic acoustic segmentation. We have published extraction of tone-related

features in [2], [3], [4], and [5]. This paper will elaborate our word segmen-

tation algorithm and the design philosophy of our pronunciation phone sets

and acoustic segmentation, along with experimental results. The core speech

technologies include discriminative training criteria with discriminative fea-

tures, multiple-pass unsupervised cross adaptation between two different

feature front ends, and language model adaptation.

This paper starts with a description of the acoustic and language model

training data used in building the system, and the lexical word segmentation

algorithm. Then we summarize our decoding architecture, which illustrates

the need for two complementary sub-systems; in this setup, one can also

clearly see where LM adaptation fits. The next three sections describe the

key components of the system: Section 4 elaborates the improvement in

automatic segmentation of long recordings into utterances of a few seconds;

Section 5 describes the two front-end features used by our two complemen-

tary acoustic models; Section 6 discusses our LM adaptation algorithm.

Finally Section 7 demonstrates the relative improvements of various compo-

nents via experiments, and presents our 2007 official evaluation result. In

Section 8 we summarize our contribution and future work.

2 Speech and Text Corpora

2.1 Acoustic Data

In this paper, we use about 866 hours of BN and BC speech data collected by

the LDC, including the corpora of the Mandarin Hub4, TDT4, and GALE

3

Page 4: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

P1 release, up to shows dated before 11/1/2006, for training our acoustic

models. The TDT4 data do not have manual transcriptions associated with

them – instead, only closed captions. We use a flexible alignment algorithm

to filter out bad segments where the search paths differ significantly from

the closed captions [6]. After the filtering, we keep 89 hours of data for

training. Table 1 summarizes our acoustic training data.

Table 1: Acoustic training data.Corpus Year DurationHub4 1997 30 hrsTDT4 2000-2001 89 hrs

GALE P1 2004-10/2006 747 hrsTotal 1997-2006 866 hrs

As shown in Table 2, we use three different data sets for system devel-

opment: the EARS RT-04 evaluation set (Eval04), GALE 2006 evaluation

set (Eval06), and GALE 2007 development set (Dev07)1. Once the system

parameters are finalized based on these development test sets, we then apply

the settings to the GALE 2007 evaluation set (Eval07).

Table 2: Summary of all test sets. The first three are used for systemdevelopment. The last one is the evaluation set for the final system.

Data year/month #shows BC BNEval04 2004/04 3 0 1 hrEval06 2006/02 24 1 hr 1.16 hrsDev07 2006/11 74 1.5 hrs 1 hrEval07 2006/11,12 83 1 hr 1.25 hrs

1The Dev07 set used here is the IBM-modified version, not the original LDC-releasedversion.

4

Page 5: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

2.2 Text Data, Word Segmentation and Lexicon

Our text corpora come from a wider range of data. In addition to those

transcripts of the acoustic training data, we add the LDC Chinese TDT2,

TDT3, Chinese News Translation Text Part 1, Multiple-Translation Chi-

nese Corpus Part 1, 2, and 3, Chinese Gigaword corpus, all GALE-related

Chinese web text releases with news dated before 11/1/2006, 183M charac-

ters of National Taiwan University (NTU) Web downloaded BN transcripts

and newswire text, 33M words BN and newswire web data collected by

Cambridge University, and the Mandarin conversation LM training data de-

scribed in [7], which includes 1M-word acoustic transcripts and 300M-word

Web downloaded conversations.

To process the text data, we first perform text normalization to remove

HTML tags and phrases with bad or corrupted codes, to convert full-width

punctuations and letters to half-width, and to convert numbers, dates and

currencies into their verbalized forms in Chinese. Word fragments, laughter,

and background noise transcriptions are mapped to a special garbage word.

To leverage English ASR technologies to Mandarin directly, our Man-

darin ASR system is based on “word” recognition. We believe word-based

ASR is better than character-based ASR for its longer acoustic and language

contexts. Chinese word segmentation aims at inserting spaces between se-

quences of Chinese characters and to form grammatical and meaningful

Chinese sentences. Starting from the 64K-word BBN-modified LDC Chi-

nese word lexicon, we manually augment it with 20K new words (both Chi-

nese and English words) over time from various sources of frequent names

5

Page 6: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

and word lists. For word segmentation, we begin with a simple longest-first

match (LFM) algorithm with this 80K-word list to segment all training text.

The most frequent 60K words are then selected as our lexicon to train an

n-gram LM, with all out-of-vocabulary (OOV) words mapped to the garbage

word. In total, after word segmentation, the LM training corpora comprise

around 1.4 billion words. Among them, 11 million words are from the BC

genre.

LFM segmentation does not take context into account and sometimes

makes inappropriate segmentation errors that result in wrong or difficult-

to-interpret semantics, as the following example show:

(English) The Green Party and Qin-Min Party reached a consensus.

(LFM) Ì�j Z* Ìj HÄ á# (wrong)

(1gram) Ì�j Z *Ìj HÄ á# (correct)

To improve word segmentation, we have been advocating maximum-

likelihood (ML) search based on an n-gram LM. A lower-order n-gram is

preferred if the LM is ML-trained on the same text, so that there is a

better chance to get out of the local optimal segmentation. In our case, we

used the unigram LM trained on LFM-segmented text to re-segment the

same training data. Having done so, the only possible OOV words now

are OOV English words and OOV single-character Chinese words. In our

experience, ML word segmentation results in only slightly better perplexity

and usually translates to no further improvement in recognition. However,

we believe that the ML word segmentation more often offers a semantically

better segmentation which helps downstream applications such as machine

6

Page 7: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

translation.2

After the ML word segmentation, we then re-train our n-gram LMs,

smoothed with the modified Kneser-Ney algorithm [8] using the SRILM

toolkit [9]. N-grams of the garbage word are also trained.

Table 3 lists the sizes of the full n-gram LMs and their pruned versions.

The pruned n-grams (qLMn) are used in fast decoding, as will be explained

in Section 3. To estimate the difficulty of the ASR task, we extract a subset

of Dev07 where there is no OOV word with respect to the 60K-lexicon,

for computing perplexity. All noises labeled in the reference are removed

when computing perplexity. This Dev07-i set contains about 44K Chinese

characters, accounting for 99% of the full Dev07 set.

Table 3: Numbers of entries in n-gram LMs. qLMn are the highly prunedversions of the full n-grams LMn.

#2grams #3grams #4grams Perplexity(Dev07-i)qLM3 6M 3M — 379.8qLM4 19M 24M 6M 331.2LM3 38M 108M — 325.7LM4 58M 316M 201M 297.8

3 Decoding Architecture

Figure 1 illustrates the flowchart of our recognition engine. We will briefly

describe the architecture in this section, while details will be presented in

the next three sections.2Following our algorithm, Y-C Pan at National Taiwan University indicated that their

system character accuracy improved from 74.42% to 75.01%, using LFM vs. ML segmen-tation.

7

Page 8: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

Figure 1: System decoding architecture. Block arrows represent N-besthypotheses.

VTLN/ CMN/CVN

Auto speaker clustering

Acoustic segmentation

PLP-SA MLLR, LM3

qLM4 Adapt/Rescore

qLM4 Adapt/Rescore

MLP-SA MLLR, LM3

MLP-SI qLM3

Confusion Network Combination

3.1 Acoustic Segmentation and Feature Normalization

GALE test data come with per-show recordings. However, only specified

segments in each recording are required to be recognized. Instead of feeding

the minute-long segment into our decoder, we re-fine each segment into ut-

terances of a few seconds long, separated by long pauses, and run utterance-

based recognition. This allows us to run wide-beam search and keep rich

search lattices per utterance. We call this first step the acoustic segmenta-

tion step.

Next we perform automatic speaker clustering using Gaussian mixture

models of static MFCC features and K-means clustering. We call these

speakers auto speakers. Note that the number of speakers in the show is

unknown. Therefore, we limit the average number of utterances per speaker

8

Page 9: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

by an empirical threshold, tuned on the development sets. Vocal tract length

normalization (VTLN) is then performed for each auto speaker, followed by

utterance-based cepstral mean normalization (CMN) and cepstral variance

normalization (CVN) on all features. Speaker boundary information is also

used in later steps for speaker-based AM adaptation.

3.2 Search with Trigrams and Cross Adaptation

Referring to Figure 1, the decoding is composed of three trigram recognition

passes:

1. MLP-SI Search: Using qLM3, we start a quick search with a speaker-

independent (SI) within-word (non cross-word, nonCW) triphone

acoustic model, based on MFCC cepstrum features, pitch features,

and multi-layer perceptron (MLP) generated phoneme posterior fea-

tures. This model is denoted the MLP model, which will be elaborated

in Section 5.1. This step generates a good hypothesis quickly.

2. PLP-SA Search: Next we use the above hypothesis to learn the

speaker-dependent SAT feature transform and to perform MLLR

adaptation [10] per speaker, on the CW triphone model with PLP cep-

strum features and pitch features. This PLP model and SAT transform

will be explained in Section 5.2.

After the acoustic model is speaker adapted (SA), we then run full-

trigram decoding to produce an N-best list for each utterance.

3. MLP-SA Search: Similar to PLP-SA Search, we run cross adaptation

first, using the trigram hypothesis from PLP-SA Search to adapt the

9

Page 10: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

CW triphone MLP-model, followed by full-trigram decoding to pro-

duce another N-best list. This model has the same feature front end

as the one at the MLP-SI step.

3.3 LM Adaptation and Confusion Network Combination

Based on topic-dependent LM adaptation, the N-best list generated by PLP-

SA Search is used to adapt the 4-gram LM on a per-utterance basis. All

of the N-best hypotheses of an utterance are then rescored by this adapted

4-gram LM. Finally all the words in each N-best hypothesis are split into

character sequences, with the adapted acoustic score and adapted language

score unchanged.

Next we apply the same 4-gram adaptation procedure to the N-best list

generated by MLP-SA Search. Finally a character-level confusion network

is obtained by combining both rescored character N-best lists, and the most

likely character in each column of the network is output as the system out-

put.

4 Acoustic Segmentation

In error analysis of our previous system, we discover that deletion errors

were particularly serious. We find that some of the deletion errors were

caused by false alarm garbage words.To control these garbage false alarms,

we introduce a garbage penalty into the decoder, which is successful in

removing some deletion errors.

However, most of our deletion errors came from dropped speech segments

10

Page 11: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

due to faulty acoustic segmentation. Therefore, we attempt to improve

acoustic segmentation so that not only fewer speech segments are dropped,

but new insertion errors are simultaneously avoided [11].

4.1 Previous Segmenter

Our previous segmenter is a speech-silence detector. A recognizer is run

with a finite state grammar as shown in Figure 2(a). There are three words

in the vocabulary of the decoder: silence, noise, and speech, whose pro-

nunciations are shown in the figure. Each pronunciation phone (bg, rej, fg)

is modeled by a 3-state HMM, with 300 Gaussians per state. The HMMs

are ML-trained on Hub4, with 39-dimensional features of MFCC and its

first-order and second-order differences. The segmenter operates without

any knowledge of the underlying phoneme sequence contained in the speech

waveform. More seriously, due to the pronunciation of speech, each speech

segment is defined as having at least 18 consecutive fgs, which forces any

speech segment to have a minimum duration of 540 ms, given that our front-

end computes one feature vector every 10ms.

After speech/silence is detected, segments composed of only silence and

noises are discarded. Then if the segmenter judges that two consecutive

speech segments are from the same speaker (based on MFCC Gaussian-

mixture models) and the pause in between is very short (under a certain

threshold), it attempts to merge them into one utterance that is at most 9

seconds long. The speaker boundary decision algorithm here is similar to

(but more micro, and independent of) the auto-speaker clustering step in

Figure 1.

11

Page 12: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

Figure 2: The finite state grammar of our acoustic segmenter: (a) previoussegmenter, (b) new segmenter.

man2 /I2 F/

man1 /I1 F/

silence /bg bg/

end

start

noise /rej rej/

(b)

silence /bg bg/

end

start

noise /rej rej/

speech /fg18+/

(a) eng /forgn forgn/

4.2 New Segmenter

Our new segmenter, shown in Figure 2(b), makes use of broad phonetic

knowledge of Mandarin and models the input recording with five words:

silence, noise, a Mandarin syllable with a voiceless initial, a Mandarin

syllable with a voiced initial, and a non-Mandarin word. Altogether there

are 6 distinct HMMs (bg rej I1 F I2 forgn) for speech-silence detection,

and the minimum speech duration is reduced to 60 ms. Except for the finite-

state grammar and the pronunciations, the rest of the segmentation process

remains the same. As shown later in Section 7.1, we are able to recover most

of the discarded speech segments via the new finite state grammar and the

new duration constraint.

5 Two Acoustic Systems

As illustrated in Figure 1, a key component of our system is cross adap-

tation and system combination between two complementary sub-systems.

12

Page 13: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

[12] showed three discriminative techniques that are effective in reducing

recognition errors: MLP features [13], MPE (MPFE) [14], [15] discrimina-

tive learning criterion, and fMPE feature transform [16]. Table 4, cited from

[12], shows that the three discriminative techniques are additive, under an SI

recognition setup. However, combining all three yields minimal further im-

provement compared with combining only two of them, after unsupervised

adaptation. In designing our two acoustic systems, we therefore decide to

choose the most effective combinations of two techniques: MLP+MPE and

fMPE+MPE, where the most effective technique, MPE training, is always

applied. Furthermore, to diversify the error behaviors, we specifically design

a new pronunciation phone set for the second system, in the hope to behave

better on BC speech.

Table 4: Contribution of three discriminative techniques. The experimentswere conducted on an English speaker independent ASR task.

MLP MPE fMPE WER17.1%

yes 15.3%yes 14.6%

yes 15.6%yes yes 13.4%yes yes 14.7%

yes yes 13.9%yes yes yes 13.1%

13

Page 14: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

5.1 System-MLP

5.1.1 The Pronunciation Phone Set

Our first AM system uses 70 phones for pronunciations, inherited from the

BBN dictionary with our minor fixes. Vowels usually have 4 distinct tones,

accounting for 4 different phones. Neutral tones are not modeled in this

phone set. Instead, the third tone is used to represent the neutral tone. The

pronunciation rule follows the main vowel idea proposed in [17] for reducing

the size of the phone set and easy parameter sharing. Table 5 lists some key

syllables for a better understanding when converting Pinyin to the main-

vowel phonetic pronunciation. Notice for Pinyin syllable final /an/ and

/ai/, we use capital phone /A/ to distinguish them from other /a/ vowels.

The rare I2 as in # is modeled by I1 as it does not exist in this phone set.

In addition to the 70 phones, there is one phone designated for silence,

and another one, rej, for modeling all noises, laughter, and foreign (non-

Mandarin) speech. Both the silence phone and the noise phone are context-

independent. The garbage word is modeled by a pronunciation graph of two

or more rej. All HMMs have the 3-state Bakis topology without skipping

arcs.

5.1.2 The HMM Observation Features

The front-end features consist of 74 dimensions per frame, including (a) 13-

dim MFCC cepstra, and its first- and second-order derivatives; (b) Spline

smoothed pitch feature [4], and its first- and second-order derivatives; and

(c) 32-dim phoneme-posterior features generated by MLPs [13, 12].

14

Page 15: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

Table 5: Main-vowel phonetic pronunciations for Chinese syllables.Sample character Pinyin Phonetic pronunciation² ba b a3� kang k a2 NG� nan n A2 N� yao y a4 W� ai A4 Yÿ e e2û ben b e1 N� beng b e1 NG er er2� ye y E1� chui ch w E1 Y· wo w o3� you y o3 W� yi y i1X bu b u4� zhi zh IH1Ï chi ch IH14 shi sh IH4� ri r IH4ý zi z I1� ci c I1� si s I1í juan j v A1 N¬ yu v yu4

The MLP feature is designed to provide discriminative phonetic infor-

mation at the frame level. Its generation involves the following three main

steps, as illustrated in Figure 3. In all MLPs mentioned in the paper, all

hidden units use the sigmoid output function and all output units use the

softmax output function.

1. The Tandem feature [18] generated by one MLP

We first, for each input frame, concatenate its neighboring 9 frames

15

Page 16: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

of PLP and pitch features as the input to an MLP. That is, there

are 42 ∗ 9 = 378 input units in this MLP. Each output unit of the

MLP models the likelihood of the central frame belonging to a certain

phone, given the 9-frame intermediate temporal acoustic evidence. We

call this vector of output probabilities the Tandem phoneme posteriors.

Particularly in our case, there are 15000 hidden units, and 71 output

units in the Tandem MLP. The noise phone rej is excluded from

the MLP output because it is presumably not a very discriminable

class. It is aligned to many kinds of noises and foreign speech, not any

particular phonetic segment. The acoustic training data are Viterbi

aligned using an existing acoustic model, to identify the target phone

label for each frame, which is then used during MLP back-propagation

training.

2. The HATs feature [13] generated by a network of MLPs

Next, we separately construct a two-stage network structure where the

first stage contains 19 MLPs and the second stage is a single MLP. The

purpose of each MLP in the first stage is to identify a different class of

phonemes, based on the log energy of a different critical band across

a long temporal context (51 frames ≈ 0.5 seconds). From [13], it is

more advantageous to feed forward to the 2nd stage the output from

the hidden units rather than the output of the output-layer units,

probably because the softmax function mitigates the discriminative

power of the output layer.

The second stage of MLP then combines the information from all

16

Page 17: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

of the hidden units of the first stage to make a grand judgment

on the phoneme identity for the central frame. The output of this

merger MLP is called the HATs phoneme posteriors (hidden activation

temporal patterns) and is illustrated in Figure 3.

3. Combining Tandem and HATs

Finally, the 71-dim Tandem and HATs posterior vectors are combined

using the Dempster-Shafer [19] algorithm. The values in the combined

vector still range from 0 to 1. We then apply logarithm to make

the posterior distributions appear more like a Gaussian. Principal

component analysis (PCA) is then applied to the log posteriors to (a)

make each dimension independent, as our HMM models use Gaussian

mixtures with diagonal co-variances, and (b) reduce the dimensionality

from 71 to 32. The 32 dimensions of phoneme posteriors are then

appended to the MFCC and pitch features. This system with 74-dim

features is thus referred to as System-MLP because of the use of the

MLP features.

MLP feature extraction as described in this paper differs from our 2006

system [20] in two principal ways: (1) the amount of training data is dou-

bled, and (2) pitch features are added to the Tandem MLP input layer.

While there is no data like more data, this introduces considerable prac-

tical complications for MLP training since the online training algorithm

cannot be easily parallelized across multiple machines. To address this, we

optimize our multi-threaded QuickNet code to run on an 8-core server in

a quasi-online batched mode: each network update is performed using the

17

Page 18: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

Figure 3: The HATs feature, computed using a two-stage MLP. Notice theoutput from the hidden units of the first-stage MLPs is the input to thesecond-stage MLP.

feed-forward error signal accumulated from a batch of 2048 randomized in-

put frames. Additionally, we decrease the training time by partitioning the

training data and applying the learning schedule described in [21].

5.1.3 Model Training

Except for the Hub4 training corpus, where we use the hand-labeled

speaker information, we apply the same automatic speaker clustering al-

gorithm in Figure 1 to all other training corpora, followed by speaker-based

VTLN/CMN/CVN.

18

Page 19: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

Then following the flowchart shown in Figure 4, we first train nonCW

triphone models with the 74-dim MLP-feature, using the ML objective func-

tion. Next through unigram recognition on the training data, we gener-

ate phone lattices for MPE training [14, 15]. The MPE training updates

only Gaussian means, while keeping ML Gaussian covariances and mixture

weights intact. The nonCW MPE model is used at Step MLP-SI.

Figure 4: The acoustic model training process.

 tx

nonCW ML Model 

 

tz  

ty  

Unigram Recognition 

Phone lattices 

MPE Training 

      cMLLR 

SAT transforms,   kk bA ,

nonCW MPE Model 

ML Training 

CW ML Model 

MPE Training 

Unigram Recognition 

CW MPE Model 

Phone lattices

Smaller ML Training 

Smaller CW ML Model, h 

fMPE Training 

fMPE Transform, M

With the nonCW ML model3, we learn the speaker-dependent SAT fea-

ture transform, yt = Akxt + bk, based on 1-class constrained MLLR [22, 10],

or cMLLR, for each speaker k. The SAT feature transform is applied to

train CW triphone ML model, which is then used with a unigram decoder

to generate CW triphone lattices on the training data for the purpose of3The ML model is used because SAT is an ML learning criterion.

19

Page 20: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

training a CW triphone MPE model. The CW triphone SAT ML model is

also used to learn the speaker-dependent MLLR model transforms at Step

MLP-SA depicted in Figure 1. Notice this MLLR transform usually con-

tains multiple regression classes, depending on the amount of SA data, and

is applied at the model domain of the MPE model during the full trigram

search.

The last column of Figure 4 is not used by System-MLP and will be

explained in the next section.

All models use decision-tree based HMM state clustering [23] to deter-

mine Gaussian parameter sharing. In our system, there are 3500 shared

states, each with 128 Gaussians. This model size is denoted as 3500x128.

5.2 System-PLP

The second acoustic system contains 42-dimensional features with static,

first- and second-order derivatives of PLP and pitch features. Similar to

System-MLP, we first train a 3500x128 nonCW triphone ML model and use

it to learn the speaker-dependent SAT feature transforms. Next referring to

the third column in Figure 4, we ML train a smaller (3500x32) CW triphone

model with the SAT transforms, for the purpose of learning the fMPE [16]

feature transform. A smaller model is adopted for its computational effi-

ciency.

5.2.1 fMPE Feature Transform

For each time frame, we compute all Gaussian density values of 5 neighboring

frames, given the 3500x32 smaller CW SAT ML model. This gives us a

20

Page 21: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

Gaussian posterior vector, ht, of 3500x32x5=560K dimensions if context-

independent phone models are excluded. The final feature used is

zt = (Akxt + bk) + Mht

where M is a sparse 42x560K matrix, and is learned through an MPE crite-

rion [16]. Finally with the zt features, we train 3500x128 CW triphone ML

and MPE models. Both models are used at Step PLP-SA: The ML model

is used to learn MLLR transforms, which are applied to the MPE model for

full-trigram search.

5.2.2 The Pronunciation Phone Set

To make the two acoustic systems behave more differently and to focus on

the more difficult BC speech, we design a different pronunciation phone set

for the PLP system. Since BC speech tends to be faster and more sloppy,

we first introduce a few diphthongs as shown in Table 6, where vowels with

no tone apply to all four tones. Combining two phones into one reduces the

minimum duration requirement by half and hence is likely better for fast

speech. The addition of diphthongs also naturally removes the need for the

syllable-ending /Y/ and /W/ sounds.

Next the 72-phone set does not model the neutral tone, but instead

the third tone is used as a replacement. As there are a few very common

characters that are 5-th tone, we add three neutral-tone phones for them.

Furthermore, we add phone /V/ for the v sound in English words, as this

phone is missing in Mandarin but not difficult at all for Chinese people to

21

Page 22: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

pronounce correctly. For the two common filled-pause characters (ã,F),

we use two separate phones to model them individually, so that they are

not sharing parameters with regular Chinese words. As there is not much

training data for V and the filled pauses, we make these three phones context-

independent, indicated by the asterisks.

Finally, to keep the size of the new phone set manageable, we merge /A/

into /a/, and both /I/ and /IH/ into /i/. We rely on triphone modeling to

distinguish these allophones of the same phoneme. With /I2/ represented

by /i2/, the second tone of the non-retroflex /i/ is now modeled correctly.

Table 6: Difference between the 72-phone and 81-phone sets. Asterisksindicate context-independent phones.

example phone-72 phone-81� a W awð E Y ey� o W ow� a Y ay� A N a N' I iÚ IH iê e3 e5m a3 a5� i3 i5victory w V ∗

ã o3 fp o∗

F e3 N fp en∗

6 Topic-Based Language Model Adaptation

After we obtain two N-best lists for each utterance, we attempt to update

the language score with an adapted higher-order n-gram.

22

Page 23: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

We perform topic-based LM adaptation using a Latent Dirichlet Allo-

cation (LDA) topic model [24, 25]. The topic inference algorithm takes as

input a weighted bag of words w (e.g. in one topic-coherent story) and an

initial topic mixture θ0 and returns a topic mixture θ. During training, we

label the topic of each individual sentence to be the one with the maximum

weight in θ, and add the sentence to this topic’s corpus. We then use the

resulting topic-specific corpora to train one n-gram LM per topic [26].

During decoding, we infer the topic mixture weights dynamically for each

utterance; we select the top few most relevant topics above a threshold, and

use their weights in θ to interpolate with the background language model

in Table 3.

In order to make topic inference more robust against recognition er-

rors, we weight the words in w based on an N-best-list derived confidence

measure. Additionally we include words not only from the utterance being

rescored but also from surrounding utterances in the same story chunk via

a decay factor, where the words of distant utterances are given less weight

than those of nearer utterances. As a heuristic, utterances that are in the

same show and less than 4 seconds apart are considered to be part of the

same story chunk. The adapted n-gram is then used to rescore the N-best

list.

23

Page 24: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

7 Experimental Results

7.1 Acoustic Segmentation

Tables 7 and 8 show the CERs with different segmenters at the MLP-SI

Search step and PLP-SA Search step, respectively, on Eval06. The error

distributions and our manual error analysis both show that the main ben-

efit of the new segmenter is in recovering lost speech segments and thus in

lowering deletion errors. However, those lost speech segments are usually of

lower speech quality and therefore lead to more substitution and insertion

errors. For comparison, we also show the CERs with the oracle segmenta-

tion as derived from the reference transcriptions. These results show that

our segmenter is very competitive.

Table 7: CERs at Step MLP-SI on Eval06.Segmenters Sub Del Ins OverallPrevious segmenter 9.7% 7.0% 1.9% 18.6%New segmenter 9.9% 6.4% 2.0% 18.3%Oracle segmenter 9.5% 6.8% 1.8% 18.1%

Table 8: CERs at Step PLP-SA on Eval06.Segmenters Sub Del Ins OverallPrevious segmenter 9.0% 5.4% 2.0% 16.4%New segmenter 9.2% 4.8% 2.1% 16.1%Oracle segmenter 8.8% 5.3% 2.0% 16.1%

24

Page 25: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

7.2 MLP Features

Since Mandarin is a tonal language, it is well known that adding pitch

information helps with speech recognition [2]. For this reason, we investigate

adding pitch into the input of the Tandem neural nets. For quick verification,

we used Hub4 to train nonCW triphone ML models. Table 9 shows the SI

bigram CER performance on Eval04. Pitch information obviously provides

extra information for both the MFCC front-end and the Tandem front-end.

It also demonstrates the discriminative power offered by the MLP feature,

compared with cepstrum features alone.

Table 9: SI bigram CERs on Eval04, using nonCW ML models trained onHub4.

HMM Feature MLP Input CERMFCC — 24.1%MFCC+F0 — 21.4%MFCC+F0+Tandem PLP 20.3%MFCC+F0+Tandem PLP+F0 19.7%

7.3 Acoustic and Language Model Adaptation

Table 10 shows the effect of acoustic model cross adaptation on Dev07,

following the decoding architecture in Figure 1.

Due to memory constraints, we are unable to adapt the full 4-gram LM.

Instead, we train 64 topic-dependent 4-grams and interpolate them with

qLM4 in Table 3. The adapted 4-gram is then applied to rescore the N-best

25

Page 26: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

Table 10: Decoding progress on Dev07. All AMs are trained on 866 hoursof speech.

LM MLP-SI PLP-SA MLP-SA CNCstatic qLM3 14.1% – – –static LM3 – 12.0% 11.9% –adapted qLM4 – 11.7% 11.4% 11.2%static LM4 – 11.9% 11.7% 11.4%

list of each utterance. The result is shown in the third row in Table 10.

Compared with the full static 4-gram in the next row, the adapted 4-gram

is slightly but consistently better.

7.4 System Combination

Finally, a character-level confusion network combination of the two rescored

N-best lists yields a 11.2% CER on Dev07, as shown in the last column of

Table 10. When this system was officially evaluated on Eval07 in June 2007,

we achieved the best GALE Mandarin ASR error rate of 9.1% (BN 3.4%,

BC 16.3%) 4 as shown in Table 11.

Table 11: Official CERs on Eval07. System parameter settings were tunedon Dev07.

LM MLP-SI PLP-SA MLP-SA CNCstatic qLM3 12.4% – – –static LM3 – 10.2% 9.6% –adapted qLM4 – 10.0% 9.6% 9.1%

4Our official result of 8.9% CER was obtained by CNC with yet another two N-bestlists generated by adapting our two systems with RWTH University top-1 hypotheses.

26

Page 27: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

8 Contribution and Future Work

This paper presents, step by step, a highly accurate Mandarin speech rec-

ognizer. It illustrates the language-independent technologies adopted and

details the language dependent modules that make our system a success.

Anecdotal error analysis on Dev07 shows that diphthongs did help in

examples such as ðL (/b ey3 d a4/, Beijing University), and merging /A/

and /a/ was not harmful. But merging /I/ and /IH/ into /i/ seemed to cause

somewhat more confusion among characters such as (4,�,�)=(shi,zhi,di).

Perhaps we need to reverse the last decision.

We have recently investigated a technique for improving the quality of

the MLP based features, adjusting the network’s outputs for training data

using knowledge of the Viterbi-aligned labels [27].

The topic-based LM adaptation had small, but not quite satisfactory,

improvement. Further refinement in the algorithm and in the implementa-

tion is needed to adapt the full 4-gram and obtain greater significance. Our

previous study [20] showed that full re-recognition with the adapted LM

offered more improvement than N-best rescoring. Yet the computation was

expensive. A lattice or word graph re-search is worth investigating.

Finally the system still has a much higher error rate on BC speech than

BN. There is not much BC text available to train conversation-specific lan-

guage models. Web crawling of more conversations is strongly needed. We

are working along all the above directions.

27

Page 28: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

Acknowledgment

The authors would like to express their gratitude to SRI International, for providing

Decipher as the backbone of this study, and particularly for all of the technical

support from A. Stolcke and Z. Jing. This material is based upon work supported

by the Defense Advanced Research Projects Agency (DARPA) under Contract No.

HR0011-06-C-0023. Any opinions, findings and conclusions or recommendations

expressed in this material are those of the authors and do not necessarily reflect

the views of DARPA.

References

[1] L. Mangu, E. Brill, and A. Stolcke, “Finding consensus in speech recogni-

tion: Word error minimization and other applications of confusion networks,”

Computer Speech and Language, pp. 373–400, 2000.

[2] M.Y. Hwang, X. Lei, T. NG, I. Bulyko, M. Ostendorf, A. Stolcke, W. Wang,

and J. Zheng, “Progress on mandarin conversational telephone speech recog-

nition,” in International Symposium on Chinese Spoken Language Processing,

2004.

[3] X. Lei, M.Y. Hwang, and M. Ostendorf, “Incorporating tone-related MLP

posteriors in feature representation for Mandarin ASR,” in Proc. Interspeech,

2005, pp. 2981–2984.

[4] X. Lei, M. Siu, M.Y. Hwang, M. Ostendorf, and T. Lee, “Improved tone mod-

eling for Mandarin broadcast news speech recognition,” in Proc. Interspeech,

2006.

[5] X. Lei and M. Ostendorf, “Word-level tone modeling for Mandarin speech

recognition,” in Proc. ICASSP, 2007.

28

Page 29: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

[6] A. Venkataraman, A. Stolcke, W. Wang, D. Vergyri, V. Gadde, and J. Zheng,

“An efficient repair procedure for quick transcriptions,” in Proc. ICSLP, 2004.

[7] T. NG, M. Ostendorf, M.Y. Hwang, M. Siu, I. Bulyko, and X. Lei, “Web data

augmented language models for Mandarin conversational speech recognition,”

in Proc. ICASSP, 2005, pp. 589–592.

[8] S. Chen and J. Goodman, “An empirical study of smoothing techniques for

language modeling,” Computer Science Group, Harvard University, TR-10-98,

1998.

[9] A. Stolcke, “SRILM-an Extensible Language Modeling Toolkit,” Seventh

International Conference on Spoken Language Processing, 2002.

[10] M.J.F. Gales, “Maximum likelihood linear transformations for HMM-based

speech recognition,” Computer Speech and Language, vol. 12, pp. 75–98, 1998.

[11] G. Peng, M.Y. Hwang, and M. Ostendorf, “Automatic acoustic segmentation

for speech recognition on broadcast recordings,” in Proc. Interspeech, 2007.

[12] J. Zheng, O. Cetin, M.Y. Hwang, X. Lei, A. Stolcke, and N. Morgan, “Combin-

ing discriminative feature, transform, and model training for large vocabulary

speech recognition,” ICASSP 2007.

[13] B. Chen, Q. Zhu, and N. Morgan, “Learning long-term temporal features in

LVCSR using neural networks,” in Proc. ICSLP, 2004.

[14] D. Povey and P.C. Woodland, “Minimum phone error and I-smoothing for

improved discriminative training,” in Proc. ICASSP, 2002.

[15] J. Zheng and A. Stolcke, “Improved discriminative training using phone lat-

tices,” in Proc. Interspeech, 2005, pp. 2125–2128.

[16] D. Povey, B. Kingsbury, L. Mangu, G. Saon, H. Soltau, and G. Zweig, “fMPE:

Discriminatively trained features for speech recognition,” in Proc. ICASSP,

2005.

29

Page 30: Building A Highly Accurate Mandarin Speech Recognizer with ...ssli.ee.washington.edu/~mhwang/pub/2008/manuscript.pdf · tomatic acoustic segmentation. We have published extraction

[17] C.J. Chen et al., “New methods in continuous Mandarin speech recognition,”

in Proc. Eur. Conf. Speech Communication Technology, 1997, vol. 3, pp. 1543–

1546.

[18] N. Morgan, B. Chen, Q. Zhu, and A. Stolcke, “Trapping conversational speech:

Extending trap/tandem approaches to conversational telephone speech recog-

nition,” in Proc. ICASSP, 2004, pp. 537–540.

[19] F. Valente and H. Hermansky, “Combination of acoustic classifiers based on

dempster-shafer theroy of evidence,” in Proc. ICASSP, 2007.

[20] M.Y. Hwang, W. Wang, X. Lei, J. Zheng, O. Cetin, and G. Peng, “Advances

in mandarin broadcast speech recognition,” in Proc. Interspeech, 2007.

[21] Q. Zhu, A. Stolcke, B.Y. Chen, and N. Morgan, “Using MLP features in SRI’s

conversational speech recognition system,” Proc. Interspeech, Lisbon, 2005.

[22] T. Anastasakos, J. McDonough, and J. Makhoul, “Speaker adaptive training:

A maximum likelihood approach to speaker normalization,” in International

Conference on Spoken Language Processing, 1996.

[23] M.Y. Hwang, X.D. Huang, and F. Alleva, “Predicting unseen triphones with

senones,” in Proc. ICASSP, 1993, pp. 311–314.

[24] T. Hofmann, “Probabilistic latent semantic analysis,” in Uncertainty in Arti-

ficial Intelligence, 1999.

[25] D.M. Blei, A.Y. NG, and M.I. Jordan, “Latent dirichlet allocation,” in The

Journal of Machine Learning Research, 2003, pp. 993–1022.

[26] A. Heidel and L.S. Lee, “Robust topic inference for latent semantic language

model adaptation,” in IEEE Automatic Speech Recognition and Understanding

Workshop, 2007.

[27] A. Faria and N. Morgan, “Corrected Tandem features for acoustic model

training,” in Proc. ICASSP, 2008.

30