automatic speech recognition on mobile devices and over

11
Zheng-Hua Tan and Berge Lindberg Automatic Speech Recognition on Mobile Devices and over Communication Networks ^Spri inger g<

Upload: others

Post on 05-Feb-2022

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Automatic Speech Recognition on Mobile Devices and over

Zheng-Hua Tan and Berge Lindberg

Automatic Speech Recognition on Mobile Devices and over Communication Networks

^ S p r i inger g<

Page 2: Automatic Speech Recognition on Mobile Devices and over

Contents

Preface v

Contributors xix

1. Network, Distributed and Embedded Speech Recognition: An Overview 1 Zheng-Hua Tan and Imre Varga 1.1 Introduction 1 1.2 ASR and Its Deployment in Devices and Networks 3

1.2.1 Automatic Speech Recognition 3 1.2.2 Resources and Constraints of Mobile Devices 5 1.2.3 Resources and Constraints of Communication Networks 7 1.2.4 Architectural Solutions for ASR in Devices and Networks 8

1.3 Network Speech Recognition 9 1.4 Distributed Speech Recognition 11

1.4.1 Feature Extraction 11 1.4.2 Source Coding 12 1.4.3 Channel Coding and Packetisation 13 1.4.4 Error Concealment 14 1.4.5 DSR Standards 14 1.4.6 A Configurable DSR System 15

1.5 Embedded Speech Recognition 15 1.5.1 ESRScenario 16 1.5.2 Applications and Platforms 16 1.5.3 Fixed-Point Arithmetic 17 1.5.4 Optimisation 18 1.5.5 Robustness 19

1.6 Discussion 20 References 21

Page 3: Automatic Speech Recognition on Mobile Devices and over

x Contents

Part I Network Speech Recognition

2. Speech Coding and Packet Loss Effects on Speech and Speaker Recognition 27 Laurent Besacier 2.1 Introduction 27 2.2 Sources of Degradation in Network Speech Recognition 28

2.2.1 Speech and Audio Coding Standards 28 2.2.2 Packet Loss 30

2.3 Effects on the Automatic Speech Recognition Task 32 2.3.1 Experimental Setup 32 2.3.2 Degradation Due to Simulated Packet Loss 32 2.3.3 Degradation with Real Transmissions 33 2.3.4 Degradation Due to Speech and Audio Codecs 34

2.4 Effect for the Automatic Speaker Verification Task 35 2.4.1 Speaker Verification Experiments Over Compressed Speech

and Packet Loss 36 2.4.2 Speaker Verification Experiments Over GSM Compressed

Speech 37 2.5 Conclusion 38 Acknowledgments 38 References 39

3. Speech Recognition Over Mobile Networks 41 Hong Kook Kim and Richard C. Rose 3.1 Introduction 41 3.2 Techniques for Improving ASR Performance Over Mobile Networks... 43 3.3 Bitstream-Based Approach 46 3.4 Feature Transform 50

3.4.1 Mel-Scaled LPCC 51 3.4.2 LPC-Based MFCC (LP-MFCC) 52 3.4.3 Pseudo-Cepstrum (PCEP) and Its Mel-Scaled Variant

(MPCEP) 53 3.5 Enhancement of ASR Performance Over Mobile Networks 53

3.5.1 Compensation for the Effect of Mobile Systems 53 3.5.2 Compensation for Speech Coding Distortion in LSP Domain 54 3.5.3 Compensation for Channel Errors 56

3.6 Conclusion 57 References 58

4. Speech Recognition Over IP Networks 63 Hong Kook Kim 4.1 Introduction 63 4.2 Speech Recognition and IP Networks 65

4.2.1 Relationship Between ASR Performance and Speech Quality 65 4.2.2 Impact of Speech Coding Distortion 66 4.2.3 Impact of Network Channel Distortion 67

Page 4: Automatic Speech Recognition on Mobile Devices and over

Contents xi

4.3 Robustness Against Packet Loss 69 4.3.1 Rate Control 69 4.3.2 Forward Error Correction 70 4.3.3 Interleaving 70 4.3.4 Error Concealment and ASRDecoder- Based Concealment 71

4.4 Speech Coder for Speech Recognition Over IP Networks 71 4.4.1 MFCC-Based Speech Coder 72 4.4.2 Efficient Vector Quantization of MFCCs 74 4.4.3 Speech Quality Comparison 78 4.4.4 ASR Performance Comparison 79

4.5 Conclusion 82 References 82

Part II Distributed Speech Recognition

5. Distributed Speech Recognition Standards 87 David Pearce 5.1 Introduction 87 5.2 Overview of the Set of DSR Standards 89 5.3 Scope of the Standards 90

5.3.1 Electro-Acoustics 91 5.3.2 Speech Detection or External Control Signal 92 5.3.3 Pre-Processing 92 5.3.4 Parameterisation 92 5.3.5 Compression and Error Protection 93 5.3.6 Formatting 93 5.3.7 Error Detection and Mitigation 93 5.3.8 Decompression 93 5.3.9 Server Side Post Processing 93 5.3.10 Feature Derivatives 93

5.4 DSR Basic Front-End ES 201 108 94 5.4.1 Feature Extraction 94 5.4.2 Compression 94 5.4.3 Error Detection and Mitigation 95

5.5 DSR Advanced Front-End ES 202 050 96 5.5.1 Feature Extraction 96 5.5.2 VAD 96 5.5.3 Compression 96

5.6 Recognition Performance of the DSR Front-Ends 97 5.6.1 Aurora Speech Databases and ETSI Performance Testing 97 5.6.2 Aurora 3: Multilingual SpeechDat-Car Digits — Small

Vocabulary Evaluation 97 5.7 3GPP Evaluations and Comparisons to AMR Coded Speech 99 5.8 ETSI DSR Extended Front-End Standards ES 202 211 and ES 202 212 102 5.9 Transport Protocols: The IETF RTP Payload Formats for DSR 104 5.10 Conclusion 105 Acknowledgements 105 References 105

Page 5: Automatic Speech Recognition on Mobile Devices and over

xii Contents

6. Speech Feature Extraction and Reconstruction 107 Ben Milner 6.1 Introduction 107 6.2 Feature Extraction 109

6.2.1 Basic Terminal-Side Feature Extraction 109 6.2.2 Advanced Terminal-Side Feature Extraction 115 6.2.3 Quantisation and Packetisation 116 6.2.4 Server-Side Processing 117

6.3 Speech Reconstruction 117 6.3.1 Analysisof Received Speech Information 118 6.3.2 Speech Reconstruction 119

6.4 Prediction of Voicing and Fundamental Frequency 123 6.4.1 Fundamental Frequency Prediction from MFCC Vectors 123 6.4.2 Voicing Prediction from MFCC Vectors 126 6.4.3 Speech Reconstruction from Predicted Fundamental

Frequency and Voicing 128 6.5 Conclusion 129 References 129

7. Quantization of Speech Features: Source Coding 131 Stephen So and Kuldip K. Paliwal 7.1 Introduction 131 7.2 Quantization Schemes 132

7.2.1 Brief Introduction to Quantization Theory 132 7.2.2 Distortion Measures for Quantization in Speech Processing 134 7.2.3 Scalar Quantization 135 7.2.4 Block Quantization 137 7.2.5 Vector Quantization 137 7.2.6 GMM-Based Block Quantization 138

7.3 Quantization of ASR Feature Vectors 141 7.3.1 Introduction and Literature Review 141 7.3.2 Statistical Properties of MFCCs 142 7.3.3 Use of Cepstral Liftering for MFCC Variance Normahzation 148 7.3.4 Relationship Between the Distortion Measure and

Recognition Performance 150 7.3.5 Improving Noise Robustness: Perceptual Weighting

of Filterbank Energies 152 7.4 Experimental Results 153

7.4.1 ETSI Aurora-2 Distributed Speech Recognition Task 153 7.4.2 Experimental Setup 154 7.4.3 Non-Uniform Scalar Quantization Using HRO Bit Allocation 154 7.4.4 Unconstrained Vector Quantization 155 7.4.5 GMM-Based Block Quantization 156 7.4.6 Multi-frame GMM-Based Block Quantization 156 7.4.7 Perceptually-Weighted Vector Quantization of Logarithmic

Filterbank Energies 157 7.5 Conclusion 158 References 159

Page 6: Automatic Speech Recognition on Mobile Devices and over

Contents xiii

8. Error Recovery: Channel Coding and Packetization 163 BengtJ. Borgström, Alexis Bernard, and Abeer Alwan 8.1 Distributed Speech Recognition Systems 163 8.2 Characterization and Modeling of Communication Channels 164

8.2.1 Signal Degradation Over Wireless Communication Channels 164 8.2.2 Signal Degradation Over IP Networks 165 8.2.3 Modeling Bursty Communication Channels 165

8.3 Media-Specific FEC 167 8.4 Media-Independent FEC 168

8.4.1 Combining FEC with Error Concealment Methods 169 8.4.2 Linear Block Codes 169 8.4.3 Cyclic Codes 174 8.4.4 Convolutional Codes 174

8.5 Unequal Error Protections 176 8.6 Frame Interleaving 177

8.6.1 Optimal Spread Block Interleavers 178 8.6.2 Convolutional Interleavers 179 8.6.3 Decorrelated Block Interleavers 180

8.7 Examples of Modern Error Recovery Standards 181 8.7.1 ETSI DSR Standard (ETSI 2000) 181 8.7.2 ETSI GSM/EFR Standard (ETSI 1998) 182

8.8Summary 183 Acknowledgements 184 References 184

9. Error Concealment 187 Reinhold Haeb-Umbach and Valentin Ion 9.1 Introduction 187 9.2 Speech Recognition in the Presence of Corrupted Features 190

9.2.1 Modified Observation Probability 190 9.2.2 Gaussian Approximation 193

9.3 Feature Posterior Estimation in a DSR Framework 194 9.3.1 ETSI DSR Standards 195 9.3.2 Source Coder Redundancy 195 9.3.3 Channel Models 196 9.3.4 Estimation of Feature Posterior 199 9.3.5 Related Work 201

9.4 Performance Evaluations 202 9.4.1 Experimental Setup 202 9.4.2 Results on GSM Data Channel 203 9.4.3 Results on Packet Erasure Channel 206

9.5 Conclusion 207 Acknowledgments 208 References 208

Page 7: Automatic Speech Recognition on Mobile Devices and over

xiv Contents

Part III Embedded Speech Recognition

10. Algorithm Optimizations: Low Computational Complexity 213 Miroslav Novak 10.1 Introduction 213 10.2 Common Limitations of Embedded Platforms 214

10.2.1 Memory Limitations 214 10.2.2 CPU Limitations 215

10.3 Overview of an ASR System 215 10.4 Front End 216 10.5 Observation Model 217

10.5.1 Model Organization 217 10.5.2 Efficient Computation Strategies 218

10.6Search 221 10.6.1 Viterbi Search Implementation 222 10.6.2 Search Graph Construction 226 10.6.3 Fast Match 228 10.6.4 Alternative Decoding Schemes 228

10.7 Conclusion 229 Acknowledgments 229 References 230

11. Algorithm Optimizations: Low Memory Footprint 233 Marcel Vasilache 11.1 Introduction 233 11.2 Notations and Problem Statement 234 11.3 Model Complexity Control 237

11.3.1 Akaike's Information Criterion 238 11.3.2 Bayesian Information Criterion 238 11.3.3 Second Order Approximation 239 11.3.4 Other Measures 239

11.4 Parameter Tying 239 11.4.1 Model Level 240 11.4.2 State Level 241 11.4.3 Density Level 241 11.4.4 Subspaces 241 11.4.5 Clustering 242

11.5 Parameter Representations 243 11.5.1 Floating Point Representation 243 11.5.2 Fixed Point Representation 244 11.5.3 Quantization 244

11.6 Quantized Parameters HMMs 245 11.6.1 Scalar Quantization 245 11.6.2 Vector Quantization 247

11.7 Subspace Distribution Clustering HMM 247 11.7.1 Subspace Partitioning 248 11.7.2 Density Clustering 249

Page 8: Automatic Speech Recognition on Mobile Devices and over

Contents xv

11.8 Computational Complexity Implications 249 11.9 Practicalities and Conclusion 250 References 251

12. Fixed-Point Arithmetic 255 Enrico Bocchieri 12.1 Introduction 255 12.2 Fixed-Point Arithmetic 257

12.2.1 Programming with Fixed-Point Numbers 257 12.2.2 Fixed-Point Representation and Quantization 259

12.3 LVCSR MAP Recognizer 259 12.3.1 HMM State Likelihoods 261 12.3.2 State Duration Model 262 12.3.3 Language Model 263 12.3.4 Viterbi Decoder 263 12.3.5 Acoustic Front-End 264

12.4 Fixed-Point Implementation of the Recognizer 264 12.4.1 Log-Likelihoods 265 12.4.2 Viterbi Frame-Synchronous Search 266 12.4.3 Gaussian Parameters 267 12.4.4 MFCC Front-End 268

12.5 Experiments 269 12.5.1 Real-Time on the Device 272

12.6 Conclusion 274 Acknowledgements 274 References 274

Part IV Systems and Applications

13. Software Architectures for Networked Mobile Speech Applications.... 279 James C. Ferrans and Jonathan Engelsma 13.1 Introduction 279

13.1.1 Embedded and Distnbuted Speech Engines 279 13.1.2 The Voice Web 280 13.1.3 Multimodal User Interfaces 283 13.1.4 Distributed Speech Recognition 284 13.1.5 Multimodal Architectures 285 13.1.6 Simultaneous and Sequential Multimodality 287 13.1.7 Mode Composition 288

13.2 Classesof Multimodal Architectures 288 13.2.1 Fully Embedded or "Fat Client" (a) 289 13.2.2 Distributed Processing Engines (b) 289 13.2.3 Thin Client (d) 291 13.2.4 Remote Visual Interface (e) 291 13.2.5 "Pudgy" Client (c) 292 13.2.6 Discussion 292

13.3 The "Plus V" Distributed Multimodal Architecture 293 13.4 Other Distributed Multimodal Architectures 295

Page 9: Automatic Speech Recognition on Mobile Devices and over

xvi Contents

13.4.1 Video Interactive Services with VoiceXML 295 13.4.2 Multimodal for Set-Top Boxes 295 13.4.3 Bare Minimum Mobile Voice Search 296 13.4.4 A Transcription-Based Architecture 297

13.5 Towards a Commercial Ecosystem 297 13.6Conclusion 298 References 298

14. Speech Recognition in Mobile Phones 301 Imre Varga and Imre Kiss 14.1 Introduction 301 14.2 Applications of Speech Recognition for Mobile Phones 302 14.3 Multilinguality and Language Support 305

14.3.1 Multilingual Speaker Independent Name Dialing 305 14.3.2 Multilinguality in Other ASR Applications 308 14.3.3 Language Resources 308

14.4 Noise Robustness 309 14.4.1 Robust HMM Models 309 14.4.2 Feature Extraction 309 14.4.3 Noise Reduction 310

14.5 Footprint and Complexity Reduction 314 14.5.1 Footprint Reduction of Acoustic Models 314 14.5.2 Footprint Reduction of Language Models 315 14.5.3 Footprint Reduction of Pronunciation Lexicon 317 14.5.4 Reduction of Computational Complexity in Embedded ASR

Systems 317 14.5.5 Low Memory, Fast Decoding 319

14.6 Platforms and an Example Application 319 14.6.1 Example Application: Large Vocabulary Isolated Word

Dictation 320 14.7 Conclusion and Outlook 323 References 323

15. Handheld Speech to Speech Translation System 327 Yuqing Gao, Bowen Zhou, Weizhong Zhu and Wei Zhang 15.1 Introduction 327 15.2 System Overview 328

15.2.1 System architecture 328 15.2.2 Hardware and OS Specifications 330 15.2.3 Interface 330

15.3 System Components and Optimization 332 15.3.1 LVCSR on Handheld Devices 332 15.3.2 Natural Language Understanding and Generation Based

Translation 334 15.3.3 Weighted Finite State Transducer Based Translation 337 15.3.4 Embedded Speech Synthesis 340

Page 10: Automatic Speech Recognition on Mobile Devices and over

Contents xvii

15.4 Experiments and Discussions 341 15.4.1 Speech Recognition Experiments 341 15.4.2 Translation Experiments 343

15.5 Conclusion 344 References 345

16. Automotive Speech Recognition 347 Harald Höge, Sascha Hohenner, Bernhard Kämmerer, Niels Kunstmann, Stefanie Schachtl, Martin Schönle, and Panji Setiawan 16.1 Introduction 347 16.2 Siemens Speech Processing — From Research to Products 348

16.2.1 Development for Performance andQuality 348 16.2.2 High-Performance Recognizer 349 16.2.3 Ultra-Compact Text-to-Speech Synthesizer 350 16.2.4 Natural Voice Dialog 351 16.2.5 Speaker Characterization and Recognition 351

16.3 Example Automotive Voice Applications: Infotainment, Navigation, Manuals, and Internet 351 16.3.1 Radio Station Selection 352 16.3.2 MP3 Title Selection 352 16.3.3 Navigation Destination Entry 353 16.3.4 Manuals and Help Systems 354 16.3.5 Access to Structured Web Content 355 16.3.6 Access to Web Services 356

16.4 Automotive Platform Issues and Challenges 357 16.4.1 Hardware Constraints 358 16.4.2 Software Constraints 359 16.4.3 User Constraints 360 16.4.4 Acoustic Channel 360

16.5 Noise Robust Recognition Technology 360 16.5.1 ASRFront-End 362 16.5.2 Minimum Mean Square Weighting Rules 363 16.5.3 Recursive Least Squares Weighting Rules 364 16.5.4 Implementation of RLS Weighting Rules 365 16.5.5 Recognition Results 366

16.6 Methodology for Evaluation of Automotive Recognizers Quality Measurement Using SNR Curves 367 16.6.1 Common Evaluation Procedures 368 16.6.2 Proposed SNR-Approach 368 16.6.3 Data Recording 368 16.6.4 Evaluation 369 16.6.5 Best Practice 371

16.7 Conclusion 372 References 372

Page 11: Automatic Speech Recognition on Mobile Devices and over

xviii Contents

17. Energy Aware Speech Recognition for Mobile Devices 375 Brian Delaney 17.1 Introduction 375

17.1.1 Battery Technology 375 17.1.2 Energy Aware Design Principles 376 17.1.3 Related Work 377

17.2 Case Study of Distributed Speech Recognition Using the HP Labs Smartbadge System 379 17.2.1 Signal Processing Front-End 379 17.2.2 Energy Consumption of DSR with IEEE 802.11 Wireless

Networks 384 17.2.3 Energy Consumption of DSR Using Bluetooth Networks 389 17.2.4 Comparison of 802.11 and Bluetooth in DSR 391

17.3 Conclusion 395 References 395

Index 397