Signal Processing Research Group
Contents
Top
Research Topics
People
Publications
Organization
NTT Communication Science Laboratories
Media Information Laboratory
Recognition Research Group
Signal Processing Research Group
Communication environment research group
Innovative Communication Laboratory
Human and Information Science Laboratory
Moriya Research Laboratory
NTT Science and Core Technology Laboratory Group
Nippon Telegraph and Telephone Corporation (NTT)
Research Topics
Speech recognition for computers
Fast and highly accurate spontaneous speech recognition


 
Automatic speech recognition (ASR) technology endeavors to endow computers with a primary human communication channel, namely speech. Given an audio signal, ASR systems identify segments that contain a speech signal. For each speech segment, a sequence of information-rich feature vectors is extracted and matched to the most likely word sequence given a set of previously-trained speech models reflecting the salient features of a languages’ phonemes. Although they may correspond to the same spoken word content, real world speech signals vary greatly depending on the speaker and on the acoustic environment, especially for spontaneous speech in natural human conversation, where speech patterns are highly diverse, ambiguous and incomplete. Human beings can absorb this variation remarkably well and can fill in missing components instantly. Our research on speech recognition technology aims to bring human faculties for high performance, high speed and robust speech recognition to computers.
 


* How is speech recognition technology used?

Speech recognition technology has many applications including the control of consumer electronic devices, the automatic generation of meeting minutes, and the captioning of audio/visual content. With effective ASR, conversational robots that can understand human speech could become a reality. Leveraging the high-speed data processing ability of computers, large amounts of speech data can be analyzed, organized, summarized, and translated. In the future, speech recognition technology will be used to give "ears" to robots, and speech analyzers will be a part of daily life. To make ASR effective for various applications, we are working to improve algorithms for speech analysis, model training, search, and backend processing.
 
* Integration technique for speech enhancement and speech recognition

In real world environments, the interference originating from various kinds of ambient noise means that it is often difficult for automatic speech recognition (ASR) to transcribe audio signals accurately. Speech enhancement techniques improve the audible quality of speech signals. However, the improvement in terms of recognition accuracy is usually limited because the enhanced speech inevitably includes distortion and residual noise. We have proposed a method to mitigate these effects on ASR by dynamically compensating the acoustic model parameters based on a reliability estimation of the enhanced speech signals. This method enables us to integrate various kinds of speech enhancement techniques with ASR.
Integration technique for speech
enhancement and speech recognition
 
* Highly accurate speech modeling based on discriminative training

We have established several methods for constructing highly accurate acoustic and language models, such as dMMI(*1)-based discriminative acoustic model training and R2D2(*2)-based discriminative language model training methods. These methods provide models with better than ever performance in terms of discrimination and generalization. We have also developed a WFST(*3)-based linear classifier that can be trained directly on a WFST composed of acoustic and language models. This reduces recognition errors that cannot be recovered using only separately trained acoustic and language models.
Procedure of dMMI-based discriminative acoustic model training
 
* Real-time meeting recognition and understanding

Since ASR is used in a variety of situations, it needs to handle multi-speaker conversations on a wide variety of topics, and yet be sufficiently fast and memory efficient. Speech applications must both convert speech signals into text, and recognize rich information such as the speaker, topic, circumstance and confidence of the ASR result. We have developed a very fast algorithm that makes it possible to achieve real-time ASR with a 10 million word vocabulary using WFSTs. We have also developed topic tracking language models and a confidence estimation method with error causes for ASR results.
Real-time meeting analysis system that recognizes who spoke when and what
 
*1 differenced Maximum Mutual Information
*2 Round Robin Duel Discrimination
*3 Weighted Finite-State Transducer
 
[ Reference ]
[1] T. Hori, C. Hori, Y. Minami, and A. Nakamura, "Efficient WFST-based one-pass decoding with on-the-fly hypothesis rescoring in extremely large vocabulary continuous speech recognition," IEEE Transactions on Audio, Speech and Language Processing, Vol. 15, pp. 1352―1365, 2007.
[2] M. Delcroix, T. Nakatani, and S. Watanabe, "Static and dynamic variance compensation for recognition of reverberant speech with dereverberation preprocessing," IEEE Transactions on Audio, Speech, and Language Processing, vol. 17, no. 2, pp. 324-334, 2009.
[3] E. McDermott, S. Watanabe, and A. Nakamura, "Discriminative training based on an integrated view of MPE and MMI in margin and error space," Proc. ICASSP'10, pp. 4894-4897, 2010.
[4] T. Hori, S. Araki, T. Yoshioka, M. Fujimoto, S. Watanabe, T. Oba, A. Ogawa, K. Otsuka, D. Mikami, K. Kinoshita, T. Nakatani, A. Nakamura, J. Yamato, "Low-latency real-time meeting recognition and understanding using distant microphones and omni-directional camera," IEEE Transactions on Audio, Speech, and Language Processing, Vol. 20, No. 2, pp. 499―513, 2012.
[5] T. Oba, T. Hori, A. Nakamura, A. Ito, "Round-Robin Duel Discriminative Language Models," IEEE Transactions on Audio, Speech and Language Processing, Vol. 20, No. 4, pp. 1244-1255, May 2012.
[6] Y. Kubo, S. Watanabe, T. Hori, A. Nakamura, "Structural Classification Methods based on Weighted Finite-State Transducers for Automatic Speech Recognition," IEEE Transactions on Audio, Speech, and Language Processing, Vol. 20, Issue 8, pp. 2240―2251, 2012.
[7] M. Delcroix, K. Kinoshita, T. Nakatani, S. Araki, A. Ogawa, T. Hori, S. Watanabe, M. Fujimoto, T. Yoshioka, T. Oba, Y. Kubo, M. Souden, S.-J. Hahm, and A. Nakamura, "Speech recognition in living rooms: Integrated speech enhancement and recognition system based on spatial, spectral & temporal modeling of sounds," Computer Speech and Language, Elsevier, vol. 27, no. 3, pp. 851-873, 2013.
[8] Y. Kubo, T. Hori, A. Nakamura, "Large Vocabulary Continuous Speech Recognition Based on WFST Structured Classifiers and Deep Bottleneck Features," Proc. ICASSP, 2013.
[9] M. Delcroix, S. Watanabe, T. Nakatani, and A. Nakamura, "Cluster-based dynamic variance adaptation for interconnecting speech enhancement pre-processor and speech recognizer," Computer Speech and Language, Elsevier, vol. 27, no. 1, pp. 350-368, 2013.
[10] T. Hori, Y. Kubo, A. Nakamura, "Real-time one-pass decoding with recurrent neural network language model for speech recognition," in Proc. ICASSP, 2014.
english   japanese