Acoustical and Environmental Robustness in Automatic Speech Recognition / Edition 1by Alex Acero
Pub. Date: 11/15/1992
Publisher: Springer US
The need for automatic speech recognition systems to be robust with respect to changes in their acoustical environment has become more widely appreciated in recent years, as more systems are finding their way into practical applications. Although the issue of environmental robustness has received only a small fraction of the attention devoted to speaker independence,… See more details below
The need for automatic speech recognition systems to be robust with respect to changes in their acoustical environment has become more widely appreciated in recent years, as more systems are finding their way into practical applications. Although the issue of environmental robustness has received only a small fraction of the attention devoted to speaker independence, even speech recognition systems that are designed to be speaker independent frequently perform very poorly when they are tested using a different type of microphone or acoustical environment from the one with which they were trained. The use of microphones other than a "close talking" headset also tends to severely degrade speech recognition -performance. Even in relatively quiet office environments, speech is degraded by additive noise from fans, slamming doors, and other conversations, as well as by the effects of unknown linear filtering arising reverberation from surface reflections in a room, or spectral shaping by microphones or the vocal tracts of individual speakers. Speech-recognition systems designed for long-distance telephone lines, or applications deployed in more adverse acoustical environments such as motor vehicles, factory floors, oroutdoors demand far greaterdegrees ofenvironmental robustness. There are several different ways of building acoustical robustness into speech recognition systems. Arrays of microphones can be used to develop a directionally-sensitive system that resists intelference from competing talkers and other noise sources that are spatially separated from the source of the desired speech signal.
- Springer US
- Publication date:
- Springer International Series in Engineering and Computer Science, #201
- Edition description:
- Product dimensions:
- 6.10(w) x 9.25(h) x 0.24(d)
Table of ContentsList of Figures. List of Tables. Foreword. 1. Introduction. 2. Experimental Procedure. 3. Frequency Domain Processing. 4. The SDCN Algorithm. 5. The CDCN Algorithm. 6. Other Algorithms. 7. Frequency Normalization. 8. Summary of Results. 9. Conclusions. Appendix I: Glossary. Appendix II: Signal Processing in Sphinx. Appendix III: The Bilinear Transform. Appendix IV: Spectral Estimation Issues. Appendix V: MMSE Estimation in the CDCN Algorithm. Appendix VI: Maximum Likelihood via the EM Algorithm. Appendix VII: Estimation of Noise and Spectral Tilt. Appendix VIII: Vocabulary and Pronunciation Dictionary. References. Index.
and post it to your social network
Most Helpful Customer Reviews
See all customer reviews >