Search results for: “stolcke”
-
A Cross-language Study on Automatic Speech Disfluency Detection
We investigate two systems for automatic disfluency detection on English and Mandarin conversational speech data. The first system combines various lexical and prosodic features in a Conditional Random Field model for detecting edit disfluencies.
-
Highly Accurate Phonetic Segmentation Using Boundary Correction Models and System Fusion
Accurate phone-level segmentation of speech remains an important task for many subfields of speech research. We investigate techniques for boosting the accuracy of automatic phonetic segmentation based on HMM acoustic-phonetic models. In prior work we were able to improve on state-of-the-art alignment accuracy by employing special phone boundary HMM models, trained on phonetically segmented training…
-
Highly Accurate Phonetic Segmentation Using Correction Models and System Fusion
We investigate techniques for boosting the accuracy of automatic phonetic segmentation based on HMM acoustic-phonetic models.
-
Automatic phonetic segmentation using boundary models
This study attempts to improve automatic phonetic segmentation within the HMM framework.
-
Articulatory trajectories for large-vocabulary speech recognition
We present a neural network model to estimate articulatory trajectories from speech signals where the model was trained using synthetic speech signals generated by Haskins Laboratories’ task-dynamic model of speech production.
-
Using multiple versions of speech input in phone recognition
This study investigates the use of multiple versions of the same speech unit in automatic phone recognition.
-
Effects of audio and ASR quality on cepstral and high-level speaker verification systems
We evaluate the effect that improved audio quality has for speaker verification performance, using a recently released full-bandwidth version of microphone data from the SRE2010 evaluation.
-
Speaker recognition with region-constrained MLLR transforms
We define regions based on phonetic and prosodic criteria, based on automatic speech recognition output, and perform MLLR estimation using only frames selected by these criteria.
-
SRILM at sixteen: Update and outlook
We review developments in the SRI Language Modeling Toolkit (SRILM) since 2002, when a previous paper on SRILM was published.
-
SRILM at 16: Update and Outlook
We review developments in the SRI Language Modeling Toolkit (SRILM) since 2002, when a previous paper on SRILM was published.
-
Effective Arabic dialect classification using diverse phonotactic models
We study the effectiveness of recently developed language recognition techniques based on speech recognition models for the discrimination of Arabic dialects.
-
Constrained cepstral speaker recognition using matched UBM and JFA training
We study constrained speaker recognition systems, or systems that model standard cepstral features that fall within particular types of speech regions.