Context dependent speech recognition
Item statusRestricted Access
MetadataShow full item record
Poor speech recognition is a problem when developing spoken dialogue systems, but several studies has showed that speech recognition can be improved by post-processing of recognition output that use the dialogue context, acoustic properties of a user utterance and other available resources to train a statistical model to use as a filter between the speech recogniser and dialogue manager. In this thesis a corpus of logged interactions between users and a dialogue system was used to extract features from previous dialogue context, acoustics from the user utterance and n-best recognition hypotheses. The features were used to train maximum entropy models with different feature sets to rerank the n-best hypotheses. The models fail to some extent to predict intended labels but using the reranked output in effect means that 94.9% of the adequate hypotheses will be sent to the dialogue manager, a decrease in relative error over baseline with 44.6% showing that contextual reranking can improve speech recognition for dialogue systems. Future work involves developing the current feature sets and maxEnt models to better classify whether a hypothesis should be accepted or rejected by the dialogue system rather than rerank them.