Abstract
A natural language spoken dialog system includes a large vocabulary automatic speech recognition (ASR) engine, whose output is used as the input of a spoken language understanding component. Two challenges in such a framework are that the ASR component is far from being perfect and the users can say the same thing in very different ways. So, it is very important to be tolerant to recognition errors and some amount of orthographic variability. In this paper, we present our work on developing new methods and investigating various ways of robust recognition and understanding of an utterance. To this end, we exploit word-level confusion networks (sausages), obtained from ASR word graphs (lattices) instead of the ASR 1-best hypothesis. Using sausages with an improved confidence model, we decreased the calltype classification error rate for AT&T's How May I Help YouSM (HMIHYSM) natural dialog system by 38%.
Original language | English (US) |
---|---|
Pages | 1137-1140 |
Number of pages | 4 |
State | Published - 2002 |
Externally published | Yes |
Event | 7th International Conference on Spoken Language Processing, ICSLP 2002 - Denver, United States Duration: Sep 16 2002 → Sep 20 2002 |
Other
Other | 7th International Conference on Spoken Language Processing, ICSLP 2002 |
---|---|
Country/Territory | United States |
City | Denver |
Period | 9/16/02 → 9/20/02 |
ASJC Scopus subject areas
- Language and Linguistics
- Linguistics and Language