TY - GEN
T1 - Alexa Conversations
T2 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021
AU - Acharya, Anish
AU - Adhikari, Suranjit
AU - Agarwal, Sanchit
AU - Auvray, Vincent
AU - Belgamwar, Nehal
AU - Biswas, Arijit
AU - Chandra, Shubhra
AU - Chung, Tagyoung
AU - Fazel-Zarandi, Maryam
AU - Gabriel, Raefer
AU - Gao, Shuyang
AU - Goel, Rahul
AU - Hakkani-Tur, Dilek
AU - Jezabek, Jan
AU - Jha, Abhay
AU - Kao, Jiun Yu
AU - Krishnan, Prakash
AU - Ku, Peter
AU - Goyal, Anuj
AU - Lin, Chien Wei
AU - Liu, Qing
AU - Mandal, Arindam
AU - Metallinou, Angeliki
AU - Naik, Vishal
AU - Pan, Yi
AU - Paul, Shachi
AU - Perera, Vittorio
AU - Sethi, Abhishek
AU - Shen, Minmin
AU - Strom, Nikko
AU - Wang, Eddie
N1 - Publisher Copyright:
© 2021 Association for Computational Linguistics.
PY - 2021
Y1 - 2021
N2 - Traditional goal-oriented dialogue systems rely on various components such as natural language understanding, dialogue state tracking, policy learning and response generation. Training each component requires annotations which are hard to obtain for every new domain, limiting scalability of such systems. Similarly, rule-based dialogue systems require extensive writing and maintenance of rules and do not scale either. End-to-End dialogue systems, on the other hand, do not require module-specific annotations but need a large amount of data for training. To overcome these problems, in this demo, we present Alexa Conversations, a new approach for building goal-oriented dialogue systems that is scalable, extensible as well as data efficient. The components of this system are trained in a data-driven manner, but instead of collecting annotated conversations for training, we generate them using a novel dialogue simulator based on a few seed dialogues and specifications of APIs and entities provided by the developer. Our approach provides out-of-the-box support for natural conversational phenomena like entity sharing across turns or users changing their mind during conversation without requiring developers to provide any such dialogue flows. We exemplify our approach using a simple pizza ordering task and showcase its value in reducing the developer burden for creating a robust experience. Finally, we evaluate our system using a typical movie ticket booking task and show that the dialogue simulator is an essential component of the system that leads to over 50% improvement in turn-level action signature prediction accuracy.
AB - Traditional goal-oriented dialogue systems rely on various components such as natural language understanding, dialogue state tracking, policy learning and response generation. Training each component requires annotations which are hard to obtain for every new domain, limiting scalability of such systems. Similarly, rule-based dialogue systems require extensive writing and maintenance of rules and do not scale either. End-to-End dialogue systems, on the other hand, do not require module-specific annotations but need a large amount of data for training. To overcome these problems, in this demo, we present Alexa Conversations, a new approach for building goal-oriented dialogue systems that is scalable, extensible as well as data efficient. The components of this system are trained in a data-driven manner, but instead of collecting annotated conversations for training, we generate them using a novel dialogue simulator based on a few seed dialogues and specifications of APIs and entities provided by the developer. Our approach provides out-of-the-box support for natural conversational phenomena like entity sharing across turns or users changing their mind during conversation without requiring developers to provide any such dialogue flows. We exemplify our approach using a simple pizza ordering task and showcase its value in reducing the developer burden for creating a robust experience. Finally, we evaluate our system using a typical movie ticket booking task and show that the dialogue simulator is an essential component of the system that leads to over 50% improvement in turn-level action signature prediction accuracy.
UR - http://www.scopus.com/inward/record.url?scp=85135792096&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85135792096&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85135792096
T3 - NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Demonstrations
SP - 125
EP - 132
BT - NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics
PB - Association for Computational Linguistics (ACL)
Y2 - 6 June 2021 through 11 June 2021
ER -