TY - JOUR
T1 - Using recurrent neural networks for slot filling in spoken language understanding
AU - Mesnil, Grégoire
AU - Dauphin, Yann
AU - Yao, Kaisheng
AU - Bengio, Yoshua
AU - Deng, Li
AU - Hakkani-Tur, Dilek
AU - He, Xiaodong
AU - Heck, Larry
AU - Tur, Gokhan
AU - Yu, Dong
AU - Zweig, Geoffrey
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2015/3/1
Y1 - 2015/3/1
N2 - Semantic slot filling is one of the most challenging problems in spoken language understanding (SLU). In this paper, we propose to use recurrent neural networks (RNNs) for this task, and present several novel architectures designed to efficiently model past and future temporal dependencies. Specifically, we implemented and compared several important RNN architectures, including Elman, Jordan, and hybrid variants. To facilitate reproducibility, we implemented these networks with the publicly available Theano neural network toolkit and completed experiments on the well-known airline travel information system (ATIS) benchmark. In addition, we compared the approaches on two custom SLU data sets from the entertainment and movies domains. Our results show that the RNN-based models outperform the conditional random field (CRF) baseline by 2% in absolute error reduction on the ATIS benchmark. We improve the state-of-the-art by 0.5% in the Entertainment domain, and 6.7% for the movies domain.
AB - Semantic slot filling is one of the most challenging problems in spoken language understanding (SLU). In this paper, we propose to use recurrent neural networks (RNNs) for this task, and present several novel architectures designed to efficiently model past and future temporal dependencies. Specifically, we implemented and compared several important RNN architectures, including Elman, Jordan, and hybrid variants. To facilitate reproducibility, we implemented these networks with the publicly available Theano neural network toolkit and completed experiments on the well-known airline travel information system (ATIS) benchmark. In addition, we compared the approaches on two custom SLU data sets from the entertainment and movies domains. Our results show that the RNN-based models outperform the conditional random field (CRF) baseline by 2% in absolute error reduction on the ATIS benchmark. We improve the state-of-the-art by 0.5% in the Entertainment domain, and 6.7% for the movies domain.
KW - Recurrent neural network (RNN)
KW - slot filling
KW - spoken language understanding (SLU)
KW - word embedding
UR - http://www.scopus.com/inward/record.url?scp=84923922436&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84923922436&partnerID=8YFLogxK
U2 - 10.1109/TASLP.2014.2383614
DO - 10.1109/TASLP.2014.2383614
M3 - Article
AN - SCOPUS:84923922436
SN - 1558-7916
VL - 23
SP - 530
EP - 539
JO - IEEE Transactions on Audio, Speech and Language Processing
JF - IEEE Transactions on Audio, Speech and Language Processing
IS - 3
M1 - 6998838
ER -