TY - GEN
T1 - Conversational Question Answering with Language Models Generated Reformulations over Knowledge Graph
AU - Liu, Lihui
AU - Hill, Blaine
AU - Du, Boxin
AU - Wang, Fei
AU - Tong, Hanghang
N1 - LL, BH and HT are partially supported by NSF (1947135,), DARPA (HR001121C0165), and NIFA (2020-67021-32799).
PY - 2024
Y1 - 2024
N2 - Conversational question answering (ConvQA) over knowledge graphs (KGs) involves answering multi-turn natural language questions about information contained in a KG. State-of-the-art methods of ConvQA often struggle with inexplicit question-answer pairs. These inputs are easy for human beings to understand given a conversation history, but hard for a machine to interpret, which can degrade ConvQA performance. To address this issue, we propose a reinforcement learning (RL) based model, CORNNET, which utilizes question reformulations generated by large language models (LLMs) to improve ConvQA performance. CORNNET adopts a teacher-student architecture where a teacher model learns question representations using human writing reformulations, and a student model to mimic the teacher model's output via reformulations generated by LLMs. The learned question representation is then used by a RL model to locate the correct answer in a KG. Extensive experimental results show that CORNNET outperforms state-of-the-art ConvQA models.
AB - Conversational question answering (ConvQA) over knowledge graphs (KGs) involves answering multi-turn natural language questions about information contained in a KG. State-of-the-art methods of ConvQA often struggle with inexplicit question-answer pairs. These inputs are easy for human beings to understand given a conversation history, but hard for a machine to interpret, which can degrade ConvQA performance. To address this issue, we propose a reinforcement learning (RL) based model, CORNNET, which utilizes question reformulations generated by large language models (LLMs) to improve ConvQA performance. CORNNET adopts a teacher-student architecture where a teacher model learns question representations using human writing reformulations, and a student model to mimic the teacher model's output via reformulations generated by LLMs. The learned question representation is then used by a RL model to locate the correct answer in a KG. Extensive experimental results show that CORNNET outperforms state-of-the-art ConvQA models.
UR - http://www.scopus.com/inward/record.url?scp=85205321877&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85205321877&partnerID=8YFLogxK
U2 - 10.18653/v1/2024.findings-acl.48
DO - 10.18653/v1/2024.findings-acl.48
M3 - Conference contribution
AN - SCOPUS:85205321877
T3 - Proceedings of the Annual Meeting of the Association for Computational Linguistics
SP - 839
EP - 850
BT - The 62nd Annual Meeting of the Association for Computational Linguistics
A2 - Ku, Lun-Wei
A2 - Martins, Andre
A2 - Srikumar, Vivek
PB - Association for Computational Linguistics (ACL)
T2 - Findings of the 62nd Annual Meeting of the Association for Computational Linguistics, ACL 2024
Y2 - 11 August 2024 through 16 August 2024
ER -