TY - GEN
T1 - One-Shot Exemplification Modeling via Latent Sense Representations
AU - Harvill, John
AU - Yoon, Hee Suk
AU - Yoon, Eunseop
AU - Hasegawa-Johnson, Mark
AU - Yoo, Chang D.
N1 - Publisher Copyright:
© 2023 Association for Computational Linguistics.
PY - 2023
Y1 - 2023
N2 - Exemplification modeling is a recently proposed task that aims to produce a viable sentence using a target word that takes on a specific meaning. This task can be particularly challenging for polysemous words since they can have multiple meanings. In this paper, we propose a one-shot variant of the exemplification modeling task such that labeled data is not needed during training, making it possible to train our system using a raw text corpus. Given one example at test time, our proposed approach can generate diverse and fluent examples where the target word accurately matches its intended meaning. We compare our approach to a fully-supervised baseline trained with different amounts of data and focus our evaluation on polysemous words. We use both automatic and human evaluations to demonstrate how each model performs on both seen and unseen words. Our proposed approach performs similarly to the fully-supervised baseline despite not using labeled data during training.
AB - Exemplification modeling is a recently proposed task that aims to produce a viable sentence using a target word that takes on a specific meaning. This task can be particularly challenging for polysemous words since they can have multiple meanings. In this paper, we propose a one-shot variant of the exemplification modeling task such that labeled data is not needed during training, making it possible to train our system using a raw text corpus. Given one example at test time, our proposed approach can generate diverse and fluent examples where the target word accurately matches its intended meaning. We compare our approach to a fully-supervised baseline trained with different amounts of data and focus our evaluation on polysemous words. We use both automatic and human evaluations to demonstrate how each model performs on both seen and unseen words. Our proposed approach performs similarly to the fully-supervised baseline despite not using labeled data during training.
UR - http://www.scopus.com/inward/record.url?scp=85174541839&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85174541839&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85174541839
T3 - Proceedings of the Annual Meeting of the Association for Computational Linguistics
SP - 303
EP - 314
BT - ACL 2023 - 8th Workshop on Representation Learning for NLP, RepL4NLP 2023 - Proceedings of the Workshop
A2 - Can, Burcu
A2 - Mozes, Maximilian
A2 - Cahyawijaya, Samuel
A2 - Saphra, Naomi
A2 - Kassner, Nora
A2 - Ravfogel, Shauli
A2 - Ravichander, Abhilasha
A2 - Zhao, Chen
A2 - Augenstein, Isabelle
A2 - Rogers, Anna
A2 - Cho, Kyunghyun
A2 - Grefenstette, Edward
A2 - Voita, Lena
PB - Association for Computational Linguistics (ACL)
T2 - 8th Workshop on Representation Learning for NLP, RepL4NLP 2023, co-located with ACL 2023
Y2 - 13 July 2023
ER -