TY - GEN
T1 - Tailor
T2 - 60th Annual Meeting of the Association for Computational Linguistics, ACL 2022
AU - Ross, Alexis
AU - Wu, Tongshuang
AU - Peng, Hao
AU - Peters, Matthew E.
AU - Gardner, Matt
N1 - Publisher Copyright:
© 2022 Association for Computational Linguistics.
PY - 2022
Y1 - 2022
N2 - Controlled text perturbation is useful for evaluating and improving model generalizability. However, current techniques rely on training a model for every target perturbation, which is expensive and hard to generalize. We present Tailor, a semantically-controlled text generation system. Tailor builds on a pretrained seq2seq model and produces textual outputs conditioned on control codes derived from semantic representations. We craft a set of operations to modify the control codes, which in turn steer generation towards targeted attributes. These operations can be further composed into higher-level ones, allowing for flexible perturbation strategies. We demonstrate the effectiveness of these perturbations in multiple applications. First, we use Tailor to automatically create high-quality contrast sets for four distinct natural language processing (NLP) tasks. These contrast sets contain fewer spurious artifacts and are complementary to manually annotated ones in their lexical diversity. Second, we show that Tailor perturbations can improve model generalization through data augmentation. Perturbing just ~2% of training data leads to a 5.8-point gain on an NLI challenge set measuring reliance on syntactic heuristics.
AB - Controlled text perturbation is useful for evaluating and improving model generalizability. However, current techniques rely on training a model for every target perturbation, which is expensive and hard to generalize. We present Tailor, a semantically-controlled text generation system. Tailor builds on a pretrained seq2seq model and produces textual outputs conditioned on control codes derived from semantic representations. We craft a set of operations to modify the control codes, which in turn steer generation towards targeted attributes. These operations can be further composed into higher-level ones, allowing for flexible perturbation strategies. We demonstrate the effectiveness of these perturbations in multiple applications. First, we use Tailor to automatically create high-quality contrast sets for four distinct natural language processing (NLP) tasks. These contrast sets contain fewer spurious artifacts and are complementary to manually annotated ones in their lexical diversity. Second, we show that Tailor perturbations can improve model generalization through data augmentation. Perturbing just ~2% of training data leads to a 5.8-point gain on an NLI challenge set measuring reliance on syntactic heuristics.
UR - http://www.scopus.com/inward/record.url?scp=85131572980&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85131572980&partnerID=8YFLogxK
U2 - 10.18653/v1/2022.acl-long.228
DO - 10.18653/v1/2022.acl-long.228
M3 - Conference contribution
AN - SCOPUS:85131572980
T3 - Proceedings of the Annual Meeting of the Association for Computational Linguistics
SP - 3194
EP - 3213
BT - ACL 2022 - 60th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers)
A2 - Muresan, Smaranda
A2 - Nakov, Preslav
A2 - Villavicencio, Aline
PB - Association for Computational Linguistics (ACL)
Y2 - 22 May 2022 through 27 May 2022
ER -