TY - GEN
T1 - Diverse Paraphrasing with Insertion Models for Few-Shot Intent Detection
AU - Chevasson, Raphaël
AU - Laclau, Charlotte
AU - Gravier, Christophe
N1 - Publisher Copyright:
© 2023, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2023/1/1
Y1 - 2023/1/1
N2 - In contrast to classic autoregressive generation, insertion-based models can predict in a order-free way multiple tokens at a time, which make their generation uniquely controllable: it can be constrained to strictly include an ordered list of tokens. We propose to exploit this feature in a new diverse paraphrasing framework: first, we extract important tokens or keywords in the source sentence; second, we augment them; third, we generate new samples around them by using insertion models. We show that the generated paraphrases are competitive with state of the art autoregressive paraphrasers, not only in diversity but also in quality. We further investigate their potential to create new pseudo-labelled samples for data augmentation, using a meta-learning classification framework, and find equally competitive result. In addition to proving non-autoregressive (NAR) viability for paraphrasing, we contribute our open-source framework as a starting point for further research into controllable NAR generation.
AB - In contrast to classic autoregressive generation, insertion-based models can predict in a order-free way multiple tokens at a time, which make their generation uniquely controllable: it can be constrained to strictly include an ordered list of tokens. We propose to exploit this feature in a new diverse paraphrasing framework: first, we extract important tokens or keywords in the source sentence; second, we augment them; third, we generate new samples around them by using insertion models. We show that the generated paraphrases are competitive with state of the art autoregressive paraphrasers, not only in diversity but also in quality. We further investigate their potential to create new pseudo-labelled samples for data augmentation, using a meta-learning classification framework, and find equally competitive result. In addition to proving non-autoregressive (NAR) viability for paraphrasing, we contribute our open-source framework as a starting point for further research into controllable NAR generation.
KW - Controllable text generation
KW - Deep Learning
KW - Insertion models
KW - Natural language processing
KW - Non-autoregressive
KW - Transformers
UR - https://www.scopus.com/pages/publications/85152520773
U2 - 10.1007/978-3-031-30047-9_6
DO - 10.1007/978-3-031-30047-9_6
M3 - Conference contribution
AN - SCOPUS:85152520773
SN - 9783031300462
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 65
EP - 76
BT - Advances in Intelligent Data Analysis XXI - 21st International Symposium on Intelligent Data Analysis, IDA 2023, Proceedings
A2 - Crémilleux, Bruno
A2 - Hess, Sibylle
A2 - Nijssen, Siegfried
PB - Springer Science and Business Media Deutschland GmbH
T2 - 21st International Symposium on Intelligent Data Analysis, IDA 2022
Y2 - 12 April 2023 through 14 April 2023
ER -