Insertion Transformer: Flexible Sequence Generation via Insertion Operations

被引:0
|
作者
Stern, Mitchell [1 ,2 ]
Chan, William [1 ]
Kiros, Jamie [1 ]
Uszkoreit, Jakob [1 ]
机构
[1] Google Brain, Berlin, Germany
[2] Univ Calif Berkeley, Berkeley, CA 94720 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present the Insertion Transformer, an iterative, partially autoregressive model for sequence generation based on insertion operations. Unlike typical autoregressive models which rely on a fixed, often left-to-right ordering of the output, our approach accommodates arbitrary orderings by allowing for tokens to be inserted anywhere in the sequence during decoding. This flexibility confers a number of advantages: for instance, not only can our model be trained to follow specific orderings such as left-to-right generation or a binary tree traversal, but it can also be trained to maximize entropy over all valid insertions for robustness. In addition, our model seamlessly accommodates both fully autoregressive generation (one insertion at a time) and partially autoregressive generation (simultaneous insertions at multiple locations). We validate our approach by analyzing its performance on the WMT 2014 English-German machine translation task under various settings for training and decoding. We find that the Insertion Transformer outperforms many prior non-autoregressive approaches to translation at comparable or better levels of parallelism, and successfully recovers the performance of the original Transformer while requiring only logarithmically many iterations during decoding.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] ENCONTER: Entity Constrained Progressive Sequence Generation via Insertion-based Transformer
    Hsieh, Lee-Hsun
    Lee, Yang-Yin
    Lim, Ee-Peng
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 3590 - 3599
  • [2] Controlled Data Generation via Insertion Operations for NLU
    Kumar, Manoj
    Khan, Haidar
    Merhav, Yuval
    Hamza, Wael
    Rumshisky, Anna
    Gupta, Rahul
    2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, NAACL-HLT 2022, 2022, : 54 - 61
  • [3] Insertion sequence nomenclature
    Chandler, M
    Mahillon, J
    ASM NEWS, 2000, 66 (06): : 324 - 324
  • [4] Insertion sequence IS21: Related insertion sequence elements, transpositional mechanisms, and application to linker insertion mutagenesis
    Haas, D
    Berger, B
    Schmid, S
    Seitz, T
    Reimmann, C
    MOLECULAR BIOLOGY OF PSEUDOMONADS, 1996, : 238 - 249
  • [5] Character mediation of story generation via protagonist insertion
    Brown, Steven
    Berry, Matthew
    Dawes, Ember
    Hughes, Alessia
    Tu, Carmen
    JOURNAL OF COGNITIVE PSYCHOLOGY, 2019, 31 (03) : 326 - 342
  • [6] Sentiment-Aware Emoji Insertion Via Sequence Tagging
    Lin, Fuqiang
    Ma, Xingkong
    Min, Erxue
    Liu, Bo
    Song, Yiping
    IEEE MULTIMEDIA, 2021, 28 (02) : 40 - 48
  • [7] An Intelligent Robotic Transformer Insertion System
    Lin, Hsien-I
    Chang, Hsuan-Jui
    2016 INTERNATIONAL CONFERENCE ON SYSTEM SCIENCE AND ENGINEERING (ICSSE), 2016,
  • [8] Generation of benzosultams via a radical process with the insertion of sulfur dioxide
    Zhou, Kaida
    Xia, Hongguang
    Wu, Jie
    ORGANIC CHEMISTRY FRONTIERS, 2017, 4 (06): : 1121 - 1124
  • [9] Insertion sequence diversity in Archaea
    Filee, J.
    Siguier, P.
    Chandler, M.
    MICROBIOLOGY AND MOLECULAR BIOLOGY REVIEWS, 2007, 71 (01) : 121 - 157
  • [10] INSNET: An Efficient, Flexible, and Performant Insertion-based Text Generation Model
    Lu, Sidi
    Meng, Tao
    Peng, Nanyun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,