A Fuzzy Training Framework for Controllable Sequence-to-Sequence Generation

被引:1
|
作者
Li, Jiajia [1 ]
Wang, Ping [2 ]
Li, Zuchao [2 ]
Liu, Xi [3 ]
Utiyama, Masao [4 ]
Sumita, Eiichiro [4 ]
Zhao, Hai [5 ]
Ai, Haojun [2 ]
机构
[1] Hankou Univ, Mus Sch, Wuhan 430212, Peoples R China
[2] Wuhan Univ, Wuhan 430072, Peoples R China
[3] Wuhan Conservatory Mus, Wuhan 430060, Peoples R China
[4] Natl Inst Informat & Commun Technol, Koganei, Tokyo 1848795, Japan
[5] Shanghai Jiao Tong Univ, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
Artificial intelligence; Decoding; Machine translation; Training data; Music; Natural languages; Computational modeling; Time factors; Fuzzy systems; Task analysis; Music lyrics generation; controllable generation; music understanding; constrained decoding; fuzzy training;
D O I
10.1109/ACCESS.2022.3202010
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The generation of music lyrics by artificial intelligence (AI) is frequently modeled as a language-targeted sequence-to-sequence generation task. Formally, if we convert the melody into a word sequence, we can consider the lyrics generation task to be a machine translation task. Traditional machine translation tasks involve translating between cross-lingual word sequences, whereas music lyrics generation tasks involve translating between music and natural language word sequences. The theme or key words of the generated lyrics are usually limited to meet the needs of the users when they are generated. This requirement can be thought of as a restricted translation problem. In this paper, we propose a fuzzy training framework that allows a model to simultaneously support both unrestricted and restricted translation by adopting an additional auxiliary training process without constraining the decoding process. This maintains the benefits of restricted translation but greatly reduces the extra time overhead of constrained decoding, thus improving its practicality. The experimental results show that our framework is well suited to the Chinese lyrics generation and restricted machine translation tasks, and that it can also generate language sequence under the condition of given restricted words without training multiple models, thereby achieving the goal of green AI.
引用
收藏
页码:92467 / 92480
页数:14
相关论文
共 50 条
  • [41] ReBoost: a retrieval-boosted sequence-to-sequence model for neural response generation
    Yutao Zhu
    Zhicheng Dou
    Jian-Yun Nie
    Ji-Rong Wen
    Information Retrieval Journal, 2020, 23 : 27 - 48
  • [42] Turkish Data-to-Text Generation Using Sequence-to-Sequence Neural Networks
    Demir, Seniz
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (02)
  • [43] Automatic Grammatical Error Correction for Sequence-to-sequence Text Generation: An Empirical Study
    Ge, Tao
    Zhang, Xingxing
    Wei, Furu
    Zhou, Ming
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 6059 - 6064
  • [44] Improving Sequence-to-Sequence Constituency Parsing
    Liu, Lemao
    Zhu, Muhua
    Shi, Shuming
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 4873 - 4880
  • [45] Sequence-to-Sequence Image Caption Generator
    Alahmadi, Rehab
    Park, Chung Hyuk
    Hahn, James
    ELEVENTH INTERNATIONAL CONFERENCE ON MACHINE VISION (ICMV 2018), 2019, 11041
  • [46] Semantic Matching for Sequence-to-Sequence Learning
    Zhang, Ruiyi
    Chen, Changyou
    Zhang, Xinyuan
    Bai, Ke
    Carin, Lawrence
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 212 - 222
  • [47] Sequence-to-sequence prediction of spatiotemporal systems
    Shen, Guorui
    Kurths, Juergen
    Yuan, Ye
    CHAOS, 2020, 30 (02)
  • [48] Assessing incrementality in sequence-to-sequence models
    Ulmer, Dennis
    Hupkes, Dieuwke
    Bruni, Elia
    4TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP (REPL4NLP-2019), 2019, : 209 - 217
  • [49] An Analysis of "Attention" in Sequence-to-Sequence Models
    Prabhavalkar, Rohit
    Sainath, Tara N.
    Li, Bo
    Rao, Kanishka
    Jaitly, Navdeep
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 3702 - 3706
  • [50] OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework
    Wang, Peng
    Yang, An
    Men, Rui
    Lin, Junyang
    Bai, Shuai
    Li, Zhikang
    Ma, Jianxin
    Zhou, Chang
    Zhou, Jingren
    Yang, Hongxia
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,