Recently,various algorithms have been developed for generating appealing music.However,the style control in the generation process has been somewhat overlooked.Music style refers to the representative and unique appea...Recently,various algorithms have been developed for generating appealing music.However,the style control in the generation process has been somewhat overlooked.Music style refers to the representative and unique appearance presented by a musical work,and it is one of the most salient qualities of music.In this paper,we propose an innovative music generation algorithm capable of creating a complete musical composition from scratch based on a specified target style.A style-conditioned linear Transformer and a style-conditioned patch discriminator are introduced in the model.The style-conditioned linear Transformer models musical instrument digital interface(MIDI)event sequences and emphasizes the role of style information.Simultaneously,the style-conditioned patch discriminator applies an adversarial learning mechanism with two innovative loss functions to enhance the modeling of music sequences.Moreover,we establish a discriminative metric for the first time,enabling the evaluation of the generated music’s consistency concerning music styles.Both objective and subjective evaluations of our experimental results indicate that our method’s performance with regard to music production is better than the performances encountered in the case of music production with the use of state-of-the-art methods in available public datasets.展开更多
基金Project supported by the Natural Science Foundation of Guangdong Province in China(No.2021A1515011888)。
文摘Recently,various algorithms have been developed for generating appealing music.However,the style control in the generation process has been somewhat overlooked.Music style refers to the representative and unique appearance presented by a musical work,and it is one of the most salient qualities of music.In this paper,we propose an innovative music generation algorithm capable of creating a complete musical composition from scratch based on a specified target style.A style-conditioned linear Transformer and a style-conditioned patch discriminator are introduced in the model.The style-conditioned linear Transformer models musical instrument digital interface(MIDI)event sequences and emphasizes the role of style information.Simultaneously,the style-conditioned patch discriminator applies an adversarial learning mechanism with two innovative loss functions to enhance the modeling of music sequences.Moreover,we establish a discriminative metric for the first time,enabling the evaluation of the generated music’s consistency concerning music styles.Both objective and subjective evaluations of our experimental results indicate that our method’s performance with regard to music production is better than the performances encountered in the case of music production with the use of state-of-the-art methods in available public datasets.