ComposerX: Multi-Agent Music Generation with LLMs
Qixin Deng (University of Rochester), Qikai Yang (University of Illinois at Urbana-Champaign), Ruibin Yuan (CMU)*, Yipeng Huang (Multimodal Art Projection Research Community), Yi Wang (CMU), Xubo Liu (University of Surrey), Zeyue Tian (Hong Kong University of Science and Technology), Jiahao Pan (The Hong Kong University of Science and Technology), Ge Zhang (University of Michigan), Hanfeng Lin (Multimodal Art Projection Research Community), Yizhi Li (The University of Sheffield), Yinghao MA (Queen Mary University of London), Jie Fu (HKUST), Chenghua Lin (University of Manchester), Emmanouil Benetos (Queen Mary University of London), Wenwu Wang (University of Surrey), Guangyu Xia (NYU Shanghai), Wei Xue (The Hong Kong University of Science and Technology), Yike Guo (Hong Kong University of Science and Technology)
Keywords: Creativity -> computational creativity; Creativity -> human-ai co-creativity, Applications -> music composition, performance, and production
Music composition represents the creative side of humanity, and itself is a complex task that requires abilities to understand and generate information with long dependency and harmony constraints. Current LLMs often struggle with this task, sometimes generating poorly written music even when equipped with modern techniques like In- Context-Learning and Chain-of-Thoughts. To further explore and enhance LLMs’ potential in music composition by leveraging their reasoning ability and the large knowledge base in music history and theory, we propose ComposerX , an agent-based symbolic music generation framework. We find that applying a multi-agent approach significantly improves the music composition quality of GPT-4. The results demonstrate that ComposerX is capable of producing coherent polyphonic music compositions with captivating melodies, while adhering to user instructions.
Reviews
No reviews available