MIDI-LLM: Adapting Large Language Models for Text-to-MIDI Music Generation
2025-11-06Code Available0· sign in to hype
Shih-Lun Wu, Yoon Kim, Cheng-Zhi Anna Huang
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/slseanwu/midi-llmOfficial★ 94
Abstract
We present MIDI-LLM, an LLM for generating multitrack MIDI music from free-form text prompts. Our approach expands a text LLM's vocabulary to include MIDI tokens, and uses a two-stage training recipe to endow text-to-MIDI abilities. By preserving the original LLM's parameter structure, we can directly leverage the vLLM library for accelerated inference. Experiments show that MIDI-LLM achieves higher quality, better text control, and faster inference compared to the recent Text2midi model. Live demo at https://midi-llm-demo.vercel.app.