FLUX that Plays Music
Zhengcong Fei, Mingyuan Fan, Changqian Yu, Junshi Huang
2024-09-04

Summary
This paper talks about FluxMusic, a new method for generating music from text using advanced AI techniques.
What's the problem?
Current methods for creating music from text often rely on complex systems that require additional information, which can make the process slow and inefficient. This limits the ability to generate music that accurately reflects the input text.
What's the solution?
FluxMusic improves upon existing techniques by using a diffusion-based model to connect text and music. It processes text and music together in a way that allows the model to learn how to create music directly from text descriptions. The method includes advanced features like attention mechanisms to focus on important parts of the input, which helps produce high-quality music sequences. The authors also demonstrate that their approach outperforms traditional methods in both automatic evaluations and human preferences.
Why it matters?
This research is important because it opens up new possibilities for AI-generated music, making it easier for creators to produce soundtracks and compositions based on written descriptions. By improving the efficiency and quality of text-to-music generation, FluxMusic can enhance applications in entertainment, education, and creative industries.
Abstract
This paper explores a simple extension of diffusion-based rectified flow Transformers for text-to-music generation, termed as FluxMusic. Generally, along with design in advanced Fluxhttps://github.com/black-forest-labs/flux model, we transfers it into a latent VAE space of mel-spectrum. It involves first applying a sequence of independent attention to the double text-music stream, followed by a stacked single music stream for denoised patch prediction. We employ multiple pre-trained text encoders to sufficiently capture caption semantic information as well as inference flexibility. In between, coarse textual information, in conjunction with time step embeddings, is utilized in a modulation mechanism, while fine-grained textual details are concatenated with the music patch sequence as inputs. Through an in-depth study, we demonstrate that rectified flow training with an optimized architecture significantly outperforms established diffusion methods for the text-to-music task, as evidenced by various automatic metrics and human preference evaluations. Our experimental data, code, and model weights are made publicly available at: https://github.com/feizc/FluxMusic.