--- license: mit --- ## Model Overview The **Melody Guided Music Generation (MG²)** model is an innovative approach that uses melody to guide music generation, achieving impressive results despite its simplicity and minimal resource requirements. MG² aligns melody with audio waveforms and text descriptions via a multimodal alignment module and conditions its diffusion module on these learned melody representations. This enables MG² to create music that matches the style of given audio and reflects the content of text descriptions. ## Demo Explore the capabilities of the MG² model through an online demo: - **Demo Link**: [Model Demo](https://awesome-mmgen.github.io/) - **Instructions**: Input a text description, then click "Generate" to see the music generated by the model. ## GitHub Repository Access the code and additional resources for the MG² model: - **GitHub Link**: [MG² on GitHub](https://github.com/shaopengw/Awesome-Music-Generation) ## Integration with Transformers and Hugging Face Hub We are currently working on integrating MG² into the **Hugging Face Transformers** library and making it available on the **Hugging Face Hub** 🤗. ### Tips: To generate high-quality music using MG², you'd better craft detailed and descriptive prompts that provide rich context and specific musical elements. ## Paper > **Title**: "Melody Is All You Need For Music Generation"   > **Authors**: Shaopeng Wei, Manzhen Wei, Haoyu Wang, Yu Zhao, Gang Kou   > **Year**: 2024   > [arXiv Link](https://arxiv.org/abs/2409.20196)   ## Citation ```bibtex @article{wei2024melodyneedmusicgeneration,       title={Melody Is All You Need For Music Generation},       author={Shaopeng Wei and Manzhen Wei and Haoyu Wang and Yu Zhao and Gang Kou},       year={2024},       eprint={2409.20196},       archivePrefix={arXiv},       primaryClass={cs.SD},       url={https://arxiv.org/abs/2409.20196}, }