Spaces:
Running
Running
File size: 2,102 Bytes
2861940 bdd5ceb 2861940 283e8f1 2861940 f5dcb9d d1d2b47 2861940 3462d76 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 |
---
title: Music Descriptor
emoji: 🚀
colorFrom: blue
colorTo: indigo
sdk: gradio
sdk_version: 3.29.0
app_file: app.py
pinned: true
license: cc-by-nc-4.0
---
<!-- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference -->
# Demo Introduction
This is an example of using the [MERT-v1-95M](https://huggingface.co/m-a-p/MERT-v1-95M) model as backbone to conduct multiple music understanding tasks with the universal represenation.
The tasks include EMO, GS, MTGInstrument, MTGGenre, MTGTop50, MTGMood, NSynthI, NSynthP, VocalSetS, VocalSetT.
More models can be referred at the [map organization page](https://huggingface.co/m-a-p).
# Known Issues
## Audio Format Support
Theorectically, all the audio formats supported by [torchaudio.load()](https://pytorch.org/audio/stable/torchaudio.html#torchaudio.load) can be used in the demo. Theese should include but not limited to `WAV, AMB, MP3, FLAC`.
## Error Output
Due the **hardware limitation** of the machine hosting our demospecification (2 CPU and 16GB RAM), there might be `Error` output when uploading long audios.
Unfortunately, we couldn't fix this in a short time since our team are all volunteer researchers.
We recommend to test audios less than 30 seconds or using the live mode if you are trying the [Music Descriptor demo](https://huggingface.co/spaces/m-a-p/Music-Descriptor) hosted online at HuggingFace Space.
This issue is expected to solve in the future by applying more community-support GPU resources or using other audio encoding strategy.
In the current stage, if you want to directly run the demo with longer audios, you could:
* clone this space `git clone https://huggingface.co/spaces/m-a-p/Music-Descriptor` and deploy the demo on your own machine with higher performance following the [official instruction](https://huggingface.co/docs/hub/spaces). The code will automatically use GPU for inference if there is GPU that can be detected by `torch.cuda.is_available()`.
* develop your own application with the MERT models if you have the experience of machine learning. |