Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -31,6 +31,12 @@ def predict(audio_file_pth):
|
|
| 31 |
title = "Music Spectrogram Diffusion: Multi-instrument Music Synthesis with Spectrogram Diffusion"
|
| 32 |
|
| 33 |
description = """
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 34 |
In this work, the authors focus on a middle ground of neural synthesizers that can generate audio from MIDI sequences with arbitrary combinations of instruments in realtime.
|
| 35 |
This enables training on a wide range of transcription datasets with a single model, which in turn offers note-level control of composition and instrumentation across a wide range of instruments.
|
| 36 |
|
|
@@ -39,12 +45,6 @@ They use a simple two-stage process: MIDI to spectrograms with an encoder-decode
|
|
| 39 |
|
| 40 |
examples = ["examples/beethoven_mond_2.mid", "examples/beethoven_hammerklavier_2.mid"]
|
| 41 |
|
| 42 |
-
gr.HTML("""
|
| 43 |
-
<p>For faster inference without waiting in the queue, you should duplicate this space and upgrade to GPU via the settings.
|
| 44 |
-
<br/>
|
| 45 |
-
<a href="https://huggingface.co/spaces/reach-vb/music-spectrogram-diffusion?duplicate=true">
|
| 46 |
-
<img style="margin-top: 0em; margin-bottom: 0em" src="https://bit.ly/3gLdBN6" alt="Duplicate Space"></a>
|
| 47 |
-
</p>""")
|
| 48 |
|
| 49 |
article = """
|
| 50 |
<div style='margin:20px auto;'>
|
|
|
|
| 31 |
title = "Music Spectrogram Diffusion: Multi-instrument Music Synthesis with Spectrogram Diffusion"
|
| 32 |
|
| 33 |
description = """
|
| 34 |
+
<p>For faster inference without waiting in the queue, you should duplicate this space and upgrade to GPU via the settings.
|
| 35 |
+
<br/>
|
| 36 |
+
<a href="https://huggingface.co/spaces/reach-vb/music-spectrogram-diffusion?duplicate=true">
|
| 37 |
+
<img style="margin-top: 0em; margin-bottom: 0em" src="https://bit.ly/3gLdBN6" alt="Duplicate Space"></a>
|
| 38 |
+
</p>
|
| 39 |
+
|
| 40 |
In this work, the authors focus on a middle ground of neural synthesizers that can generate audio from MIDI sequences with arbitrary combinations of instruments in realtime.
|
| 41 |
This enables training on a wide range of transcription datasets with a single model, which in turn offers note-level control of composition and instrumentation across a wide range of instruments.
|
| 42 |
|
|
|
|
| 45 |
|
| 46 |
examples = ["examples/beethoven_mond_2.mid", "examples/beethoven_hammerklavier_2.mid"]
|
| 47 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 48 |
|
| 49 |
article = """
|
| 50 |
<div style='margin:20px auto;'>
|