Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -5,14 +5,14 @@ from audio2midi.librosa_pitch_detector import Normal_Pitch_Det , Guitar_Pitch_De
|
|
| 5 |
from audio2midi.melodia_pitch_detector import Melodia
|
| 6 |
from audio2midi.pop2piano import Pop2Piano
|
| 7 |
from audio2midi.violin_pitch_detector import Violin_Pitch_Det
|
| 8 |
-
from audio2midi.mt3_music_transcription import MT3
|
| 9 |
from os import environ
|
| 10 |
from huggingface_hub import hf_hub_download
|
| 11 |
from shutil import unpack_archive
|
| 12 |
from pathlib import Path
|
| 13 |
from platform import system as platform_system , architecture as platform_architecture
|
| 14 |
-
import nest_asyncio
|
| 15 |
-
nest_asyncio.apply()
|
| 16 |
|
| 17 |
unpack_archive(hf_hub_download("shethjenil/Audio2Midi_Models",f"melodia_vamp_plugin_{'win' if (system := platform_system()) == 'Windows' else 'mac' if system == 'Darwin' else 'linux64' if (arch := platform_architecture()[0]) == '64bit' else 'linux32' if arch == '32bit' else None}.zip"),"vamp_melodia",format="zip")
|
| 18 |
unpack_archive(hf_hub_download("shethjenil/Audio2Midi_Models","mt3.zip"),"mt3_model",format="zip")
|
|
@@ -64,6 +64,6 @@ gr.TabbedInterface([
|
|
| 64 |
gr.Interface(Crepe(getenv("crepe_model_capacity","full")).predict,[gr.Audio(type="filepath",label="Input Audio"),gr.Checkbox(False,label="viterbi",info="Apply viterbi smoothing to the estimated pitch curve"),gr.Checkbox(True,label="center"),gr.Number(10,label="step size",info="The step size in milliseconds for running pitch estimation."),gr.Number(0.8,label="minimum confidence"),gr.Number(32,label="batch size")],gr.File(label="Midi File")),
|
| 65 |
gr.Interface(CrepeTF(getenv("crepe_model_capacity","full")).predict,[gr.Audio(type="filepath",label="Input Audio"),gr.Checkbox(False,label="viterbi",info="Apply viterbi smoothing to the estimated pitch curve"),gr.Checkbox(True,label="center"),gr.Number(10,label="step size",info="The step size in milliseconds for running pitch estimation."),gr.Number(0.8,label="minimum confidence"),gr.Number(32,label="batch size")],gr.File(label="Midi File")),
|
| 66 |
gr.Interface(Pop2Piano(device).predict,[gr.Audio(label="Input Audio",type="filepath"),gr.Number(1, minimum=1, maximum=21, label="Composer"),gr.Number(2,label="Details in Piano"),gr.Number(1,label="Efficiency of Piano"),gr.Radio([1,2,4],label="steps per beat",value=2)],gr.File(label="MIDI File")),
|
| 67 |
-
gr.Interface(MT3(str(Path("mt3_model").absolute())).predict,[gr.Audio(label="Input Audio",type="filepath"),gr.Number(0,label="seed")],gr.File(label="MIDI File")),
|
| 68 |
midi_viz_ui
|
| 69 |
-
],["Normal Pitch Detection","Guitar Based Pitch Detection","Melodia","Spotify Pitch Detection","Violin Based Pitch Detection","Crepe Pitch Detection","Crepe Pitch Detection TF","Pop2Piano","
|
|
|
|
| 5 |
from audio2midi.melodia_pitch_detector import Melodia
|
| 6 |
from audio2midi.pop2piano import Pop2Piano
|
| 7 |
from audio2midi.violin_pitch_detector import Violin_Pitch_Det
|
| 8 |
+
# from audio2midi.mt3_music_transcription import MT3
|
| 9 |
from os import environ
|
| 10 |
from huggingface_hub import hf_hub_download
|
| 11 |
from shutil import unpack_archive
|
| 12 |
from pathlib import Path
|
| 13 |
from platform import system as platform_system , architecture as platform_architecture
|
| 14 |
+
# import nest_asyncio
|
| 15 |
+
# nest_asyncio.apply()
|
| 16 |
|
| 17 |
unpack_archive(hf_hub_download("shethjenil/Audio2Midi_Models",f"melodia_vamp_plugin_{'win' if (system := platform_system()) == 'Windows' else 'mac' if system == 'Darwin' else 'linux64' if (arch := platform_architecture()[0]) == '64bit' else 'linux32' if arch == '32bit' else None}.zip"),"vamp_melodia",format="zip")
|
| 18 |
unpack_archive(hf_hub_download("shethjenil/Audio2Midi_Models","mt3.zip"),"mt3_model",format="zip")
|
|
|
|
| 64 |
gr.Interface(Crepe(getenv("crepe_model_capacity","full")).predict,[gr.Audio(type="filepath",label="Input Audio"),gr.Checkbox(False,label="viterbi",info="Apply viterbi smoothing to the estimated pitch curve"),gr.Checkbox(True,label="center"),gr.Number(10,label="step size",info="The step size in milliseconds for running pitch estimation."),gr.Number(0.8,label="minimum confidence"),gr.Number(32,label="batch size")],gr.File(label="Midi File")),
|
| 65 |
gr.Interface(CrepeTF(getenv("crepe_model_capacity","full")).predict,[gr.Audio(type="filepath",label="Input Audio"),gr.Checkbox(False,label="viterbi",info="Apply viterbi smoothing to the estimated pitch curve"),gr.Checkbox(True,label="center"),gr.Number(10,label="step size",info="The step size in milliseconds for running pitch estimation."),gr.Number(0.8,label="minimum confidence"),gr.Number(32,label="batch size")],gr.File(label="Midi File")),
|
| 66 |
gr.Interface(Pop2Piano(device).predict,[gr.Audio(label="Input Audio",type="filepath"),gr.Number(1, minimum=1, maximum=21, label="Composer"),gr.Number(2,label="Details in Piano"),gr.Number(1,label="Efficiency of Piano"),gr.Radio([1,2,4],label="steps per beat",value=2)],gr.File(label="MIDI File")),
|
| 67 |
+
# gr.Interface(MT3(str(Path("mt3_model").absolute())).predict,[gr.Audio(label="Input Audio",type="filepath"),gr.Number(0,label="seed")],gr.File(label="MIDI File")),
|
| 68 |
midi_viz_ui
|
| 69 |
+
],["Normal Pitch Detection","Guitar Based Pitch Detection","Melodia","Spotify Pitch Detection","Violin Based Pitch Detection","Crepe Pitch Detection","Crepe Pitch Detection TF","Pop2Piano","Midi Vizulizer"]).launch()
|