|
import gradio as gr |
|
import spaces |
|
import json |
|
import re |
|
import random |
|
import numpy as np |
|
from gradio_client import Client |
|
|
|
MAX_SEED = np.iinfo(np.int32).max |
|
|
|
def check_api(model_name): |
|
if model_name == "MAGNet": |
|
try : |
|
client = Client("https://fffiloni-magnet.hf.space/") |
|
return "api ready" |
|
except : |
|
return "api not ready yet" |
|
elif model_name == "AudioLDM-2": |
|
try : |
|
client = Client("https://haoheliu-audioldm2-text2audio-text2music.hf.space/") |
|
return "api ready" |
|
except : |
|
return "api not ready yet" |
|
elif model_name == "Riffusion": |
|
try : |
|
client = Client("https://fffiloni-spectrogram-to-music.hf.space/") |
|
return "api ready" |
|
except : |
|
return "api not ready yet" |
|
elif model_name == "Mustango": |
|
try : |
|
client = Client("https://declare-lab-mustango.hf.space/") |
|
return "api ready" |
|
except : |
|
return "api not ready yet" |
|
elif model_name == "MusicGen": |
|
try : |
|
client = Client("https://facebook-musicgen.hf.space/") |
|
return "api ready" |
|
except : |
|
return "api not ready yet" |
|
|
|
from moviepy.editor import * |
|
import cv2 |
|
|
|
def extract_firstframe(video_in): |
|
vidcap = cv2.VideoCapture(video_in) |
|
success,image = vidcap.read() |
|
count = 0 |
|
while success: |
|
if count == 0: |
|
cv2.imwrite("first_frame.jpg", image) |
|
else: |
|
break |
|
success,image = vidcap.read() |
|
print ('Read a new frame: ', success) |
|
count += 1 |
|
print ("Done extracted first frame!") |
|
return "first_frame.jpg" |
|
|
|
def extract_audio(video_in): |
|
input_video = video_in |
|
output_audio = 'audio.wav' |
|
|
|
|
|
video_clip = VideoFileClip(input_video) |
|
audio_clip = video_clip.audio |
|
|
|
|
|
audio_clip.write_audiofile(output_audio, fps=44100) |
|
print("Audio extraction complete.") |
|
|
|
return 'audio.wav' |
|
|
|
|
|
|
|
def get_caption(image_in): |
|
kosmos2_client = Client("https://ydshieh-kosmos-2.hf.space/") |
|
kosmos2_result = kosmos2_client.predict( |
|
image_in, |
|
"Detailed", |
|
fn_index=4 |
|
) |
|
|
|
print(f"KOSMOS2 RETURNS: {kosmos2_result}") |
|
|
|
with open(kosmos2_result[1], 'r') as f: |
|
data = json.load(f) |
|
|
|
reconstructed_sentence = [] |
|
for sublist in data: |
|
reconstructed_sentence.append(sublist[0]) |
|
|
|
full_sentence = ' '.join(reconstructed_sentence) |
|
|
|
|
|
|
|
pattern = r'^Describe this image in detail:\s*(.*)$' |
|
|
|
match = re.search(pattern, full_sentence) |
|
if match: |
|
description = match.group(1) |
|
print(description) |
|
else: |
|
print("Unable to locate valid description.") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
return description |
|
|
|
def get_caption_from_MD(image_in): |
|
client = Client("https://vikhyatk-moondream1.hf.space/") |
|
result = client.predict( |
|
image_in, |
|
"Describe precisely the image.", |
|
api_name="/answer_question" |
|
) |
|
print(result) |
|
return result |
|
|
|
def get_magnet(prompt): |
|
|
|
client = Client("https://fffiloni-magnet.hf.space/") |
|
result = client.predict( |
|
"facebook/magnet-small-10secs", |
|
"", |
|
prompt, |
|
3, |
|
0.9, |
|
10, |
|
1, |
|
20, |
|
10, |
|
10, |
|
10, |
|
"prod-stride1 (new!)", |
|
api_name="/predict_full" |
|
) |
|
print(result) |
|
return result[1] |
|
|
|
def get_audioldm(prompt): |
|
client = Client("https://haoheliu-audioldm2-text2audio-text2music.hf.space/") |
|
seed = random.randint(0, MAX_SEED) |
|
result = client.predict( |
|
prompt, |
|
"Low quality.", |
|
10, |
|
6.5, |
|
seed, |
|
3, |
|
fn_index=1 |
|
) |
|
print(result) |
|
audio_result = extract_audio(result) |
|
return audio_result |
|
|
|
def get_riffusion(prompt): |
|
client = Client("https://fffiloni-spectrogram-to-music.hf.space/") |
|
result = client.predict( |
|
prompt, |
|
"", |
|
None, |
|
10, |
|
api_name="/predict" |
|
) |
|
print(result) |
|
return result[1] |
|
|
|
def get_mustango(prompt): |
|
client = Client("https://declare-lab-mustango.hf.space/") |
|
result = client.predict( |
|
prompt, |
|
200, |
|
6, |
|
api_name="/predict" |
|
) |
|
print(result) |
|
return result |
|
|
|
def get_musicgen(prompt): |
|
client = Client("https://facebook-musicgen.hf.space/") |
|
result = client.predict( |
|
prompt, |
|
None, |
|
fn_index=0 |
|
) |
|
print(result) |
|
return result[1] |
|
|
|
import re |
|
import torch |
|
from transformers import pipeline |
|
|
|
zephyr_model = "HuggingFaceH4/zephyr-7b-beta" |
|
mixtral_model = "mistralai/Mixtral-8x7B-Instruct-v0.1" |
|
|
|
pipe = pipeline("text-generation", model=zephyr_model, torch_dtype=torch.bfloat16, device_map="auto") |
|
|
|
standard_sys = f""" |
|
You are a musician AI whose job is to help users create their own music which its genre will reflect the character or scene from an image described by users. |
|
In particular, you need to respond succintly with few musical words, in a friendly tone, write a musical prompt for a music generation model. |
|
|
|
For example, if a user says, "a picture of a man in a black suit and tie riding a black dragon", provide immediately a musical prompt corresponding to the image description. |
|
Immediately STOP after that. It should be EXACTLY in this format: |
|
"A grand orchestral arrangement with thunderous percussion, epic brass fanfares, and soaring strings, creating a cinematic atmosphere fit for a heroic battle" |
|
""" |
|
|
|
mustango_sys = f""" |
|
You are a musician AI whose job is to help users create their own music which its genre will reflect the character or scene from an image described by users. |
|
In particular, you need to respond succintly with few musical words, in a friendly tone, write a musical prompt for a music generation model, you MUST include chords progression. |
|
|
|
For example, if a user says, "a painting of three old women having tea party", provide immediately a musical prompt corresponding to the image description. |
|
Immediately STOP after that. It should be EXACTLY in this format: |
|
"The song is an instrumental. The song is in medium tempo with a classical guitar playing a lilting melody in accompaniment style. The song is emotional and romantic. The song is a romantic instrumental song. The chord sequence is Gm, F6, Ebm. The time signature is 4/4. This song is in Adagio. The key of this song is G minor." |
|
""" |
|
|
|
@spaces.GPU(enable_queue=True) |
|
def get_musical_prompt(user_prompt, chosen_model): |
|
|
|
""" |
|
if chosen_model == "Mustango" : |
|
agent_maker_sys = standard_sys |
|
else : |
|
agent_maker_sys = standard_sys |
|
""" |
|
agent_maker_sys = standard_sys |
|
|
|
instruction = f""" |
|
<|system|> |
|
{agent_maker_sys}</s> |
|
<|user|> |
|
""" |
|
|
|
prompt = f"{instruction.strip()}\n{user_prompt}</s>" |
|
outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) |
|
pattern = r'\<\|system\|\>(.*?)\<\|assistant\|\>' |
|
cleaned_text = re.sub(pattern, '', outputs[0]["generated_text"], flags=re.DOTALL) |
|
|
|
print(f"SUGGESTED Musical prompt: {cleaned_text}") |
|
return cleaned_text.lstrip("\n") |
|
|
|
|
|
def blend_vmsc(video_in, audio_result): |
|
audioClip = AudioFileClip(audio_result) |
|
print(f"AUD: {audioClip.duration}") |
|
clip = VideoFileClip(video_in) |
|
print(f"VID: {clip.duration}") |
|
if clip.duration < audioClip.duration : |
|
audioClip = audioClip.subclip((0.0), (clip.duration)) |
|
elif clip.duration > audioClip.duration : |
|
clip = clip.subclip((0.0), (audioClip.duration)) |
|
final_clip = clip.set_audio(audioClip) |
|
|
|
codec = 'libx264' |
|
audio_codec = 'aac' |
|
final_clip.write_videofile('final_video_with_music.mp4', codec=codec, audio_codec=audio_codec) |
|
return "final_video_with_music.mp4" |
|
|
|
def infer(video_in, chosen_model, api_status): |
|
if video_in == None : |
|
raise gr.Error("Please provide a video input") |
|
|
|
if chosen_model == [] : |
|
raise gr.Error("Please pick a model") |
|
|
|
if api_status == "api not ready yet" : |
|
raise gr.Error("This model is not ready yet, you can pick another one instead :)") |
|
|
|
image_in = extract_firstframe(video_in) |
|
|
|
gr.Info("Getting image caption with Kosmos2...") |
|
user_prompt = get_caption(image_in) |
|
|
|
gr.Info("Building a musical prompt according to the image caption ...") |
|
musical_prompt = get_musical_prompt(user_prompt, chosen_model) |
|
|
|
if chosen_model == "MAGNet" : |
|
gr.Info("Now calling MAGNet for music...") |
|
music_o = get_magnet(musical_prompt) |
|
elif chosen_model == "AudioLDM-2" : |
|
gr.Info("Now calling AudioLDM-2 for music...") |
|
music_o = get_audioldm(musical_prompt) |
|
elif chosen_model == "Riffusion" : |
|
gr.Info("Now calling Riffusion for music...") |
|
music_o = get_riffusion(musical_prompt) |
|
elif chosen_model == "Mustango" : |
|
gr.Info("Now calling Mustango for music...") |
|
music_o = get_mustango(musical_prompt) |
|
elif chosen_model == "MusicGen" : |
|
gr.Info("Now calling MusicGen for music...") |
|
music_o = get_musicgen(musical_prompt) |
|
|
|
final_res = blend_vmsc(video_in, music_o) |
|
return gr.update(value=musical_prompt, interactive=True), gr.update(visible=True), music_o, final_res |
|
|
|
def retry(video_in, chosen_model, caption): |
|
musical_prompt = caption |
|
|
|
if chosen_model == "MAGNet" : |
|
gr.Info("Now calling MAGNet for music...") |
|
music_o = get_magnet(musical_prompt) |
|
elif chosen_model == "AudioLDM-2" : |
|
gr.Info("Now calling AudioLDM-2 for music...") |
|
music_o = get_audioldm(musical_prompt) |
|
elif chosen_model == "Riffusion" : |
|
gr.Info("Now calling Riffusion for music...") |
|
music_o = get_riffusion(musical_prompt) |
|
elif chosen_model == "Mustango" : |
|
gr.Info("Now calling Mustango for music...") |
|
music_o = get_mustango(musical_prompt) |
|
elif chosen_model == "MusicGen" : |
|
gr.Info("Now calling MusicGen for music...") |
|
music_o = get_musicgen(musical_prompt) |
|
final_res = blend_vmsc(video_in, music_o) |
|
return music_o, final_res |
|
|
|
|
|
css = """ |
|
#col-container { |
|
margin: 0 auto; |
|
max-width: 980px; |
|
text-align: left; |
|
} |
|
footer { |
|
visibility: hidden; |
|
} |
|
#inspi-prompt textarea { |
|
font-size: 20px; |
|
line-height: 24px; |
|
font-weight: 600; |
|
} |
|
/* fix examples gallery width on mobile */ |
|
div#component-11 > .gallery > .gallery-item > .container > img { |
|
width: auto!important; |
|
} |
|
""" |
|
|
|
with gr.Blocks(css=css) as demo: |
|
|
|
with gr.Column(elem_id="col-container"): |
|
with gr.Row(): |
|
|
|
with gr.Column(): |
|
video_in = gr.Video(sources=["upload"], label="Video Input") |
|
|
|
with gr.Row(): |
|
|
|
chosen_model = gr.Dropdown( |
|
label="Choose a Model", |
|
choices=[ |
|
"MAGNet", |
|
"AudioLDM-2", |
|
"Riffusion", |
|
"Mustango", |
|
"MusicGen" |
|
], |
|
value=None, |
|
filterable=False |
|
) |
|
|
|
check_status = gr.Textbox( |
|
label="API Status", |
|
interactive=False |
|
) |
|
|
|
submit_btn = gr.Button("Generate Music") |
|
|
|
with gr.Column(): |
|
|
|
caption = gr.Textbox( |
|
label="Inspirational Musical Prompt", |
|
interactive=False, |
|
elem_id="inspi-prompt" |
|
) |
|
|
|
retry_btn = gr.Button("Retry with Edited Prompt", visible=False) |
|
|
|
result = gr.Audio(label="Music") |
|
video_o = gr.Video(label="Video with SoundFX") |
|
|
|
chosen_model.change( |
|
fn=check_api, |
|
inputs=chosen_model, |
|
outputs=check_status, |
|
queue=False |
|
) |
|
|
|
retry_btn.click( |
|
fn=retry, |
|
inputs=[video_in, chosen_model, caption], |
|
outputs=[result, video_o] |
|
) |
|
|
|
submit_btn.click( |
|
fn=infer, |
|
inputs=[video_in, chosen_model, check_status], |
|
outputs=[caption, retry_btn, result, video_o], |
|
concurrency_limit=4 |
|
) |
|
|
|
demo.queue(max_size=16).launch(show_api=False) |