import gradio as gr import spaces import torch import os import subprocess import shlex from src.demo.model import DesignEdit os.makedirs('models', exist_ok=True) subprocess.run(shlex.split('wget https://huggingface.co./Adapter/DragonDiffusion/resolve/main/model/efficient_sam_vits.pt -O models/efficient_sam_vits.pt')) from src.demo.demo import * import shlex import cv2 pretrained_model_path = "stabilityai/stable-diffusion-xl-base-1.0" model = DesignEdit(pretrained_model_path=pretrained_model_path) DESCRIPTION_1 = """
🌿D e s i g n E d i t🌿
""" DESCRIPTION_2 = """

Multi-Layered Latent Decomposition and Fusion for Unified & Accurate Image Editing

""" DESCRIPTION_3 = """

Gradio demo for DesignEdit

""" with gr.Blocks(css='style.css') as demo: gr.HTML(DESCRIPTION_1) gr.HTML(DESCRIPTION_2) gr.HTML(DESCRIPTION_3) with gr.Tabs(): with gr.TabItem('1️⃣ Object Removal'): create_demo_remove(model.run_remove) with gr.TabItem('2️⃣ Zooming Out'): create_demo_zooming(model.run_zooming) with gr.TabItem('3️⃣ Camera Panning'): create_demo_panning(model.run_panning) with gr.TabItem('4️⃣ Object Moving, Resizing and Flipping'): create_demo_moving(model.run_moving) with gr.TabItem('5️⃣ 🚩 Multi-Layered Editing 🚩'): create_demo_layer(model.run_layer) with gr.TabItem('🔧 Mask Preparation: Draw or Sketch'): create_demo_mask_box(model.run_mask) demo.queue(max_size=20) demo.launch(max_threads=3, server_name="0.0.0.0")