Manjushri commited on
Commit
c961a5c
1 Parent(s): d1d8cb5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -3
app.py CHANGED
@@ -3,22 +3,26 @@ import torch
3
  import numpy as np
4
  import modin.pandas as pd
5
  from PIL import Image
6
- from diffusers import DiffusionPipeline
7
 
8
  device = "cuda" if torch.cuda.is_available() else "cpu"
9
  pipe = DiffusionPipeline.from_pretrained("prompthero/openjourney-v4", safety_checker=None)
 
 
10
  pipe = pipe.to(device)
11
 
12
  def genie (prompt, scale, steps, seed):
13
  generator = torch.Generator(device=device).manual_seed(seed)
14
  images = pipe(prompt, num_inference_steps=steps, guidance_scale=scale, generator=generator).images[0]
15
- return images
 
 
16
 
17
  gr.Interface(fn=genie, inputs=[gr.Textbox(label='What you want the AI to generate. 77 Token Limit.'),
18
  gr.Slider(1, maximum=15, value=10, step=.25),
19
  gr.Slider(1, maximum=50, value=25, step=1),
20
  gr.Slider(minimum=1, step=1, maximum=987654321, randomize=True)],
21
- outputs='image',
22
  title="OpenJourney V4 CPU",
23
  description="OJ V4 CPU. <b>WARNING:</b> Extremely Slow. 35s/Iteration. Expect 8-16mins an image for 15-30 iterations respectively. 50 iterations takes ~28mins.",
24
  article = "Code Monkey: <a href=\"https://huggingface.co/Manjushri\">Manjushri</a>").launch(debug=True, max_threads=True)
 
3
  import numpy as np
4
  import modin.pandas as pd
5
  from PIL import Image
6
+ from diffusers import DiffusionPipeline, StableDiffusionLatentUpscalePipeline
7
 
8
  device = "cuda" if torch.cuda.is_available() else "cpu"
9
  pipe = DiffusionPipeline.from_pretrained("prompthero/openjourney-v4", safety_checker=None)
10
+ upscaler = StableDiffusionLatentUpscalePipeline.from_pretrained("stabilityai/sd-x2-latent-upscaler", torch_dtype=torch.float16, safety_checker=None)
11
+ upscaler = upscaler.to(device)
12
  pipe = pipe.to(device)
13
 
14
  def genie (prompt, scale, steps, seed):
15
  generator = torch.Generator(device=device).manual_seed(seed)
16
  images = pipe(prompt, num_inference_steps=steps, guidance_scale=scale, generator=generator).images[0]
17
+ low_res_latents = pipe(prompt, generator=generator, output_type="latent").images
18
+ upscaled_image = upscaler(prompt=prompt, image=low_res_latents, num_inference_steps=20, guidance_scale=0, generator=generator).images[0]
19
+ return (images, upscaled_image)
20
 
21
  gr.Interface(fn=genie, inputs=[gr.Textbox(label='What you want the AI to generate. 77 Token Limit.'),
22
  gr.Slider(1, maximum=15, value=10, step=.25),
23
  gr.Slider(1, maximum=50, value=25, step=1),
24
  gr.Slider(minimum=1, step=1, maximum=987654321, randomize=True)],
25
+ outputs=('image', 'image')
26
  title="OpenJourney V4 CPU",
27
  description="OJ V4 CPU. <b>WARNING:</b> Extremely Slow. 35s/Iteration. Expect 8-16mins an image for 15-30 iterations respectively. 50 iterations takes ~28mins.",
28
  article = "Code Monkey: <a href=\"https://huggingface.co/Manjushri\">Manjushri</a>").launch(debug=True, max_threads=True)