Update app.py
Browse files
app.py
CHANGED
@@ -220,7 +220,7 @@ def start_tryon(dict, garm_img, garment_des, is_checked, is_checked_crop, denois
|
|
220 |
negative_pooled_prompt_embeds=negative_pooled_prompt_embeds.to(device, torch.float16),
|
221 |
num_inference_steps=denoise_steps,
|
222 |
generator=generator,
|
223 |
-
strength=1.
|
224 |
pose_img=pose_img.to(device, torch.float16),
|
225 |
text_embeds_cloth=prompt_embeds_c.to(device, torch.float16),
|
226 |
cloth=garm_tensor.to(device, torch.float16),
|
@@ -229,7 +229,7 @@ def start_tryon(dict, garm_img, garment_des, is_checked, is_checked_crop, denois
|
|
229 |
height=1024,
|
230 |
width=768,
|
231 |
ip_adapter_image=garm_img.resize((768, 1024)),
|
232 |
-
guidance_scale=
|
233 |
)[0]
|
234 |
|
235 |
if is_checked_crop:
|
|
|
220 |
negative_pooled_prompt_embeds=negative_pooled_prompt_embeds.to(device, torch.float16),
|
221 |
num_inference_steps=denoise_steps,
|
222 |
generator=generator,
|
223 |
+
strength=1.5,
|
224 |
pose_img=pose_img.to(device, torch.float16),
|
225 |
text_embeds_cloth=prompt_embeds_c.to(device, torch.float16),
|
226 |
cloth=garm_tensor.to(device, torch.float16),
|
|
|
229 |
height=1024,
|
230 |
width=768,
|
231 |
ip_adapter_image=garm_img.resize((768, 1024)),
|
232 |
+
guidance_scale=1.5,
|
233 |
)[0]
|
234 |
|
235 |
if is_checked_crop:
|