|
|
|
from typing import Tuple |
|
|
|
import requests |
|
import random |
|
import numpy as np |
|
import gradio as gr |
|
import spaces |
|
import os |
|
import torch |
|
from PIL import Image |
|
from diffusers import FluxInpaintPipeline |
|
from diffusers import FluxImg2ImgPipeline |
|
|
|
MAX_SEED = np.iinfo(np.int32).max |
|
IMAGE_SIZE = 1024 |
|
DEVICE = "cuda" if torch.cuda.is_available() else "cpu" |
|
|
|
|
|
def remove_background(image: Image.Image, threshold: int = 50) -> Image.Image: |
|
image = image.convert("RGBA") |
|
data = image.getdata() |
|
new_data = [] |
|
for item in data: |
|
avg = sum(item[:3]) / 3 |
|
if avg < threshold: |
|
new_data.append((0, 0, 0, 0)) |
|
else: |
|
new_data.append(item) |
|
|
|
image.putdata(new_data) |
|
return image |
|
|
|
|
|
|
|
|
|
pipe2 = FluxImg2ImgPipeline.from_pretrained( |
|
"black-forest-labs/FLUX.1-dev", torch_dtype=torch.bfloat16).to(DEVICE) |
|
|
|
def resize_image_dimensions( |
|
original_resolution_wh: Tuple[int, int], |
|
maximum_dimension: int = IMAGE_SIZE |
|
) -> Tuple[int, int]: |
|
|
|
w, h = original_resolution_wh |
|
w_original, h_original = w, h |
|
|
|
if w > h: |
|
aspect_ratiow = h / w |
|
w = IMAGE_SIZE |
|
h = IMAGE_SIZE * aspect_ratiow |
|
|
|
if h > w: |
|
aspect_ratioh = w / h |
|
w = IMAGE_SIZE * aspect_ratioh |
|
h = IMAGE_SIZE |
|
|
|
if h == w: |
|
w = IMAGE_SIZE |
|
h = IMAGE_SIZE |
|
|
|
|
|
|
|
w = w - w % 8 |
|
h = h - h % 8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
return int(w), int(h) |
|
|
|
|
|
@spaces.GPU(duration=80) |
|
def process( |
|
input_image_editor: dict, |
|
input_text: str, |
|
seed_slicer: int, |
|
randomize_seed_checkbox: bool, |
|
strength_slider: float, |
|
num_inference_steps_slider: int, |
|
num_influence: float, |
|
progress=gr.Progress(track_tqdm=True) |
|
): |
|
|
|
input_text = "A military COR2 "+input_text |
|
image = input_image_editor['background'] |
|
mask = input_image_editor['layers'][0] |
|
|
|
if not image: |
|
gr.Info("Please upload an image.") |
|
return None |
|
|
|
width, height = resize_image_dimensions(original_resolution_wh=image.size) |
|
resized_image = image.resize((width, height), Image.LANCZOS) |
|
if randomize_seed_checkbox: |
|
seed_slicer = random.randint(0, MAX_SEED) |
|
generator = torch.Generator().manual_seed(seed_slicer) |
|
|
|
|
|
|
|
|
|
pipe2.load_lora_weights("SIGMitch/KIT") |
|
result = pipe2( |
|
prompt=input_text, |
|
image=resized_image, |
|
width=width, |
|
height=height, |
|
num_images_per_prompt =1, |
|
strength=strength_slider, |
|
generator=generator, |
|
joint_attention_kwargs={"scale": num_influence}, |
|
num_inference_steps=num_inference_steps_slider, |
|
guidance_scale=3.5, |
|
) |
|
print('INFERENCE DONE') |
|
|
|
return result.images[0] |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
with gr.Blocks() as demo: |
|
with gr.Row(): |
|
with gr.Column(): |
|
input_image_editor_component = gr.ImageEditor( |
|
label='Image', |
|
type='pil', |
|
sources=["upload"], |
|
image_mode='RGB', |
|
layers=False |
|
) |
|
|
|
with gr.Row(): |
|
input_text_component = gr.Text( |
|
label="Prompt", |
|
show_label=False, |
|
max_lines=1, |
|
placeholder="Enter your prompt", |
|
container=False, |
|
) |
|
lorasteps = gr.Slider(label="Influence", minimum=0, maximum=2, step=0.1, value=1) |
|
submit_button_component = gr.Button( |
|
value='Submit', variant='primary', scale=0) |
|
|
|
with gr.Accordion("Advanced Settings", open=False): |
|
seed_slicer_component = gr.Slider( |
|
label="Seed", |
|
minimum=0, |
|
maximum=MAX_SEED, |
|
step=1, |
|
value=42, |
|
) |
|
|
|
randomize_seed_checkbox_component = gr.Checkbox( |
|
label="Randomize seed", value=True) |
|
|
|
with gr.Row(): |
|
strength_slider_component = gr.Slider( |
|
label="Strength", |
|
info="Indicates extent to transform the reference `image`. " |
|
"Must be between 0 and 1. `image` is used as a starting " |
|
"point and more noise is added the higher the `strength`.", |
|
minimum=0, |
|
maximum=1, |
|
step=0.01, |
|
value=0.8, |
|
) |
|
|
|
num_inference_steps_slider_component = gr.Slider( |
|
label="Number of inference steps", |
|
info="The number of denoising steps. More denoising steps " |
|
"usually lead to a higher quality image at the", |
|
minimum=1, |
|
maximum=50, |
|
step=1, |
|
value=28, |
|
) |
|
with gr.Column(): |
|
output_image_component = gr.Image( |
|
type='pil', image_mode='RGB', label='Generated image', format="png") |
|
|
|
|
|
|
|
|
|
|
|
submit_button_component.click( |
|
fn=process, |
|
inputs=[ |
|
input_image_editor_component, |
|
input_text_component, |
|
seed_slicer_component, |
|
randomize_seed_checkbox_component, |
|
strength_slider_component, |
|
num_inference_steps_slider_component, |
|
lorasteps |
|
], |
|
outputs=[ |
|
output_image_component |
|
|
|
] |
|
) |
|
|
|
|
|
demo.launch(share=True, debug=False, show_error=True) |