31
test.py
Normal file
31
test.py
Normal file
@@ -0,0 +1,31 @@
|
||||
import datetime
|
||||
import time
|
||||
|
||||
import torch
|
||||
from PIL import Image
|
||||
from diffusers import Flux2KleinPipeline
|
||||
|
||||
device = "cuda"
|
||||
dtype = torch.bfloat16
|
||||
|
||||
pipe = Flux2KleinPipeline.from_pretrained("black-forest-labs/FLUX.2-klein-4B", torch_dtype=dtype, is_distilled=False)
|
||||
pipe.to(device) # save some VRAM by offloading the model to CPUsave some VRAM by offloading the model to CPU
|
||||
|
||||
timestamp = datetime.datetime.now().strftime("%Y%m%d_%H%M%S")
|
||||
prompt = ""
|
||||
num_inference_steps = 4
|
||||
input_image = Image.open("result1.png")
|
||||
|
||||
start_time = time.time()
|
||||
image = pipe(
|
||||
image=input_image,
|
||||
prompt=prompt,
|
||||
height=768,
|
||||
width=512,
|
||||
guidance_scale=1.0,
|
||||
num_inference_steps=num_inference_steps,
|
||||
# generator=torch.Generator(device=device).manual_seed(3)
|
||||
).images[0]
|
||||
|
||||
image.save(f"{timestamp}_{num_inference_steps}steps.png")
|
||||
print(f"infer time : {time.time() - start_time}")
|
||||
Reference in New Issue
Block a user