Can you tell me the possible reason i am getting for this type of output:
doing for 20 inference step.
prompt: A cat holding a sign that says hell world, post-apocalyptic, dark-night
takes around 9 seconds
doing text to image only
with torch.no_grad():
result = pipeline(
prompt=payload.prompt,
generation_mode="image",
num_inference_steps=steps,
height=payload.height,
width=payload.width,
generator=generator,
)
dependencies:
dependencies = [
"diffusers==0.35.1",
"transformers<5",
"accelerate",
"timm",
"einops",
"sentencepiece",
"protobuf",
"pillow",
"https://github.com/Dao-AILab/flash-attention/releases/download/v2.8.3/flash_attn-2.8.3+cu12torch2.9cxx11abiTRUE-cp312-cp312-linux_x86_64.whl"
]