Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -42,8 +42,8 @@ dtype = torch.bfloat16
|
|
42 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
43 |
taef1 = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=dtype).to(device)
|
44 |
# Merge of (black-forest-labs/FLUX.1-dev + https://huggingface.co/black-forest-labs/FLUX.1-Krea-dev)
|
45 |
-
good_vae = AutoencoderKL.from_pretrained("prithivMLmods/Flux.1-Krea-Merged-
|
46 |
-
pipe_krea = DiffusionPipeline.from_pretrained("prithivMLmods/Flux.1-Krea-Merged-
|
47 |
|
48 |
# Define the flux_pipe_call_that_returns_an_iterable_of_images for flux.1-krea
|
49 |
@torch.inference_mode()
|
|
|
42 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
43 |
taef1 = AutoencoderTiny.from_pretrained("madebyollin/taef1", torch_dtype=dtype).to(device)
|
44 |
# Merge of (black-forest-labs/FLUX.1-dev + https://huggingface.co/black-forest-labs/FLUX.1-Krea-dev)
|
45 |
+
good_vae = AutoencoderKL.from_pretrained("prithivMLmods/Flux.1-Krea-Merged-Dev", subfolder="vae", torch_dtype=dtype).to(device)
|
46 |
+
pipe_krea = DiffusionPipeline.from_pretrained("prithivMLmods/Flux.1-Krea-Merged-Dev", torch_dtype=dtype, vae=taef1).to(device)
|
47 |
|
48 |
# Define the flux_pipe_call_that_returns_an_iterable_of_images for flux.1-krea
|
49 |
@torch.inference_mode()
|