LPX55 commited on
Commit
6110d64
·
verified ·
1 Parent(s): 88260e2

Update raw.py

Browse files
Files changed (1) hide show
  1. raw.py +5 -2
raw.py CHANGED
@@ -6,6 +6,7 @@ from diffusers import FluxControlNetModel, FluxControlNetPipeline, AutoencoderKL
6
  from diffusers import BitsAndBytesConfig as DiffusersBitsAndBytesConfig
7
  from transformers import T5EncoderModel
8
  from transformers import BitsAndBytesConfig as TransformersBitsAndBytesConfig
 
9
 
10
  import gradio as gr
11
  huggingface_token = os.getenv("HUGGINFACE_TOKEN")
@@ -36,12 +37,14 @@ pipe = FluxControlNetPipeline.from_pretrained(
36
  )
37
  adapter_id = "alimama-creative/FLUX.1-Turbo-Alpha"
38
  adapter_id2 = "XLabs-AI/flux-RealismLora"
 
39
 
40
  pipe.to("cuda")
41
  pipe.load_lora_weights(adapter_id, adapter_name="turbo")
42
  pipe.load_lora_weights(adapter_id2, adapter_name="real")
43
- pipe.set_adapters(["turbo", "real"], adapter_weights=[0.9, 0.6])
44
- pipe.fuse_lora(adapter_names=["turbo", "real"], lora_scale=1.0)
 
45
  pipe.unload_lora_weights()
46
  # pipe.enable_xformers_memory_efficient_attention()
47
  # save to the Hub
 
6
  from diffusers import BitsAndBytesConfig as DiffusersBitsAndBytesConfig
7
  from transformers import T5EncoderModel
8
  from transformers import BitsAndBytesConfig as TransformersBitsAndBytesConfig
9
+ from peft import PeftModel, PeftConfig
10
 
11
  import gradio as gr
12
  huggingface_token = os.getenv("HUGGINFACE_TOKEN")
 
37
  )
38
  adapter_id = "alimama-creative/FLUX.1-Turbo-Alpha"
39
  adapter_id2 = "XLabs-AI/flux-RealismLora"
40
+ adapter_id3 = "enhanceaiteam/Flux-uncensored-v2"
41
 
42
  pipe.to("cuda")
43
  pipe.load_lora_weights(adapter_id, adapter_name="turbo")
44
  pipe.load_lora_weights(adapter_id2, adapter_name="real")
45
+ pipe.load_lora_weights(adapter_id3, weight_name="lora.safetensors", adapter_name="enhance")
46
+ pipe.set_adapters(["turbo", "real", "enhance"], adapter_weights=[0.9, 0.66, 0.6])
47
+ pipe.fuse_lora(adapter_names=["turbo", "real", "enhance"], lora_scale=1.0)
48
  pipe.unload_lora_weights()
49
  # pipe.enable_xformers_memory_efficient_attention()
50
  # save to the Hub