Update app.py
Browse files
app.py
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
#!/usr/bin/env python
|
2 |
-
#
|
3 |
import os
|
4 |
import random
|
5 |
import uuid
|
@@ -74,17 +74,18 @@ NUM_IMAGES_PER_PROMPT = 1
|
|
74 |
|
75 |
if torch.cuda.is_available():
|
76 |
pipe = DiffusionPipeline.from_pretrained(
|
77 |
-
"
|
78 |
torch_dtype=torch.float16,
|
79 |
use_safetensors=True,
|
80 |
-
add_watermarker=False
|
|
|
81 |
)
|
82 |
if ENABLE_CPU_OFFLOAD:
|
83 |
pipe.enable_model_cpu_offload()
|
84 |
else:
|
85 |
-
pipe.to(device)
|
86 |
print("Loaded on Device!")
|
87 |
-
|
88 |
if USE_TORCH_COMPILE:
|
89 |
pipe.unet = torch.compile(pipe.unet, mode="reduce-overhead", fullgraph=True)
|
90 |
print("Model Compiled!")
|
|
|
1 |
#!/usr/bin/env python
|
2 |
+
#Patch0.1x
|
3 |
import os
|
4 |
import random
|
5 |
import uuid
|
|
|
74 |
|
75 |
if torch.cuda.is_available():
|
76 |
pipe = DiffusionPipeline.from_pretrained(
|
77 |
+
"yodayo-ai/kivotos-xl-2.0",
|
78 |
torch_dtype=torch.float16,
|
79 |
use_safetensors=True,
|
80 |
+
add_watermarker=False,
|
81 |
+
variant="fp16"
|
82 |
)
|
83 |
if ENABLE_CPU_OFFLOAD:
|
84 |
pipe.enable_model_cpu_offload()
|
85 |
else:
|
86 |
+
pipe.to(device)
|
87 |
print("Loaded on Device!")
|
88 |
+
|
89 |
if USE_TORCH_COMPILE:
|
90 |
pipe.unet = torch.compile(pipe.unet, mode="reduce-overhead", fullgraph=True)
|
91 |
print("Model Compiled!")
|