Spaces:
Running
on
Zero
Running
on
Zero
Requirements updates -testing
Browse files- assets/examples/favicon.ico +0 -0
- assets/favicon.ico +0 -0
- requirements.txt +1 -1
- utils/ai_generator_diffusers_flux.py +2 -2
- utils/constants.py +1 -1
- utils/version_info.py +2 -0
assets/examples/favicon.ico
ADDED
|
|
assets/favicon.ico
CHANGED
|
|
|
|
requirements.txt
CHANGED
|
@@ -17,7 +17,7 @@ Pillow
|
|
| 17 |
numpy
|
| 18 |
requests
|
| 19 |
# git+https://github.com/huggingface/diffusers
|
| 20 |
-
diffusers
|
| 21 |
peft
|
| 22 |
opencv-python
|
| 23 |
open3d
|
|
|
|
| 17 |
numpy
|
| 18 |
requests
|
| 19 |
# git+https://github.com/huggingface/diffusers
|
| 20 |
+
diffusers[torch]
|
| 21 |
peft
|
| 22 |
opencv-python
|
| 23 |
open3d
|
utils/ai_generator_diffusers_flux.py
CHANGED
|
@@ -114,14 +114,14 @@ def generate_image_lowmem(
|
|
| 114 |
strength=0.75,
|
| 115 |
additional_parameters=None
|
| 116 |
):
|
| 117 |
-
# Retrieve the pipeline class from the mapping
|
| 118 |
-
print(f"device:{device}\nmodel_name:{model_name}\nlora_weights:{lora_weights}\n")
|
| 119 |
print(f"\n {get_torch_info()}\n")
|
|
|
|
| 120 |
pipeline_class = PIPELINE_CLASSES.get(pipeline_name)
|
| 121 |
if not pipeline_class:
|
| 122 |
raise ValueError(f"Unsupported pipeline type '{pipeline_name}'. "
|
| 123 |
f"Available options: {list(PIPELINE_CLASSES.keys())}")
|
| 124 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
|
|
|
| 125 |
|
| 126 |
# Disable gradient calculations
|
| 127 |
with torch.no_grad():
|
|
|
|
| 114 |
strength=0.75,
|
| 115 |
additional_parameters=None
|
| 116 |
):
|
|
|
|
|
|
|
| 117 |
print(f"\n {get_torch_info()}\n")
|
| 118 |
+
# Retrieve the pipeline class from the mapping
|
| 119 |
pipeline_class = PIPELINE_CLASSES.get(pipeline_name)
|
| 120 |
if not pipeline_class:
|
| 121 |
raise ValueError(f"Unsupported pipeline type '{pipeline_name}'. "
|
| 122 |
f"Available options: {list(PIPELINE_CLASSES.keys())}")
|
| 123 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 124 |
+
print(f"device:{device}\nmodel_name:{model_name}\nlora_weights:{lora_weights}\n")
|
| 125 |
|
| 126 |
# Disable gradient calculations
|
| 127 |
with torch.no_grad():
|
utils/constants.py
CHANGED
|
@@ -15,7 +15,7 @@ os.environ["PYTORCH_NVML_BASED_CUDA_CHECK"] = "1"
|
|
| 15 |
os.environ["CUDA_VISIBLE_DEVICES"] = "0,1,2,3"
|
| 16 |
|
| 17 |
# constants.py contains all the constants used in the project
|
| 18 |
-
os.environ["HF_TOKEN"] = ""
|
| 19 |
HF_API_TOKEN = os.getenv("HF_TOKEN")
|
| 20 |
default_lut_example_img = "./LUT/daisy.jpg"
|
| 21 |
|
|
|
|
| 15 |
os.environ["CUDA_VISIBLE_DEVICES"] = "0,1,2,3"
|
| 16 |
|
| 17 |
# constants.py contains all the constants used in the project
|
| 18 |
+
#os.environ["HF_TOKEN"] = ""
|
| 19 |
HF_API_TOKEN = os.getenv("HF_TOKEN")
|
| 20 |
default_lut_example_img = "./LUT/daisy.jpg"
|
| 21 |
|
utils/version_info.py
CHANGED
|
@@ -76,6 +76,8 @@ def versions_html():
|
|
| 76 |
 • 
|
| 77 |
transformers: {get_transformers_version()}
|
| 78 |
 • 
|
|
|
|
|
|
|
| 79 |
gradio: {gr.__version__}
|
| 80 |
 • 
|
| 81 |
{toggle_dark_link}
|
|
|
|
| 76 |
 • 
|
| 77 |
transformers: {get_transformers_version()}
|
| 78 |
 • 
|
| 79 |
+
xformers: {get_xformers_version()}
|
| 80 |
+
 • 
|
| 81 |
gradio: {gr.__version__}
|
| 82 |
 • 
|
| 83 |
{toggle_dark_link}
|