zhzluke96
commited on
Commit
·
10a102f
1
Parent(s):
f367757
update
Browse files- requirements.txt +1 -1
- webui.py +8 -0
requirements.txt
CHANGED
|
@@ -27,4 +27,4 @@ python-box
|
|
| 27 |
ftfy
|
| 28 |
librosa
|
| 29 |
pyrubberband
|
| 30 |
-
https://github.com/Dao-AILab/flash-attention/releases/download/v2.5.9.post1/flash_attn-2.5.9.post1+cu118torch1.12cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
|
|
|
|
| 27 |
ftfy
|
| 28 |
librosa
|
| 29 |
pyrubberband
|
| 30 |
+
# https://github.com/Dao-AILab/flash-attention/releases/download/v2.5.9.post1/flash_attn-2.5.9.post1+cu118torch1.12cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
|
webui.py
CHANGED
|
@@ -27,6 +27,14 @@ from modules.utils.torch_opt import configure_torch_optimizations
|
|
| 27 |
from modules.webui import webui_config
|
| 28 |
from modules.webui.app import create_interface, webui_init
|
| 29 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 30 |
dcls_patch()
|
| 31 |
ignore_useless_warnings()
|
| 32 |
|
|
|
|
| 27 |
from modules.webui import webui_config
|
| 28 |
from modules.webui.app import create_interface, webui_init
|
| 29 |
|
| 30 |
+
import subprocess
|
| 31 |
+
|
| 32 |
+
subprocess.run(
|
| 33 |
+
"pip install flash-attn --no-build-isolation",
|
| 34 |
+
env={"FLASH_ATTENTION_SKIP_CUDA_BUILD": "TRUE"},
|
| 35 |
+
shell=True,
|
| 36 |
+
)
|
| 37 |
+
|
| 38 |
dcls_patch()
|
| 39 |
ignore_useless_warnings()
|
| 40 |
|