tommulder commited on
Commit
ab66d7a
·
1 Parent(s): 77b66b9

chore(docker): pin flash-attn prebuilt wheel to cp311 torch2.7 cu126; remove pip fallback

Browse files
Files changed (1) hide show
  1. Dockerfile +2 -3
Dockerfile CHANGED
@@ -5,7 +5,7 @@ FROM ${BASE_IMAGE}
5
  # Enable by default for Hugging Face Spaces GPU builds; override locally with
6
  # --build-arg INSTALL_FLASH_ATTN=false
7
  ARG INSTALL_FLASH_ATTN=true
8
- ARG FLASH_ATTN_WHEEL_URL=https://github.com/mjun0812/flash-attention-prebuild-wheels/releases/download/v0.0.8/flash_attn-2.7.4.post1+cu126torch2.7-cp310-cp310-linux_x86_64.whl
9
 
10
  # Persist caches and model storage in Spaces, and enable fast transfers
11
  ENV HF_HUB_ENABLE_HF_TRANSFER=1 \
@@ -73,8 +73,7 @@ RUN if [ "$INSTALL_FLASH_ATTN" = "true" ]; then \
73
  echo "Installing flash-attn from $FLASH_ATTN_WHEEL_URL" && \
74
  pip install --no-cache-dir "$FLASH_ATTN_WHEEL_URL"; \
75
  else \
76
- echo "flash-attn wheel tag mismatch for $PYTAG. Trying pip install flash-attn (no-build-isolation)." && \
77
- (pip install --no-cache-dir --no-build-isolation flash-attn || echo "flash-attn pip install failed; continuing without flash-attn"); \
78
  fi; \
79
  else \
80
  echo "Skipping flash-attn installation"; \
 
5
  # Enable by default for Hugging Face Spaces GPU builds; override locally with
6
  # --build-arg INSTALL_FLASH_ATTN=false
7
  ARG INSTALL_FLASH_ATTN=true
8
+ ARG FLASH_ATTN_WHEEL_URL=https://github.com/mjun0812/flash-attention-prebuild-wheels/releases/download/v0.0.8/flash_attn-2.7.4.post1+cu126torch2.7-cp311-cp311-linux_x86_64.whl
9
 
10
  # Persist caches and model storage in Spaces, and enable fast transfers
11
  ENV HF_HUB_ENABLE_HF_TRANSFER=1 \
 
73
  echo "Installing flash-attn from $FLASH_ATTN_WHEEL_URL" && \
74
  pip install --no-cache-dir "$FLASH_ATTN_WHEEL_URL"; \
75
  else \
76
+ echo "flash-attn wheel tag mismatch for $PYTAG. Skipping flash-attn install."; \
 
77
  fi; \
78
  else \
79
  echo "Skipping flash-attn installation"; \