Spaces:
Sleeping
Sleeping
Update Dockerfile
Browse files- Dockerfile +37 -41
Dockerfile
CHANGED
|
@@ -1,54 +1,50 @@
|
|
| 1 |
-
#
|
| 2 |
-
FROM
|
| 3 |
|
| 4 |
# Set environment variables to prevent interactive prompts during apt operations
|
| 5 |
ENV DEBIAN_FRONTEND=noninteractive
|
| 6 |
|
| 7 |
-
# Install system dependencies:
|
| 8 |
-
RUN apt update && apt install -y
|
| 9 |
|
| 10 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
RUN useradd -m -u 1000 user
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
#
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
|
| 22 |
-
|
| 23 |
-
#
|
| 24 |
-
# Start ollama serve in background, wait for it, pull the model, then kill ollama.
|
| 25 |
-
# This ensures the model is downloaded and cached within the image.
|
| 26 |
-
RUN ollama serve & \
|
| 27 |
-
echo "Waiting for Ollama to start for model pull..." && \
|
| 28 |
-
sleep 5 && \
|
| 29 |
-
while ! curl -s http://localhost:11434 > /dev/null; do \
|
| 30 |
-
sleep 1; \
|
| 31 |
-
done && \
|
| 32 |
-
echo "Ollama started. Pulling model: ${OLLAMA_HF_MODEL}" && \
|
| 33 |
-
ollama pull ${OLLAMA_HF_MODEL} && \
|
| 34 |
-
echo "Model pull complete. Stopping Ollama for build process." && \
|
| 35 |
-
pkill ollama || true # '|| true' to prevent build failure if pkill returns non-zero when ollama already stopped
|
| 36 |
-
|
| 37 |
-
# Copy the application files
|
| 38 |
-
COPY app.py .
|
| 39 |
-
COPY requirements.txt .
|
| 40 |
-
COPY run.sh .
|
| 41 |
-
|
| 42 |
-
# Install Python dependencies required by your Gradio application
|
| 43 |
RUN pip install --no-cache-dir -r requirements.txt
|
| 44 |
|
| 45 |
-
# Make the startup script executable
|
| 46 |
RUN chmod +x run.sh
|
| 47 |
|
| 48 |
-
#
|
| 49 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 50 |
EXPOSE 7860
|
| 51 |
|
| 52 |
-
# Set the
|
| 53 |
-
# This script will start Ollama and then your application.
|
| 54 |
CMD ["./run.sh"]
|
|
|
|
| 1 |
+
# Start with a Python base image, which is convenient for running the Gradio app.
|
| 2 |
+
FROM python:3.9
|
| 3 |
|
| 4 |
# Set environment variables to prevent interactive prompts during apt operations
|
| 5 |
ENV DEBIAN_FRONTEND=noninteractive
|
| 6 |
|
| 7 |
+
# Install system dependencies: curl for Ollama and wget for model download.
|
| 8 |
+
RUN apt-get update && apt-get install -y curl wget
|
| 9 |
|
| 10 |
+
# Install Ollama using its official installation script.
|
| 11 |
+
RUN curl -fsSL https://ollama.com/install.sh | sh
|
| 12 |
+
|
| 13 |
+
# Set environment variables for the model for easy modification.
|
| 14 |
+
ENV MODEL_REPO="unsloth/gemma-3-4b-it-qat-GGUF"
|
| 15 |
+
ENV MODEL_FILENAME="gemma-3-4b-it-qat.Q4_K_M.gguf"
|
| 16 |
+
|
| 17 |
+
# Create a non-root user and switch to its home directory. This is a best practice for security and avoids permission errors with model caching.
|
| 18 |
+
# As seen in the Hugging Face Docker guide. [1]
|
| 19 |
RUN useradd -m -u 1000 user
|
| 20 |
+
WORKDIR /home/user/app
|
| 21 |
+
|
| 22 |
+
# Download the GGUF model file directly from Hugging Face Hub.
|
| 23 |
+
RUN wget -O ${MODEL_FILENAME} https://huggingface.co/${MODEL_REPO}/resolve/main/${MODEL_FILENAME}
|
| 24 |
+
|
| 25 |
+
# Copy the application files and set the correct ownership to the new user. [1]
|
| 26 |
+
COPY --chown=user:user Modelfile .
|
| 27 |
+
COPY --chown=user:user app.py .
|
| 28 |
+
COPY --chown=user:user requirements.txt .
|
| 29 |
+
COPY --chown=user:user run.sh .
|
| 30 |
+
|
| 31 |
+
# Install Python dependencies.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
RUN pip install --no-cache-dir -r requirements.txt
|
| 33 |
|
| 34 |
+
# Make the startup script executable.
|
| 35 |
RUN chmod +x run.sh
|
| 36 |
|
| 37 |
+
# Switch to the non-root user. [1]
|
| 38 |
+
USER user
|
| 39 |
+
|
| 40 |
+
# Set environment variables for the user and for Ollama.
|
| 41 |
+
# This ensures cached models are stored in a writable directory and that Ollama is accessible within the container network. [1]
|
| 42 |
+
ENV HOME=/home/user
|
| 43 |
+
ENV PATH=/home/user/.local/bin:$PATH
|
| 44 |
+
ENV OLLAMA_HOST=0.0.0.0
|
| 45 |
+
|
| 46 |
+
# Expose the port for the Gradio application.
|
| 47 |
EXPOSE 7860
|
| 48 |
|
| 49 |
+
# Set the command to run our startup script.
|
|
|
|
| 50 |
CMD ["./run.sh"]
|