Spaces:
Sleeping
Sleeping
| # Start the Ollama server in the background. | |
| # The 'serve' command starts the Ollama API server. | |
| ollama serve & | |
| # Wait for the Ollama server to become available. | |
| # We loop and check the Ollama API endpoint until it responds. | |
| echo "Waiting for Ollama to start..." | |
| while ! curl -s http://localhost:11434 > /dev/null; do | |
| sleep 1 | |
| done | |
| echo "Ollama started." | |
| # Create and register the custom model with Ollama. | |
| # 'gemma-unsloth' is the name you will use to refer to this model in your application. | |
| # It uses the Modelfile we created. | |
| ollama create gemma-unsloth -f Modelfile | |
| # Run the Gradio application. | |
| # The Gradio app will then communicate with the Ollama server running locally. | |
| python app.py |