Spaces:
Paused
Paused
| title: PaliGemma Demo | |
| emoji: 🤲 | |
| colorFrom: green | |
| colorTo: yellow | |
| sdk: gradio | |
| sdk_version: 4.22.0 | |
| app_file: app.py | |
| pinned: false | |
| license: apache-2.0 | |
| # PaliGemma Demo | |
| See [Blogpost] and [`big_vision README.md`] for details about the model. | |
| [Blogpost]: https://huggingface.co/blog/paligemma | |
| [`big_vision README.md`]: https://github.com/google-research/big_vision/blob/main/big_vision/configs/proj/paligemma/README.md | |
| ## Development | |
| Local testing (CPU, Python 3.12): | |
| ```bash | |
| pip -m venv env | |
| . env/bin/activate | |
| pip install -qr requirements-cpu.txt | |
| python app.py | |
| ``` | |
| Environment variables: | |
| - `MOCK_MODEL=yes`: For quick UI testing. | |
| - `RAM_CACHE_GB=18`: Enables caching of 3 bf16 models in memory: a single bf16 | |
| model is about 5860 MB. Use with care on spaces with little RAM. For example, | |
| on a `A10G large` space you can cache five models in RAM, so you would set | |
| `RAM_CACHE_GB=30`. | |
| - `HOST_COLOCATION=4`: If host RAM/disk is shared between 4 processes (e.g. the | |
| Huggingface `A10 large` Spaces). | |
| Loading models: | |
| - The set of models loaded is defined in `./models.py`. | |
| - You must first acknowledge usage conditions to access models. | |
| - When testing locally, you'll have to run `huggingface_cli login`. | |
| - When running in a Huggingface Space, you'll have to set a `HF_TOKEN` secret. | |