Spaces:
Sleeping
Sleeping
| # --- Core LLM backend (switch here only) --- | |
| # Files written by TTS go here (and get cleaned by our code) | |
| VOICE_AUDIO_DIR=runtime/audio | |
| # Optional cloud keys (not needed for local MVP) | |
| # OPENAI_API_KEY= | |
| # GROQ_API_KEY= | |
| API_BACKEND=sim # sim | mock | http | |
| # Switch between old rule router and LLM-driven flow (no code changes needed) | |
| ROUTER_MODE=llm | |
| SAY_VOICE=Samantha | |
| # Which backend to use | |
| BACKEND_LLM=llamacpp | |
| # Path where the GGUF model will be saved after download | |
| LLAMACPP_MODEL_PATH=models/qwen2.5-1.5b-instruct-q4_k_m.gguf | |
| # HF repo to download the model from | |
| HF_MODEL_REPO=Qwen/Qwen2.5-1.5B-Instruct-GGUF | |
| HF_MODEL_FILE=qwen2.5-1.5b-instruct-q4_k_m.gguf | |
| # llama.cpp runtime knobs | |
| N_CTX=4096 | |
| N_THREADS=4 | |
| N_GPU_LAYERS=0 | |
| # Audio + misc | |
| ASR_DEVICE=cpu | |
| TTS_ENGINE=piper | |
| PIPER_MODEL=models/piper/en_US-amy-medium.onnx | |
| PIPER_BIN=piper | |
| IS_HF_SPACE=true | |
| DEBUG=false |