Spaces:
Runtime error
Runtime error
| # For more information, see the Configuration Guide: | |
| # https://www.librechat.ai/docs/configuration/librechat_yaml | |
| # Configuration version (required) | |
| version: 1.2.1 | |
| # Cache settings: Set to true to enable caching | |
| cache: true | |
| # File strategy s3/firebase | |
| # fileStrategy: "s3" | |
| # Custom interface configuration | |
| interface: | |
| # Privacy policy settings | |
| privacyPolicy: | |
| externalUrl: 'https://librechat.ai/privacy-policy' | |
| openNewTab: true | |
| # Terms of service | |
| termsOfService: | |
| externalUrl: 'https://librechat.ai/tos' | |
| openNewTab: true | |
| modalAcceptance: true | |
| modalTitle: "Capabilities and Terms of Service for Librechat" | |
| modalContent: | | |
| ## Overview and Key Features | |
| This platform is designed to facilitate user exploration of the Model Context Protocol (MCP) and the capabilities offered by LibreChat. It includes several powerful, built-in features accessible without requiring user-provided API keys or credentials. | |
| ### Core Capabilities: | |
| * **Integrated Agents:** | |
| * **Smarty:** An intelligent search companion optimized for web searches, providing robust performance comparable to specialized search AI tools. | |
| * **CodeSandbox:** A free, powerful alternative to proprietary code interpreter APIs. This local Python MCP server enables code execution, package installation, and multi-language support within a secure sandbox environment. | |
| * **Broad LLM Access:** Provides complimentary access (no API keys required) to a selection of state-of-the-art Large Language Models, including: | |
| * Qwen3 | |
| * Gemini 2.5 Pro | |
| * GPT-4o | |
| * GPT-4.1 | |
| * DeepSeek R1 | |
| * *(Note: Availability may be subject to change)* | |
| * **Document Processing:** | |
| * **RAG (Retrieval-Augmented Generation):** Enabled for enhanced information retrieval from provided documents. | |
| * **OCR (Optical Character Recognition):** Allows processing of text within uploaded image-based documents. | |
| * Supports uploading and processing of large files. | |
| * **Simplified Access:** All features are pre-configured, eliminating the need for users to manage or provide API keys or credentials. | |
| Note:By using the Website, you acknowledge that you have read these Terms of Service and agree to be bound by them. | |
| endpointsMenu: true | |
| modelSelect: true | |
| parameters: true | |
| sidePanel: true | |
| presets: true | |
| prompts: true | |
| bookmarks: true | |
| multiConvo: true | |
| agents: true | |
| # Example Registration Object Structure (optional) | |
| registration: | |
| socialLogins: ['github', 'google', 'discord', 'openid', 'facebook', 'apple'] | |
| # allowedDomains: | |
| # - "gmail.com" | |
| # Example Balance settings | |
| #balance: | |
| # enabled: true | |
| # startBalance: 1000 | |
| # autoRefillEnabled: true | |
| # refillIntervalValue: 1 | |
| # refillIntervalUnit: 'hours' | |
| # refillAmount: 5000 | |
| # speech: | |
| # tts: | |
| # openai: | |
| # url: '' | |
| # apiKey: '${TTS_API_KEY}' | |
| # model: '' | |
| # voices: [''] | |
| # | |
| # stt: | |
| # openai: | |
| # url: '' | |
| # apiKey: '${STT_API_KEY}' | |
| # model: '' | |
| # rateLimits: | |
| # fileUploads: | |
| # ipMax: 100 | |
| # ipWindowInMinutes: 60 # Rate limit window for file uploads per IP | |
| # userMax: 50 | |
| # userWindowInMinutes: 60 # Rate limit window for file uploads per user | |
| # conversationsImport: | |
| # ipMax: 100 | |
| # ipWindowInMinutes: 60 # Rate limit window for conversation imports per IP | |
| # userMax: 50 | |
| # userWindowInMinutes: 60 # Rate limit window for conversation imports per user | |
| # Example Actions Object Structure | |
| actions: | |
| allowedDomains: | |
| - "swapi.dev" | |
| - "librechat.ai" | |
| - "google.com" | |
| mcpServers: | |
| exa: | |
| type: stdio | |
| command: npx | |
| args: | |
| - -y | |
| - "/app/exa-mcp-server/build/index.js" | |
| - --tools=web_search_exa,research_paper_search,twitter_search,company_research,crawling,competitor_finder | |
| timeout: 60000000 | |
| env: | |
| EXA_API_KEY: '${EXA_API_KEY}' | |
| PATH: "/usr/local/bin:/usr/bin:/bin" | |
| NODE_PATH: "/usr/local/lib/node_modules" | |
| twilio: | |
| type: stdio | |
| command: npx | |
| args: | |
| - -y | |
| - '@twilio-alpha/mcp' | |
| - ACeea63a423d0c2d38fe0e21689f8dbcad/SK6cf2cb7b3f8b8a439c7f610712af3c11:QdhPW4CCq6GnORENPX8uE6t8t20UjNGy | |
| timeout: 60000000 | |
| env: | |
| PATH: "/usr/local/bin:/usr/bin:/bin" | |
| NODE_PATH: "/usr/local/lib/node_modules" | |
| memory: | |
| type: stdio | |
| command: npx | |
| args: | |
| - -y | |
| - "@modelcontextprotocol/server-memory" | |
| timeout: 60000000 | |
| codesandbox: | |
| type: stdio | |
| command: python | |
| args: | |
| - tests.py | |
| timeout: 60000000 | |
| env: | |
| JAVA_LIB_PATH: "/usr/lib/jvm/java-21-openjdk/lib" | |
| JAVA_SERVER_LIB_PATH: "/usr/lib/jvm/java-21-openjdk/lib/server" | |
| LD_LIBRARY_PATH : "/usr/lib/jvm/java-21-openjdk/lib/server:/usr/lib/jvm/java-21-openjdk/lib:" | |
| # Definition of custom endpoints | |
| endpoints: | |
| # assistants: | |
| # disableBuilder: false # Disable Assistants Builder Interface by setting to `true` | |
| # pollIntervalMs: 3000 # Polling interval for checking assistant updates | |
| # timeoutMs: 180000 # Timeout for assistant operations | |
| # # Should only be one or the other, either `supportedIds` or `excludedIds` | |
| # supportedIds: ["asst_supportedAssistantId1", "asst_supportedAssistantId2"] | |
| # # excludedIds: ["asst_excludedAssistantId"] | |
| # # Only show assistants that the user created or that were created externally (e.g. in Assistants playground). | |
| # # privateAssistants: false # Does not work with `supportedIds` or `excludedIds` | |
| # # (optional) Models that support retrieval, will default to latest known OpenAI models that support the feature | |
| # retrievalModels: ["gpt-4-turbo-preview"] | |
| # # (optional) Assistant Capabilities available to all users. Omit the ones you wish to exclude. Defaults to list below. | |
| # capabilities: ["code_interpreter", "retrieval", "actions", "tools", "image_vision"] | |
| agents: | |
| # (optional) Default recursion depth for agents, defaults to 25 | |
| recursionLimit: 50 | |
| # (optional) Max recursion depth for agents, defaults to 25 | |
| maxRecursionLimit: 100 | |
| # (optional) Disable the builder interface for agents | |
| disableBuilder: false | |
| # (optional) Agent Capabilities available to all users. Omit the ones you wish to exclude. Defaults to list below. | |
| capabilities: ["file_search", "tools", "ocr", "chain"] | |
| custom: | |
| - name: 'Tiny-DEV' # Unique name for the endpoint | |
| # For `apiKey` and `baseURL`, you can use environment variables that you define. | |
| # recommended environment variables: | |
| apiKey: '${TINY_API_KEY}' | |
| baseURL: '${TINY_API_BASE}' | |
| # Models configuration | |
| models: | |
| # List of default models to use. At least one value is required. | |
| default: ['Qwen3-235B','Qwen3-235B-A22B','Qwen3-30B-A3B','Qwen3-32B','gemini-2.5-pro-exp-03-25','gemini-2.5-pro-preview-05-06','gemini-2.5-flash-preview-04-17','gemini-2.0-flash-thinking-exp-01-21' ,'deepseek.r1','deepseek-reasoner','MAI-DS-R1-FP8', 'DeepSeek-R1-Zero', 'DeepSeek-R1T-Chimera' , 'QwQ-32B-ArliAI-RpR-v1' , 'DeepSeek-V3-0324','deepseek-chat','gpt4.1','gpt-4o','o4-mini','gemini-2.0-pro-exp-02-05','GLM-4-32B-0414','GLM-Z1-32B-0414','deepseek-r1-distill-llama-70b', 'qwq-32b','llama-3.3-70b-versatile','meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8','microsoft/Phi-4-multimodal-instruct','google/gemma-3-27b-it','meta-llama/Llama-4-Scout-17B-16E-Instruct'] | |
| # Fetch option: Set to true to fetch models from API. | |
| # fetch: false # Defaults to false. | |
| # Optional configurations | |
| # Title Conversation setting | |
| titleConvo: true # Set to true to enable title conversation | |
| modelDisplayLabel: 'Tiny' # Default is "AI" when not set. | |
| ocr: | |
| mistralModel: "mistral-ocr-latest" # Optional: Specify Mistral model, defaults to "mistral-ocr-latest" | |
| apiKey: '${OCR_API_KEY}' # Optional: Defaults to OCR_API_KEY env variable | |
| baseURL: "https://api.mistral.ai/v1" # Optional: Defaults to OCR_BASEURL env variable, or Mistral's API if no variable set | |
| strategy: "mistral_ocr" | |
| fileConfig: | |
| endpoints: | |
| agents: | |
| fileLimit: 5 | |
| fileSizeLimit: 100 # Maximum size for an individual file in MB | |
| totalSizeLimit: 500 # Maximum total size for all files in a single request in MB | |
| supportedMimeTypes: | |
| - "image/.*" | |
| - "application/pdf" | |
| - "video/.*" | |
| - "application/vnd.ms-excel" | |
| - "audio/mp3" | |
| - "audio/mpeg" | |
| - "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet" | |
| - "application/vnd.openxmlformats-officedocument.wordprocessingml.document" | |
| - "application/msword" | |
| # fileConfig: | |
| # endpoints: | |
| # assistants: | |
| # fileLimit: 5 | |
| # fileSizeLimit: 10 # Maximum size for an individual file in MB | |
| # totalSizeLimit: 50 # Maximum total size for all files in a single request in MB | |
| # supportedMimeTypes: | |
| # - "image/.*" | |
| # - "application/pdf" | |
| # openAI: | |
| # disabled: true # Disables file uploading to the OpenAI endpoint | |
| # default: | |
| # totalSizeLimit: 20 | |
| # YourCustomEndpointName: | |
| # fileLimit: 2 | |
| # fileSizeLimit: 5 | |
| # serverFileSizeLimit: 100 # Global server file size limit in MB | |
| # avatarSizeLimit: 2 # Limit for user avatar image size in MB | |
| # # See the Custom Configuration Guide for more information on Assistants Config: | |
| # # https://www.librechat.ai/docs/configuration/librechat_yaml/object_structure/assistants_endpoint |