Spaces:
Running
Running
| title: Anycoder | |
| emoji: π’ | |
| colorFrom: indigo | |
| colorTo: indigo | |
| sdk: gradio | |
| sdk_version: 5.23.3 | |
| app_file: app.py | |
| pinned: false | |
| disable_embedding: true | |
| Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference | |
| # Anycoder - AI Code Generation with Hugging Face Inference | |
| An ultra-clean AI-powered code generation application using Hugging Face inference providers. Minimal files for maximum simplicity. | |
| ## Features | |
| - **Hugging Face Models**: Uses DeepSeek-V3-0324 via Novita provider | |
| - **Modern UI**: Built with Gradio and ModelScope Studio components | |
| - **Code Generation**: Generates working code based on user requirements | |
| - **Live Preview**: Renders generated HTML code in real-time | |
| - **History Management**: Keeps track of conversation history | |
| - **Streaming**: Real-time code generation with streaming responses | |
| ## Project Structure | |
| ``` | |
| anycoder/ | |
| βββ app.py # Main application (everything included) | |
| βββ app.css # Basic styling | |
| βββ pyproject.toml # Dependencies | |
| βββ README.md # This file | |
| ``` | |
| ## Setup | |
| 1. Set your Hugging Face API token: | |
| ```bash | |
| export HF_TOKEN="your_huggingface_token_here" | |
| ``` | |
| 2. Install dependencies: | |
| ```bash | |
| uv sync | |
| ``` | |
| 3. Run the application: | |
| ```bash | |
| uv run python app.py | |
| ``` | |
| ## Usage | |
| 1. Enter your application requirements in the text area | |
| 2. Click "send" to generate code | |
| 3. View the generated code in the code drawer | |
| 4. See the live preview in the sandbox area | |
| 5. Use example cards for quick prompts | |
| ## Code Example | |
| ```python | |
| import os | |
| from huggingface_hub import InferenceClient | |
| client = InferenceClient( | |
| provider="novita", | |
| api_key=os.environ["HF_TOKEN"], | |
| bill_to="huggingface" | |
| ) | |
| completion = client.chat.completions.create( | |
| model="deepseek-ai/DeepSeek-V3-0324", | |
| messages=[ | |
| { | |
| "role": "user", | |
| "content": "Create a simple todo app" | |
| } | |
| ], | |
| ) | |
| ``` | |
| ## Architecture | |
| The application uses: | |
| - **Gradio**: For the web interface | |
| - **Hugging Face Hub**: For model inference | |
| - **ModelScope Studio**: For UI components | |
| - **Streaming**: For real-time code generation |