AI Service Templates
The AI Services category contains 23 templates covering LLM inference servers, chat interfaces, AI automation platforms, and specialized AI tools.
Featured Templates
These flagship templates are highlighted in the catalog:
| Template | Description | Min RAM |
|---|---|---|
| OpenClaw | Open-source AI assistant platform with multi-model support | 2 GB |
| LibreChat | Multi-provider AI chat interface (OpenAI, Anthropic, local models) | 1 GB |
| Ollama | Local LLM inference server — run models on your hardware | 4 GB |
| Dify | LLM application development platform with visual workflow builder | 2 GB |
Complete Template List
| Template | Description | Category | Min RAM | Security |
|---|---|---|---|---|
| OpenClaw | Open-source AI assistant platform | Chat | 2 GB | secure |
| LibreChat | Multi-provider AI chat interface | Chat | 1 GB | secure |
| Ollama | Local LLM inference server | Inference | 4 GB | advanced |
| Dify | LLM app development platform | Platform | 2 GB | secure |
| Open WebUI | Web interface for Ollama and OpenAI-compatible APIs | Chat | 512 MB | secure |
| LiteLLM | OpenAI-compatible proxy for 100+ LLM providers | Proxy | 512 MB | secure |
| LocalAI | Self-hosted OpenAI-compatible API | Inference | 4 GB | advanced |
| Text Generation WebUI | Gradio web UI for large language models | Chat | 4 GB | advanced |
| Flowise | Drag-and-drop LLM flow builder | Platform | 1 GB | secure |
| Langflow | Visual framework for building LLM applications | Platform | 1 GB | secure |
| AnythingLLM | All-in-one AI desktop app turned self-hosted | Chat | 1 GB | secure |
| Jan | Open-source ChatGPT alternative, runs offline | Chat | 2 GB | secure |
| GPT4All | Local LLM chat with privacy focus | Chat | 4 GB | advanced |
| Serge | Self-hosted chat interface using llama.cpp | Chat | 2 GB | secure |
| SillyTavern | AI chat and roleplay frontend | Chat | 512 MB | secure |
| Stable Diffusion WebUI | Image generation with Stable Diffusion models | Image | 8 GB | advanced |
| ComfyUI | Node-based Stable Diffusion workflow editor | Image | 8 GB | advanced |
| Whisper | OpenAI Whisper speech-to-text server | Audio | 2 GB | secure |
| Coqui TTS | Deep learning text-to-speech server | Audio | 2 GB | secure |
| Haystack | LLM orchestration framework for RAG pipelines | Framework | 1 GB | secure |
| Chroma | AI-native open-source vector database | Database | 1 GB | secure |
| Qdrant | High-performance vector similarity search engine | Database | 1 GB | secure |
| Weaviate | Vector database for AI applications | Database | 2 GB | secure |
GPU Support
Templates marked with advanced security profile may support GPU passthrough. For GPU-accelerated inference (Ollama, LocalAI, Stable Diffusion), ensure your server has:
- NVIDIA GPU with CUDA support
nvidia-container-toolkitinstalled- Docker configured with the NVIDIA runtime
The template variables include GPU configuration options when applicable.
Connecting AI Services
Many AI services work well together. Common combinations:
- Ollama + Open WebUI — Local inference with a polished chat interface
- Ollama + LiteLLM — Local inference exposed as an OpenAI-compatible API
- Dify + Ollama — Visual AI workflow builder using local models
- LibreChat + LiteLLM — Multi-provider chat routing through a single proxy
Services deployed on the same panel can communicate over the internal Docker network using their service names as hostnames.