
Deploy OpenWebUI [Updated May ’26]
[May'26] Deploy and Host Open Web UI with Railway
open-webui
Just deployed
/app/backend/data
Deploy and Host Open WebUI on Railway
Open WebUI is a feature-rich, self-hosted AI platform that lets you run and interact with large language models entirely on your own infrastructure. It supports Ollama, OpenAI, Anthropic, and any OpenAI-compatible API — all from one clean, ChatGPT-style interface you fully control.
About Hosting Open WebUI
Hosting Open WebUI requires a persistent runtime environment, a volume mount for conversation and model data, and a connection to an LLM backend such as Ollama or an external API provider. The application runs as a single Docker container exposed on port 8080, with an SQLite database handling chat history and user sessions by default. On Railway, the container is deployed directly from GitHub Container Registry, environment variables handle all provider connections, and persistent storage keeps your data intact across redeploys — no manual server configuration required.
Common Use Cases
- Private AI assistant for teams — Deploy a shared ChatGPT alternative with user roles, access controls, and conversation history your org fully owns
- RAG-powered knowledge base — Upload internal documents, PDFs, and notes and query them with citations using Open WebUI's built-in retrieval-augmented generation engine
- Multi-model AI workspace — Connect Ollama, OpenAI, Anthropic, and other providers simultaneously and switch or compare models mid-conversation
- AI dev environment — Use the built-in Python tool runner and MCP/OpenAPI integrations to build and test AI-powered automations without leaving the UI
Dependencies for Open WebUI Hosting
- Docker image — ghcr.io/open-webui/open-webui:main (GitHub Container Registry)
- Persistent volume — mounted at /app/backend/data for chat history, user data, and uploaded files
- LLM backend — Ollama (self-hosted) or an external API key for OpenAI, Anthropic, or any OpenAI-compatible provider
Deployment Dependencies
- Open WebUI Documentation: https://docs.openwebui.com/
- Open WebUI GitHub Repository: https://github.com/open-webui/open-webui
- Docker image on GHCR: https://github.com/open-webui/open-webui/pkgs/container/open-webui
- Ollama on Railway: https://railway.com/template/ollama
Implementation Details
Set the following environment variables when deploying:
PORT=8080
WEBUI_SECRET_KEY=your-random-secret-key
OLLAMA_BASE_URL=http://your-ollama-service:11434
OPENAI_API_KEY=your-key
DATA_DIR=/app/backend/data
The container exposes port 8080 by default. Railway will generate a public URL automatically on deploy.
Why Deploy Open WebUI on Railway?
Railway is a singular platform to deploy your infrastructure stack. Railway will host your infrastructure so you don't have to deal with configuration, while allowing you to vertically and horizontally scale it.
By deploying Open WebUI on Railway, you are one step closer to supporting a complete full-stack application with minimal burden. Host your servers, databases, AI agents, and more on Railway.
Template Content
open-webui
ghcr.io/open-webui/open-webui:main