Deploy LLM Council

Multi-model AI deliberation for higher-quality decisions

Deploy LLM Council

Just deployed

Deploy and Host LLM Council on Railway

Stop relying on a single AI model. LLM Council orchestrates multiple large language models through a 3-stage deliberation process—parallel responses, anonymised peer review, and chairman synthesis—to produce more reliable, balanced answers than any single model alone.

About Hosting LLM Council

LLM Council runs as a stateless HTTP server that coordinates queries across AI providers via OpenRouter. Deploy on Railway to get a dedicated API endpoint for integrating AI deliberation into your applications, automation workflows, or development tools.

The server handles:

  • Model coordination across 4+ LLMs in parallel
  • Response anonymization to eliminate model favoritism
  • Consensus building with configurable verdict types
  • Webhook notifications for async workflow integration

Build time is approximately 2-3 minutes. No database required—fully stateless.

Common Use Cases

  • Workflow Automation - Trigger council deliberations from webhooks, process results downstream. Perfect for decision gates in automated pipelines.
  • Automated Code Review - Submit PRs for multi-model review. Get consensus recommendations with binary verdicts (approve/reject) or detailed synthesis feedback.
  • Support Ticket Triage - Route tickets through the council for priority and category consensus. Multiple models cross-check to reduce misclassification.
  • Design Decision Validation - Present architectural trade-offs to the council. Peer review ensures no single model's biases dominate.
  • Content Quality Gates - Evaluate copy, documentation, or user content with multi-perspective analysis and aggregate scoring.

Dependencies for LLM Council Hosting

Note: See also Gateway Options for alternative LLM providers and gateways.

Deployment Dependencies

VariableRequiredDescription
OPENROUTER_API_KEYYesYour OpenRouter API key
LLM_COUNCIL_API_TOKENRecommendedBearer token for API auth

Implementation Details

Stack: Python 3.11 with FastAPI, running as non-root user

Health Check:

  • (GET) /health returns {"status": "ok"}

API Endpoints:

  • (POST) /v1/council/run — Synchronous deliberation
  • (GET) /v1/council/stream — Server-Sent Events streaming

Verdict Types:

  • synthesis — Natural language consensus (default)
  • binary — Go/no-go with confidence score
  • tie_breaker — Chairman resolves deadlocks

Security: All endpoints (except /health) require Bearer token authentication when LLM_COUNCIL_API_TOKEN is set.

Getting Started

  1. Click Deploy and connect your GitHub account
  2. Add your OPENROUTER_API_KEY from OpenRouter
  3. Generate an API token: openssl rand -hex 16
  4. Set LLM_COUNCIL_API_TOKEN with your generated token
  5. Deploy and test: curl https://your-app.railway.app/health

Resources

Why Deploy LLM Council on Railway?

Railway is a singular platform to deploy your infrastructure stack. Railway will host your infrastructure so you don't have to deal with configuration, while allowing you to vertically and horizontally scale it.

By deploying LLM Council on Railway, you are one step closer to supporting a complete full-stack application with minimal burden. Host your servers, databases, AI agents, and more on Railway.


Template Content

More templates in this category

View Template
Chat Chat
Chat Chat, your own unified chat and search to AI platform.

View Template
openui
Deploy OpenUI: AI-powered UI generation with GitHub OAuth and OpenAI API.

View Template
firecrawl
firecrawl api server + worker without auth, works with dify