Own Your AIInfrastructure
A self-hosted, open-source platform for deploying, managing, and scaling conversational AI across your organization. Multi-model. Enterprise-grade. Yours to control.
Why SecondStack
Built for Organizations that take AI seriously
Keep sensitive data (chat threads, attached documents, usage logs) inside your organization's perimeter.
But we still use external LLM API providers...?
Even when using external LLM providers, API access carries a smaller risk and exposure profile than full SaaS platforms. API requests are retained for a shorter time and might be better protected then full ChatGPT-like Web App platforms hosting your entire data indefinitely, and having much larger attack surface. Many reputable LLM providers also offer Zero Data Retention (ZDR) guarantees.
A specialized guardrails module compatible with LiteLLM, and configurable via the ControlTower admin dashboard.
LLM classifier for Secrets (example)
Classifies whether a detected candidate string is a real credential (password, API key, token, private key) or a harmless match (code identifier, hash, UUID, public key, color code, placeholder). Examples of NOT a credential: documentation placeholders, commit hashes, package hashes, version numbers, color codes, PNG data, SVG strokes, file names, dates. Examples of a real credential: randomly-generated passwords, API keys, tokens, secrets, private keys that don't look like demo values. The classifier receives both the flagged string and surrounding context to make an informed decision.
LLM classifier for Customer PII (example)
Classifies whether a detected candidate string is a sensitive customer PII that should not be sent to external LLM API. Examples of NOT PII for this purpose: placeholder values ("John Doe", "test@test.com"), company names, product names, employee business contacts e.g. from email headings or signatures, business email addresses, internal Customer IDs, fictional characters, general mentions of public figures (e.g. "President George W Bush"). Depending on a company policy, even an isolated one-off occurence of low-sensitivity PII elements e.g. a customer's email address may be allowed e.g. in troubleshooting context if not accompanied by other highly sensitive revelaing data. Examples of real PII: data extracted from database records or user profiles, customer PII appearing at scale in structured data, high-sensitivity identifiers like SSNs or credit card numbers. The classifier uses surrounding context to distinguish a customer database dump from routine business correspondence.
Pay for actual compute consumption to your LLM API providers, not per-user subscriptions.
In larger organizations, most users are low-volume; API consumption-based pricing is typically far more cost-effective than flat per-seat fees.
Centralize all LLM traffic from chat users, internal apps, agents, and IDE plugins through LiteLLM.
Users create self-serve virtual API keys for authentication, with budget enforcement, logging, usage analytics, and guardrails. Your apps call standard OpenAI-compatible endpoints, and the gateway handles routing, failover, load balancing, observability, and policy across all supported models
An open platform means simplified integration with your existing toolchain, no vendor lock-in, and full ability to customize.
Inspect every line, contribute back, or fork. Your deployment is never held hostage by a vendor's roadmap.
Platform
Architecture Overview
A modular, self-hosted stack built on proven open-source foundations.
Features
Everything your team needs
A complete conversational AI platform, not just a chat interface.
Multi-model chat
Claude, GPT, Gemini, local models — chat, agents, and image generation all in one place.
Cowork Agent
Cowork Agent runs in sandboxed containers with dozens of Skills — file operations, browser automation, image generation, diagrams authoring, office documents, persistent memory, etc.
Budget & spend controls
Per-user, per-team, and per-API-key budgets. Threshold alerts, spend forecasting, and usage analytics down to the request level.
LiteLLM-based API Routing
Centralize LLM API traffic from enterprise apps and local agents with user virtual API keys, guardrails, spend observability and budgets enforcement.
Admin dashboard
Centralized model management, provider configuration, system prompts, and deployment controls. One panel for your entire AI stack.
Models
| Model | Display Name | Access | Status | Health | Order | Actions |
|---|---|---|---|---|---|---|
| cowork-agent(agent) | Cowork Agent | 2 Teams | active | 2/2 OK | ||
| llm-council(agent) | LLM Council | 2 Teams | active | 1/1 OK | ||
| claude-opus-4-6 | Claude Opus 4.6 | All Teams | active | 2/2 OK | ||
| gpt-5-2 | GPT-5.2 | All Teams | active | 1/1 OK | ||
| gemini-3-1-pro | Gemini 3.1 Pro | All Teams | active | 1/1 OK | ||
| gemini-3-flash | Gemini 3 Flash | All Teams | active | 1/1 OK | ||
| gpt-image-1-5(image_generation) | GPT Image 1.5 | All Teams | active | 1/1 OK | ||
| nano-banana-2(image_generation) | Nano Banana 2 | Hidden | disabled | — |
Users
| User | Role | Teams | Last Active | Actions | |
|---|---|---|---|---|---|
| JC Jane Chen | jane.chen@acme.com | Admin | Platform | 2 min ago | |
| MR Mike Rodriguez | mike.r@acme.com | User | Engineering | 14 min ago | |
| SP Sarah Park | sarah.p@acme.com | User | Design | 1 hr ago | |
| AK Alex Kim | alex.kim@acme.com | Team Manager | Engineering | 3 hrs ago | |
| LW Lisa Wang | lisa.w@acme.com | User | Marketing | 1 day ago |
Providers
Teams
Groups
| Group | Source | Members | Teams | Actions |
|---|---|---|---|---|
| engineering-all | Azure AD | 24 | Engineering, Platform | |
| design-team | Azure AD | 8 | Design | |
| marketing-all | Azure AD | 12 | Marketing |
Analytics & Logs
| Datetime | Model | User | Cached | Input Tok | Output Tok | Duration | Status |
|---|---|---|---|---|---|---|---|
| 2026-03-01 14:32:01 | claude-sonnet-4-6 | jane.chen | 12,480 | 1,204 | 847 | 1.2s | 200 |
| 2026-03-01 14:31:58 | gpt-4.1 | mike.r | 0 | 3,820 | 1,204 | 0.8s | 200 |
| 2026-03-01 14:31:45 | claude-opus-4-6 | alex.kim | 8,192 | 2,560 | 5,102 | 3.4s | 200 |
| 2026-03-01 14:31:32 | gemini-2.5-flash | sarah.p | 4,096 | 512 | 892 | 0.4s | 200 |
General Configuration
Notifications
MCP Servers
| Server | Type | Tools | Status | Actions |
|---|---|---|---|---|
| Context7 | stdio | 2 tools | active | |
| Web Search and Fetch | sse | 2 tools | active |
Guardrails
| Rule | Scope | Action | Status | |
|---|---|---|---|---|
| PII Detection | All Models | Block & Alert | active | … |
| Prompt Injection | Cowork Agent | Block | active | … |
| Secrets Detection | All Models | Block & Alert | active | … |
Localized Content
Manage system prompts, help content, and other localized strings
| Name | Preview |
|---|---|
| Budget Alert Key Last updated: 3/2/2026 | Your {platformName} API Key "{displayName}" has used {pct}% ... |
| Setup Instructions Last updated: 3/2/2026 | export ANTHROPIC_BASE_URL={{baseUrl}} export ANTHROPIC_A... |
| Name | Preview |
|---|---|
| Default Last updated: 3/2/2026 | You are Chat App, an access interface to AI Language Models ... |
| Cowork Agent Last updated: 3/2/2026 | COWORK AGENT SYSTEM PROMPT (AMENDS PREVIOUS INSTRUC... |
Configs Deployment
Setup
Up and Running in Minutes
Three steps from zero to a production AI platform.
Deploy
Clone the repo and run Docker Compose. The full platform spins up with PostgreSQL, Redis, Meilisearch, etc — all pre-configured.
Connect providers
Add your LLM provider API keys through ControlTower. Supports Anthropic, OpenAI, Google Gemini, OpenRouter - or your own inference endpoints. Manage access and pricing.
Manage Users and Teams
Set up user teams, assign budgets, and synchronize with your enterprise IdP groups. Control spend and budgets at the team level.
Pricing
Open Source. Deploy Anywhere.
SecondStack is free open source licensed. Run it on your hardware or let us help.
Community
Open source, forever
Full platform for teams getting started with self-hosted AI.
- All platform features
- Docker Compose deployment
- Community support via GitHub
- Multi-model chat interface
- Team & budget management
- Identity federation
Enterprise
Tailored to your needs
For organizations requiring dedicated support and advanced deployment.
- Everything in Community
- Kubernetes deployment support
- Priority support & SLA
- Custom integrations
- Deployment assistance
- Security review & hardening
Why self-host AI instead of using ChatGPT Enterprise or similar SaaS?
Why does an organization need an LLM routing gateway?
What does deployment look like?
What are the minimal deployment requirements?
How does user authentication work?
Which LLM providers and models are supported?
What is Cowork Agent?
What about data privacy and compliance?
Ready to own your AI infrastructure?
Deploy SecondStack in minutes. Open source, self-hosted, and built for teams that demand control.