Steward (Gateway)
A transparent HTTP proxy that sits between your application and LLM providers. Steward handles all request traffic: it proxies calls to OpenAI, Anthropic, and Gemini; parses token usage; calculates cost using live pricing data; writes full request/response bodies to your S3/GCS bucket; and logs only metadata (tokens, cost, latency, model, custom tags) to Majordomo. Streaming is handled transparently. Managed — Majordomo operates Steward on its own infrastructure. You configure your cloud storage bucket, create an API key, and point your SDK at the gateway endpoint. No servers to run. Self-hosted — You run Steward inside your own VPC. Your prompts and completions never leave your network — not even to Majordomo. Only metadata (token counts, cost, latency, model name) is sent outbound to Majordomo Cloud. The right choice for teams with data residency requirements or enterprise contracts that specify where AI data is processed. Both modes write request/response bodies to your bucket. The difference is where Steward runs. See also:Butler (Control Plane API)
Majordomo’s cloud service backing the dashboard and advanced features.- Stores usage metadata and serves analytics for the Web app
- Manages API keys, steward tokens, and provider mappings (encrypted at rest)
- Orchestrates Replay and Evals jobs and stores their results
- Maintains metadata key discovery and indexing (for fast filters)
Web (Dashboard)
The Majordomo dashboard UI, powered by Butler’s APIs.- Usage explorer, cost breakdowns, and request detail
- Metadata Keys management (discover, label, activate/index)
- Replay and Evals creation, status, and results
- Settings for Cloud Body Storage (S3/GCS) and Provider API keys