- ✓ 1 application
- ✓ 2,000 API calls / week
- ✓ 1 MCP server endpoint
- ✓ 1M tokens / mo included
- ✗ No overage (upgrade to add)
- ✓ Basic Glass Box tracing
- ✓ 1 team member
- ✓ Community support
Pay a flat base fee for the platform, then only for what your agents actually consume. Bring your own AI client and pay per application — not per token.
Run AI workflows through OBTO's hosted models. You pay a base platform fee that includes a monthly token allowance, then utility billing on usage above the threshold. Powered by Groq GPT-OSS 120B — $0.15 input / $0.60 output per million tokens, passed through at cost + 20%.
Connect Claude Desktop, ChatGPT, Cursor, or any MCP-compatible client to your OBTO endpoint. You keep your existing LLM subscription — we charge per deployed application, not per token. Each app gets its own MCP server, rate limits, and Glass Box audit trail.
The raw Glass Box platform. Start free, scale predictably on utility billing.
No hidden markups beyond the 20% platform fee. You can verify the base rate at groq.com/pricing at any time. That's what Glass Box means.
Already paying for Claude, ChatGPT, or Cursor? Connect them to OBTO's MCP endpoint and use OBTO as your deployment infrastructure — not your inference provider. You keep your existing AI subscription. We charge per application hosted.
yourapp.obto.coApplications count against your plan's app limit. Rate limits apply per app per week. No token charges — OBTO doesn't touch your LLM inference.
Most dev agencies charge you for 40 hours of manual coding. We don't. Our team builds using the OBTO AI platform — so we execute 10x faster. You pay for expert architecture and rapid assembly, not slow typing.
Perfect for launching an MVP. Buy a dedicated block of hours where our team sits with you, uses our platform, and ships your app in days — not months.
Calculate my Sprint →A dedicated OBTO expert acts as your fractional CTO — helping you refine workflows, integrate complex APIs, and build MCP servers behind the scenes.
Discuss a Retainer →One deployed OBTO app with its own MCP endpoint, domain, and backend. Each app can have multiple pages, routes, and server scripts. The limit is per deployed app, not per page or feature.
You pay the flat base fee regardless of usage — that covers the platform, your included token allowance, and your app slots. If you consume more tokens than your plan includes, you're billed at the overage rate at end of month. No surprise spikes — your Glass Receipt shows usage in real-time.
No. When you connect a BYO AI client via MCP, OBTO doesn't route your inference — your client talks to its own LLM provider. OBTO only charges for the application hosting, rate limits, and infrastructure. Your token bill stays with Anthropic, OpenAI, or whoever you're using.
It's the best price-performance model for agentic workloads right now — 500 tokens/sec, 128K context, full tool-calling support, and the lowest cost per capable token on the market. We pass the rate through at cost + 20% with no hidden markup. You can verify the base rate at groq.com/pricing.
Yes — that's what Enterprise is for. You run the entire OBTO runtime on your own Kubernetes cluster. You still need an Enterprise agreement for support and updates, but you're not paying per-token or per-app to us.
Rate limits apply per application per week — not per user or per account. Builder: 2,000 calls/week. Team: 50,000 calls/week per app. Business: fair use (no hard limit, but we reserve the right to throttle runaway loads). Enterprise: unlimited with dedicated infrastructure.
Start building on the platform yourself, or let our experts help you map out your first architecture sprint.