The API gateway built for the AI era. Meter usage, support LLM response streaming, and charge exactly for what your users consume.
14-day free trial · No credit card required
POST /v1/completions
Tokens
1,247
Cost
$0.037
Latency
142ms
Adding a single line to your Nadles configuration enables automatic LLM token metering. Nadles can meter token usage automatically for many popular streaming response formats — OpenAI Completions API, OpenAI Responses API, Anthropic API, Ollama, Gemini, and more.
Read the docs →
Every feature designed for the realities of LLM and AI API monetization.
Nadles proxies streamed responses natively — no buffering, no dropped chunks, no latency penalty.
Meter by tokens, API calls, characters, or any unit. Let Nadles meter usage automatically or manually report usage per request via the Nadles API.
For complex AI pipelines (multi-step, batch processing, agents), report usage asynchronously after the fact. Nadles reconciles it with the correct billing period.
Charge a base fee + per-token overage, offer prepaid token bundles, or go pure pay-as-you-go. Model any pricing structure without code changes.
Your users don't consume your API in predictable monthly chunks. They stream responses, burn tokens at variable rates, and expect to pay only for what they use.
Standard billing tools can't handle streamed responses or per-request usage reporting.
One heavy user calling GPT-4 can wipe out your margin on a fixed plan.
Building idempotent, async usage pipelines yourself takes months.
Nadles solves all three.
Native LLM streaming support, token metering, and usage-based billing — built into the gateway itself. No custom code required.
AI-specific features are just the start. Nadles is a complete revenue stack.
You already have the model. Here's how you start charging for it.
Link your Stripe or Paddle account. Revenue goes directly to you.
Enter your API origin. Nadles deploys a gateway with streaming support.
Set per-token rates, prepaid bundles, or hybrid plans in the dashboard.
Share a checkout link. Customers buy tokens and start calling your model.
All AI features — streaming, token metering, usage-based billing — are included on every plan.
Starter Plan
14-day free trial · No credit card required
Yes. Nadles natively proxies Server-Sent Events (SSE) and chunked transfer encoding — no buffering, no extra latency. After each request completes, report token usage via the Nadles API for accurate per-request billing.
Yes. Define separate meters for prompt tokens and completion tokens, then price them independently. Useful for models where input and output costs differ significantly.
Use async usage reporting. After your pipeline completes, report the total usage via API. Nadles reconciles it with the billing period and customer automatically.
No. The gateway runs globally at the edge. Auth is verified without database lookups. For streaming responses, Nadles forwards chunks as they arrive — no buffering overhead.
Yes. Offer prepaid credits (e.g., "Buy 1M tokens for $49"), pay-as-you-go, or hybrid models (base fee + token overage). All configurable from the dashboard.
Stop building billing infrastructure. Start charging for tokens.
14-day free trial · No credit card required