Kong Ai Gateway Bridge
Routes Naftiko-side LLM calls through Kong AI Gateway's universal LLM API (OpenAI / Anthropic / Gemini / Bedrock / Azure / Databricks / Mistral / HuggingFace under one OpenAI-compatible surface) — automatic semantic caching, token budgets, prompt firewalls, PII guardrails, automated RAG injection, and per-token observability with no per-capability LLM-vendor wiring.
What You Can Do
MCP Tools
chat-completion
Run an OpenAI-compatible chat completion through Kong AI Gateway (semantic cache + token budget + prompt firewall + PII guardrails).
completion
Run an OpenAI-compatible text completion through Kong AI Gateway.
embedding
Compute embeddings through Kong AI Gateway.
list-models
List the LLM models available through Kong AI Gateway (provider-agnostic).
list-token-budgets
List Kong AI Gateway token budgets (per-agent / per-period token caps).
create-token-budget
Create a new Kong AI Gateway token budget.
get-token-usage
Get Kong AI Gateway token consumption stats per model / agent / period.