Announcements

Stay informed about Portkey's newest releases, features, and improvements

  1. Portkey has raised $15M in Series A funding 🎉

    We have some exciting news to share: Portkey has officially closed our $15M Series A !!

    Huge thanks to our investors at Elevation and Lightspeed, as well as our team and community for getting us here.

    You have pushed us to be better, given us helpful feedback, and suggested the features that have made Portkey what it is today!

    This funding will continue expanding our go-to-market presence, deepen our governance capabilities for agentic AI, and scale the gateway that enterprises rely on to run AI in production.

    What’s more, we’re making our core enterprise gateway available for free so that all teams have access to governance and observability when they need it!!

    SM Post (1)

  2. 🆕 Sticky Load Balancing

    You can now route requests with the same identifier consistently to the same target when load balancing.

    This means:

    → Conversation context stays intact across multiple requests

    → A/B tests produce consistent, reliable results

    → Users get session-based routing for personalized experiences

    See how to set up this config ->

  3. Day 0 support for Claude Sonnet 4.6 ⚡

    Claude Sonnet 4.6 is Anthropic's new default model, delivering near-Opus level reasoning and a 1M token context window, all at mid-tier pricing. Built for agentic workflows, computer use, and long-context tasks.

    With Portkey, bring Claude Sonnet 4.6 into production with:

    ✅ Unified access via the AI Gateway

    ✅ Consistent guardrails and policy enforcement

    ✅ Full observability across logs, latency, and spend

    ✅ Team-level budgets and rate-limits

    claude sonnet 4.6 (1)
  4. GPT-5.3-Codex is now live on Portkey

    GPT-5.3-Codex is OpenAI’s latest model, designed for agentic, long-running coding workflows, covering everything from multi-step refactors to debugging across large repositories.

    With Portkey, you can bring GPT-5.3-Codex into production with:

    ✅ Unified access via the AI Gateway

    ✅ Consistent guardrails and policy enforcement

    ✅ Full observability across logs, latency, and spend

    ✅ Team-level budgets and rate-limit controls

    gpt-5.3 (1)
  5. Heads up if you’re using Gemini models 👀

    Google is deprecating Gemini 2.0 Flash and Gemini 2.0 Flash Lite on March 31, 2026 across Gemini API and Google AI Studio.

    If your apps or agents rely on these models, you’ll need to migrate to supported alternatives like Gemini 2.5 Flash or 2.5 Flash Lite before that date.

    image
  6. 🚀 The MCP Gateway is now GA!

    MCP adoption is accelerating rapidly, with servers and tools now being used across teams, environments, and agent workflows.

    As MCP usage scales, teams start running into a consistent set of challenges:

    • authentication varies across MCP servers
    • access to tools becomes difficult to govern
    • visibility into MCP usage is limited

    Portkey’s MCP Gateway solves this!

    It is a control layer to securely connect MCP servers, standardize authentication, manage tool access, and get clear visibility into MCP usage, without changing existing agents or MCP servers.

    Check it out ->

  7. Open-sourcing the LLM pricing database!!

    Keeping LLM pricing accurate is difficult because it isn’t static. Models change, pricing shifts, and new billing dimensions are introduced over time.

    To handle this, we built a never-goes-old database that powers cost calculations ~$250,000 in LLM spend, every day.

    We’ve now open-sourced the pricing database produced by this system so others can inspect how pricing is tracked, reuse the data, and understand the assumptions behind cost calculations.

    The database covers 2,000+ models across 40+ providers, including thinking tokens, cache pricing, context tiers, and non-token fees. It’s available via a free public API with no authentication required.

    Take a look ->https://portkey.ai/models

    llms.new
  8. OpenCode Integration!

    Running OpenCode in shared or production environments?

    By using Portkey alongside OpenCode, teams can add the controls needed to run it at scale, without changing how developers work.

    You get:

    ✅ Access control, budgets, and usage limits at the platform layer

    ✅ Centralized governance without modifying OpenCode itself

    ✅ Full observability across usage, latency, and spend

    ✅ Guardrails to keep usage predictable in shared environments

    See how you can run OpenCode with governance

    3

  9. Anthropic models on Azure are now accessible via Portkey!

    You can now access Anthropic models on Azure via the native /messages endpoint, so you can keep the standard Anthropic request format while getting:

    ✅ Unified access through the Portkey Gateway

    ✅ Guardrails for safe, compliant usage

    ✅ Full observability across logs, latency, and spend

    ✅ Team-level budgets and rate-limit controls

    2