Announcements

Stay informed about Portkey's newest releases, features, and improvements

  1. Configure request logging!

    Not every team wants to log everything. Some need full visibility into every LLM request. Others have strict compliance rules and want to keep things minimal.

     

    Portkey now gives org owners control over what gets logged — from full request details to just high-level metrics.

     

    You can apply this org-wide, or allow individual workspace managers to modify it for their teams.

     

     Configure it in your org  

     

    Request Logging
  2. Kimi K2 is now live on Portkey!

    Moonshot’s latest release, Kimi K2, is a low-cost, open-source model with capabilities that surpass top models on coding tasks.

    Bring Kimi K2 in production with Portkey’s AI gateway that provides:

     

    ✅ Smart routing, retries, and failover

    ✅ Guardrails for safe, compliant usage

    ✅ Full observability — logs, latency, and cost

    ✅ Budget and rate-limit controls across teams

     

     Integrate Kimi K2 with Portkey -> 

    Moonshot
  3. ⚡ Day 0 support for Grok 4!

    You can now start using xAI's Grok 4 via Portkey’s AI gateway.

     

    Grok 4 currently tops the Artificial Intelligence Index and outperforms others in coding and math tasks.

    With Portkey, bring Grok 4 into production with:

    ✅ Smart routing, failover, and retries built in

    ✅ Guardrails for safe, compliant interactions

    ✅ Full observability with logging, latency, and cost insights

    ✅ Budget and rate-limit controls across use cases

    xAI (2)
  4. Circuit Breaker is live!

    Handling millions of requests a day?

     

    Retries and fallbacks work for quick failures, but when a target has prolonged failures, they can trigger cascading effects that may lead to increased load and latency across the board.

    Circuit breakers fix this. They monitor failures and latency, and stop routing to degraded targets until they recover.

     

    Portkey now supports circuit breakers in configs, because at scale, slow is the new downtime. Learn more

    Untitled diagram _ Mermaid Chart-2025-07-08-104701
  5. New integration alert: Portkey 🤝 Future AGI

    Operational metrics tell you what happened, evals tell you how well it worked.

     

    With Portkey × Future AGI integration, you now get both together!!

    Portkey already tracks every request: which model was used, how long it took, and how much it cost. With FutureAGI plugged in, you can evaluate the quality of those responses, scoring for task success, clarity, and accuracy on real traffic.

    No changes to your infrastructure. No rerouting.

     

     See how you can set it up quickly ->  

     

    image
  6. Better availability, fewer 429s.

    Vertex AI has rolled out global endpoints, and they’re now fully supported on Portkey. If you’ve hit “resource exhausted” errors before, switching to the global endpoint can significantly improve availability and reduce throttling under load.

     

    To enable it on Portkey, just set the region to global in your Vertex AI virtual key. Everything else works the same.

     

    Models supported via global endpoints:

    – Gemini 2.5 Flash-Lite

    – Gemini 2.5 Pro

    – Gemini 2.5 Flash

    – Gemini 2.0 Flash

    – Gemini 2.0 Flash-Lite

     

    vertex AI
  7. ⚡ Day 0 support for Gemini 2.5 Pro, Flash, and the new Flash-Lite!

    Gemini 2.5 Pro and 2.5 Flash models are now stable and generally available. 2.5 Flash-Lite is the most cost-efficient and fastest Gemini 2.5 model yet.

    These models offer long context lengths, making them especially useful for coding.

    You can now access these models via Portkey and get:

     

    ✅ Smart routing and fallback options

    ✅ Full logs, usage metrics, and cost tracking

    ✅ Org-wide rate limits, retries, and guardrails

    ✅ Prompt playground to test, version, and deploy prompts

    gemini (3)
  8. 🔥 Now available on Portkey

     gemini-2.5-pro-preview-05-06 and gemini-2.5-flash-preview-05-20 Models are now live and ready to use via the AI Gateway, along with:

     

    ✅ Org-wide guardrails, retries, and rate limits

    ✅ Full logging and metadata for every call

    ✅ Usage, latency, and cost tracking across all providers

    gemini (1)
  9. Langroid now supports Portkey out of the box!

    Langroid is a powerful Python framework purpose-built for agentic workflows with multi-agent programming.

    Teams building on Langroid can now easily plug into Portkey's AI Gateway and get:

     

    ✅ Unified access to 1600+ models

    ✅ Caching, retries, and fallbacks

    ✅ Prompt-level safety guardrails

    ✅ Logs, cost metrics, and full observability

    ✅ Budget enforcement and routing

     

    Explore the integration

     

  10. 🚨 Model Deprecation Alert

    OpenAI is deprecating gpt-4o–realtime-preview-2024-10-01 model.

    The endpoint will stop responding entirely after September 10, 2025.

     

    Quick action items:

    ✅ Audit any usage of this specific model

    ✅ Migrate to gpt-4o-realtime-preview release

    ✅ Complete all tests before the cutoff