Announcements

Stay informed about Portkey's newest releases, features, and improvements

  1. 🚀 Lessons from Building a Multi-Billion Scale AI Gateway

    From 0 to 2Bn+ requests on Portkey’s AI Gateway!

    Listen to Rohit Agarwal as he shares lessons from building a billion-scale AI Gateway, in conversation with Shaw Talebi from The Data Entrepreneurs

     

    Why do 75% of AI projects get stuck in the proof of concept stage? Rohit shares how Portkey is helping 600+ teams overcome the production challenges in AI deployment.

     

     📺 Watch here 

     

    Like

  2. Launching Prompt Folders

    New Feature

    This was a long standing demand from many users: The ability to organize prompt templates inside folders.

    This is available to EVERY org now!

    • Create multiple folders as you need
    • Move any prompt from one folder to another

    Try it out now!

    Like

  3. New Integration: Portkey x Open WebUI

    Announcement

    Open WebUI (45k+ stars) is one of the most beloved open-source ChatGPT clones.

    Now you use Portkey with Open WebUI natively. Connect to 250+ LLMs (ChatGPT, Claude, Ollama) through a single interface in Open WebUI.

    Track costs, monitor token usage, measure response latency, and collect user feedback - all through 40+ on Portkey's dashboard.

     

    Learn more at Portkey docs 

    Figma Group 1000006801

    Like

  4. ✨ New Integration: Portkey AI × vLLM

    Announcement

    vLLM is a fast and easy-to-use library for LLM inference and serving, with State-of-the-art serving throughput.

    Portkey now integrates seamlessly with vLLM.

    Run Llama, Mistral, Qwen, and other open-source LLMs with Portkey to:

    1. Save $$$s on your GPU costs using semantic caching
    2. Full stack observability (logs, tokens, latency)
    3. Built-in reliability features: load-balancing, fallbacks, etc.
    4. Prompt management & versioning
    5. Guardrails to protect your data

    and much more...

     Link to docs 

     

    Like

  5. New Integration: Nvidia's Triton Inference Server on AI Gateway

    Announcement

    🎉 Exciting news!

    Portkey now supports Nvidia's Triton Inference Server.

    Big Thanks to Tapan Sharma for contributing to Portkey's open-source AI gateway with the integration.

    Triton is NVIDIA's inference server that makes AI model deployment lightning-fast and scalable across multiple GPUs and CPUs.

     Link to docs 

    Group from Figma (1)

     

     

    Like

  6. New Integration: MindsDB 🤝 Portkey AI

    We're excited to share that MindsDB now integrates natively with Portkey!

    Easily connect your databases, vector stores, and apps to 250+ LLMs with enterprise-grade monitoring and reliability built-in.

     

    Transform your existing data infrastructure into advanced, AI-powered systems today! 🚀  

     Link to docs 

    Like