December '23 Migration
Last updated
Was this helpful?
Last updated
Was this helpful?
Date: 8th Dec, 2023
This December, we're pushing out some exciting new updates to Portkey's SDKs, APIs, and Configs.
are upped to major version 1.0 bringing parity with the new OpenAI SDK structure and adding Portkey production features to it. We are also bringing native Langchain & Llamaindex integrations inside the SDK. This is a Breaking Change that Requires Migration.
are upgraded with new endpoints, making it simpler to do /chat/completions
and /completions
calls and adding Portkey's production functionalities to them.
This is a Breaking Change that Requires Migration.
are upgraded to version 2.0, bringing nested gateway strategies with granular handling. For Configs saved in the Portkey dashboard, this is NOT a Breaking Change and we will Auto Migrate your old Configs. For Configs directly defined at the time of making a call, through the old SDKs or old APIS, they will fail on the new APIs & SDKs and require migration.
API (Old)
/v1/proxy
/v1/complete
/v1/chatComplete
/v1/embed
/v1/prompts/ID/generate
✅ SDK (Old) ❌ SDK (New) ✅ Configs (Old) ✅ Configs (New)
Q2 '24
API (New)
/v1
/v1/completions
/v1/chat/completions
/v1/embeddings
/v1/prompts/ID/completions
❌ SDK (Old) ✅ SDK (New) ❌ Configs (Old) ✅ Configs (New)
-
SDK Version < 1 (Old)
✅ API (Old) ❌ API (New) ✅ Configs (Old) ✅ Configs (new)
Q2 '24
SDK Version = 1 (New)
❌ API (Old) ✅ API (New) ✅ Configs (Old) ✅ Configs (new)
-
Configs 1.0 (Old)
✅ API (Old)
❌ API (New) ✅ SDK (Old) ❌ SDK (new) === Configs saved through the Portkey UI will be auto migrated.
Q2 '24
Configs 2.0 (New)
✅ API (Old)
✅ API (New) ✅ SDK (Old) ✅ SDK (new)
-
We recommend upgrading to these new versions promptly to take full advantage of their capabilities. While your existing code will continue to work until the deprecation date around Q2 '24, transitioning now ensures you stay ahead of the curve and avoid any future service interruptions. Follow along with this guide!
More extensible SDK that can be used with many more LLM providers
Out-of-the-box support for streaming
Completely follows OpenAI's SDK signature reducing your technical debt
Native support for Langchain & Llamaindex within the SDK (Python)
Support for the Portkey Feedback endpoint
Support for Portkey Prompt Templates
Older SDK versions to be deprecated soon
FROM ➡
TO ⬇
Installing the New SDK,
Introduced 3 new routes /chat/completions
, /completions
, and /embeddings
Simplified the headers:
x-portkey-mode
header is deprecated and replaced with x-portkey-provider
Which takes values: openai
, anyscale
, cohere,
palm
, azure-openai
, and more.
New header x-portkey-virtual-key
is introduced.
/complete
and /chatComplete
endpoints to be deprecated soon
Prompts endpoint /prompts/$PROMPT_ID/generate
is upgraded to /prompts/$PROMPT_ID/completions
and the old route will be deprecated soon
We now support updating the model params on-the-fly (i.e. changing temperature etc at the time of making a call)
Prompt response object on the /completions
route is now fully OpenAI compliant
New /gateway
endpoint that lets you make calls to third-party LLM providers easily
FROM ➡
TO ⬇
FROM ➡
TO ⬇
New concept of strategy
instead of standalone mode
. You can now build bespoke gateway strategies and nest them in a single config.
You can also trigger a specific strategy on specific error codes.
New concept of targets
that replace options
in the previous Config
If you are adding virtual_key
to the target array, you no longer need to add provider
,Portkey will pick up the Provider directly from the Virtual Key!
For Azure, only now pass the virtual_key
- it takes care of all other Azure params like Deployment name, API version etc.
FROM ➡
TO ⬇
Shoot ANY questions or queries you have on the migration to the Portkey team and we will try to get back to you ASAP.