Replies: 2 comments 3 replies
-
|
Hi @kharbat, what do you mean with prompt caching? What's the use case you are thinking of? |
Beta Was this translation helpful? Give feedback.
-
|
Hi @kharbat You can enable it by passing the prompt_cache_key inside the parameters array, for example: This allows the Responses API to cache repeated parts of your prompt according to OpenAI’s prompt caching guidelines. One thing that is still missing, though, is registering the cached_tokens field inside the chat_history. @ilvalerione — if this is something you’d like to have in the provider, I’d be happy to contribute and open a PR implementing it. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello, thanks for the nice work. I'm looking to replace my from-scratch architecture and give this a try - I was wondering if it supports prompt caching
Beta Was this translation helpful? Give feedback.
All reactions