idea: Add Claude Prompt Caching Support to Jan #3715
Labels
category: model support
Support new model, or fix broken model
category: providers
Local & remote inference providers
type: feature request
A new feature
Problem Statement
Using an API to access Claude without requiring a $20/month Claude Pro subscription saves users money in the long run. However, this benefit diminishes whenever users enter long context multi-turn conversations with Claude, especially with file uploads
Feature Idea
By integrating Anthropic's newer prompt caching feature, the response latency and cost for users could be dramatically reduced. Users may engage with longer multi turn conversations without starting a new thread and losing context. Cached prompts can be an optional feature that costs 90% less compared to uncached prompts per API call
The text was updated successfully, but these errors were encountered: