[Feature] Prompt Caching
Hi @andrewpareles, could you please assign this issue to me? I'd like to start working on it right away and get it resolved as soon as possible. Thanks!
@andrewpareles I prompt caching means thinking in this direction like prompt caching refers to storing the results of previously sent prompts (or messages) to an LLM (Large Language Model) so that subsequent requests with the same or similar prompts can retrieve the cached response instead of sending the request to the LLM again.
Sure, just assigned you!
Yep, prompt caching here means sending the provider (Anthropic, OpenAI, OpenRouter) a message to cache the conversation. A lot of providers offer discount rates for cached inputs.
@karankulshrestha hey, how is this going?
hey @karankulshrestha.. hope everything’s going well! Just checking in on the prompt caching feature. there’s been a bit of buzz around it lately, especially since it ties into cost management (and indirectly to cost tracking PR #485 ). Totally understand if other things have come up... if you’re swamped no worries at all. would be great to hear your feedback. really appreciate you taking it on!