Skip to content

[Enhancement]: Implement AWS Bedrock prompt caching #7109

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
1 task done
dvejsada opened this issue Apr 28, 2025 · 0 comments
Open
1 task done

[Enhancement]: Implement AWS Bedrock prompt caching #7109

dvejsada opened this issue Apr 28, 2025 · 0 comments
Labels
✨ enhancement New feature or request

Comments

@dvejsada
Copy link

What features would you like to see added?

Amazon Bedrock has made a prompt caching feature publicly available see here.

More details

The prompt caching works by adding a cache marker to a part of the prompt that remains static. In case of LibreChat, I believe that this would be beneficial for:

  • system prompts - namely system prompts for agents with many tools can get quite long and remain static over each message
  • documents added as text - if a user adds documents as text to conversation, this remains static over the conversation
  • tool call results (?)

The prompt caching can lead to decrease of API usage costs as cache read input tokens are 10 times cheaper that regular input tokens.

Which components are impacted by your request?

Endpoints

Pictures

No response

Code of Conduct

  • I agree to follow this project's Code of Conduct
@dvejsada dvejsada added the ✨ enhancement New feature or request label Apr 28, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
✨ enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant