Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: adding cache to LLMEndpoint #3555

Merged
merged 1 commit into from
Jan 27, 2025
Merged

feat: adding cache to LLMEndpoint #3555

merged 1 commit into from
Jan 27, 2025

Conversation

jacopo-chevallard
Copy link
Collaborator

@jacopo-chevallard jacopo-chevallard commented Jan 27, 2025

This allows us to avoid repeating expensive operations, such as reloading the tokenizers, at each call

Closes ENT-394

@dosubot dosubot bot added the size:S This PR changes 10-29 lines, ignoring generated files. label Jan 27, 2025
Copy link

linear bot commented Jan 27, 2025

ENT-394 Cache tokenizer

@dosubot dosubot bot added the lgtm This PR has been approved by a maintainer label Jan 27, 2025
@AmineDiro AmineDiro merged commit 6072907 into main Jan 27, 2025
7 checks passed
@AmineDiro AmineDiro deleted the feat/cache_LLMEndpoint branch January 27, 2025 10:47
StanGirard added a commit that referenced this pull request Jan 27, 2025
🤖 I have created a release *beep* *boop*
---


##
[0.0.30](core-0.0.29...core-0.0.30)
(2025-01-27)


### Features

* adding cache to LLMEndpoint
([#3555](#3555))
([6072907](6072907))

---
This PR was generated with [Release
Please](https://github.com/googleapis/release-please). See
[documentation](https://github.com/googleapis/release-please#release-please).
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
lgtm This PR has been approved by a maintainer size:S This PR changes 10-29 lines, ignoring generated files.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants