You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It would be nice to have an optional mode permitting to monitor the number of token the Semantic Cache permits to save.
To do so, today we can store the token as metadata item.
Interesting idea. @justin-cechmanek curious your take on this. Is there a way we could build this as a generalized extension on the cache so that we could handle this elegantly (not just OpenAI obviously)
We could elevate the token_count to be a permanent optional field, similar to prompt and response, instead of requiring users to define it in meta data. Users will still have to pass in the token count per prompt/response because that will vary by LLM tokenizer, but we can handle the incrementing.
It would be nice to have an optional mode permitting to monitor the number of token the Semantic Cache permits to save.
To do so, today we can store the token as metadata item.
(...)
I am wondering if it would make sense to make such easier as it permits to highlight the Semantic Cache value.
The text was updated successfully, but these errors were encountered: