You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
promptulate need cache for LLM generating. Having a cache means that the large model output is used on the first input, and the output is cached, and on the second input, if the same data has been entered before, the previously cached data is used directly.
🚀 Feature Request
promptulate need cache for LLM generating. Having a cache means that the large model output is used on the first input, and the output is cached, and on the second input, if the same data has been entered before, the previously cached data is used directly.
Method 1
For example:
The answer is generated by the gpt-4o driver during the first run and then into the cache, and the cached data is used directly during the second run.
Default no cache, if you want to open cache, use the following pattern:
When your cache_seed is 111, your cache is queried.
Method 2
Use enable_cache parameter. For exmaple:
The answer is generated by the gpt-4o driver during the first run and then into the cache, and the cached data is used directly during the second run.
Compare
The first approach is a little more granular and can be cached based on different user ids? Useless, prompt key is the same.
So method2 is simple and enough.
The text was updated successfully, but these errors were encountered: