### Describe the feature refer to: - https://github.com/triton-inference-server/server/blob/main/docs/user_guide/response_cache.md Some ML models might benefit from the cache. As for the storage part, I think ideally we should support both local and remote cache. ### Why do you need this feature? _No response_ ### Additional context _No response_
Describe the feature
refer to:
Some ML models might benefit from the cache.
As for the storage part, I think ideally we should support both local and remote cache.
Why do you need this feature?
No response
Additional context
No response