sagemaker-inference-toolkit
sagemaker-inference-toolkit copied to clipboard
Default output function encodes results to JSON and that seems to add to response latency.
Describe the bug By default the accept type in inference container seems to be application/json. The default encoder which converts results to JSON seems to add significantly to the response latencies. Is there a way to reduce the default response's latencies?
up, did you find the solution?