model_server
model_server copied to clipboard
A scalable inference server for models optimized with OpenVINO™
* Benchmark client documentation update * Benchmark client doc add precisions and png dump * Fix doc Co-authored-by: katarzyna.papis Co-authored-by: kmatejuk Co-authored-by: Katarzyna Papis
I have a feature extractor with input shape (-1,-1,-1,-1) which refers to (b,c,h,w) and output of (-1,-1,512) refers to (b,num_steps,512). I write a config file with following model_config_list and try...
Hi, I run benchmark client in python to measure performance of model server, but I am a bit confused by the meaning of various metrics e.g. 'worker: mean_latency', 'worker: mean_latency2'...