clipper icon indicating copy to clipboard operation
clipper copied to clipboard

Benchmark against SageMaker and Tensorflow Serving

Open simon-mo opened this issue 6 years ago • 4 comments

A blog post?

simon-mo avatar Oct 12 '18 00:10 simon-mo

in the paper the object recognition experiment was setup using cifar-10, imagenet, would you consider using bench marking using coco data instead? would you be open source model network that you use?

wcwang07 avatar Feb 13 '19 19:02 wcwang07

ran uniform load with single arrival process sending one randomly generated color image of size 224x224 for every 120 seconds

image

wcwang07 avatar Feb 17 '19 01:02 wcwang07

In TFS, here they set batching parameter to test model latency https://github.com/tensorflow/serving/issues/344

we know that clipper also use adaptive batching mechanism that is determined by https://github.com/ucbrise/clipper/blob/3c5a1cc6ce59e0ccd778f526a50808d0e7b2576f/src/libclipper/src/containers.cpp#L128 (https://github.com/ucbrise/clipper/issues/548)

could you show whether batching size can be set at runtime just like TFS? if so what different sizes have you tried for image recognition benchmark?

Thank you.

wcwang07 avatar Feb 17 '19 02:02 wcwang07

Clipper cannot dynamically change arbitrary batchsize at runtime, however you can set the max_batch_size similar to TFS when you deploy the model: http://docs.clipper.ai/en/v0.3.0/clipper_connection.html#clipper_admin.ClipperConnection.build_and_deploy_model

batch_size (int, optional) – The user-defined query batch size for the model. Replicas of the model will attempt to process at most batch_size queries simultaneously. They may process smaller batches if batch_size queries are not immediately available. If the default value of -1 is used, Clipper will adaptively calculate the batch size for individual replicas of this model.

simon-mo avatar Feb 17 '19 05:02 simon-mo