multithreaded-estimators icon indicating copy to clipboard operation
multithreaded-estimators copied to clipboard

Multithreading inference in Tensorflow Estimators. This is a ServiceNow Research project that was started at Element AI.

ServiceNow completed its acquisition of Element AI on January 8, 2021. All references to Element AI in the materials that are part of this project should refer to ServiceNow.

Multithreaded-estimators

Code demonstrating how to use multithreading to speedup inference for Tensorflow estimators.

Installation

A Dockerfile is provided. First build the image from the root directory:

docker build . -t threaded

Then run the tests:

docker run threaded

License

This code is released under an Apache 2 license. See the license in full.