go-infer icon indicating copy to clipboard operation
go-infer copied to clipboard

Go framework for DL model inference and API deployment

Go framework for DL model inference and API deployment

中文

Deep learning models are usually deployed in the cloud and the inference services are provided through APIs. This framework provides the basic architectural components for deploying APIs and achieves several goals:

  • The API processing module is decoupled from the model inference module to reduce the risk of network and computing blocking caused by high concurrency.
  • The API processing module and model inference module can be deployed as a distributed architecture, and can achieve horizontal expansion.
  • The framework implemented using Go language to achieve execution efficiency and simplify deployment and maintenance.
  • Custom logic is implemented using callback, hiding common logic. Developers only need to focus on custom logic

Other features:

  • Yaml is used for server-side configuration, which can be configured separately during distributed deployment.
  • API signature supports SHA256 and SM2 algorithms.
  • Model examples:
    • BERT model inference example
    • CNN model inference example
    • Tensorflow model weight conversion example
    • Keras model weight conversion example
    • ONNX model weight inference example
    • PyTorch model conversion to ONNX format example

Distributed deployment architecture

Distributed deployment architecture

Development documentation

  1. Development Guide
  2. API documentation template
  3. Local testing
  4. Tensorflow runtime environment

Star History

Star History Chart