/go-infer

Go framework for DL model inference and API deployment

Primary LanguageGoBSD 3-Clause "New" or "Revised" LicenseBSD-3-Clause

Go framework for DL model inference and API deployment

中文

Deep learning models are usually deployed in the cloud and the inference services are provided through APIs. This framework provides the basic architectural components for deploying APIs and achieves several goals:

  • The API processing module is decoupled from the model inference module to reduce the risk of network and computing blocking caused by high concurrency.
  • The API processing module and model inference module can be deployed as a distributed architecture, and can achieve horizontal expansion.
  • The framework implemented using Go language to achieve execution efficiency and simplify deployment and maintenance.
  • Custom logic is implemented using callback, hiding common logic. Developers only need to focus on custom logic

Other features:

Distributed deployment architecture

Distributed deployment architecture

Development documentation

  1. Development Guide
  2. API documentation template
  3. Local testing
  4. Tensorflow runtime environment

Star History

Star History Chart