Skip to content

Code to make running model inference easy in different backends. ONNX and Triton Inference Server are now available.

Notifications You must be signed in to change notification settings

TheConstant3/EasyInfer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

EasyInfer

Code to make running model inference easy in different backends. You don't have to worry about input/output names, dtypes and batch size. ONNX and Triton Inference Server are now available.

Quick start

  1. Run run_env.sh for building docker image and run docker container.
  2. In docker environment from example directory run python3.8 prepare_model.py for export resnet18 to onnx with dynamic and static batch size
  3. Out of docker environment from example directory run run_triton.sh for start Triton Inference Server with two exported models
  4. In docker environment from workdir run python3.8 main.py for send batch with size 12 to models with size 8 in triton and onnx format.

About

Code to make running model inference easy in different backends. ONNX and Triton Inference Server are now available.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published