In this project I:
- Converted MS COCO annotations to YOLO annotations
- Built object detection model
- Accelerated the model (quantization, smaller inference size, optimization),
- Suggested metrics for tracking model degradation and data consistency in production
- Converted model to ONNX format
- Deployed model on Nvidia Triton with TensorRT backend engine (for maximum inference speed)
- Wrote grpc client and tested inference FPS
Main Goal: fastest inference
Test machine config:
- No GPU
- Core i7-8th
- 16 GB RAM
Learning artefacts + metric plots:
https://wandb.ai/svtdanny/yolov5/overview
Results (imahes sended sequentially, batch_size = 1):
Object_Detection.ipynb [MS COCO -> YOLO labels, transfer learning, accelerating, converting]
Online_validation.ipynb [online metrics]