-
Notifications
You must be signed in to change notification settings - Fork 473
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[models] Add model compression utils #5
Comments
ONNX conversion seems to be incompatible with TF 2.4.* as per onnx/keras-onnx#662. I tried on my end and encountered the same problem. |
A good lead for ONNX support would be to use https://github.com/onnx/tensorflow-onnx (might have to create a savemodel to use it but it's worth a look) |
@frgfm i think we can remove the tensorrt point If we support onnx wdyt ? |
Yes sure! We'll need to take a look at pruning at some point |
yeah pruning is fine but tensorrt is a bit to much (should do the user on his own side if we can provide onnx this should be not so tricky) |
Add a
doctr.models.utils
module to compress existing models and improve their latency / memory load for inference purposes on CPU. Some interesting leads to investigate:Optional: TensorRT export (cf. https://developer.nvidia.com/blog/speeding-up-deep-learning-inference-using-tensorflow-onnx-and-tensorrt/)
The text was updated successfully, but these errors were encountered: