Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[models] Add model compression utils #5

Open
5 of 6 tasks
fg-mindee opened this issue Jan 11, 2021 · 5 comments
Open
5 of 6 tasks

[models] Add model compression utils #5

fg-mindee opened this issue Jan 11, 2021 · 5 comments
Assignees
Labels
framework: pytorch Related to PyTorch backend framework: tensorflow Related to TensorFlow backend help wanted Extra attention is needed module: models Related to doctr.models
Milestone

Comments

@fg-mindee
Copy link
Contributor

fg-mindee commented Jan 11, 2021

Add a doctr.models.utils module to compress existing models and improve their latency / memory load for inference purposes on CPU. Some interesting leads to investigate:

Optional: TensorRT export (cf. https://developer.nvidia.com/blog/speeding-up-deep-learning-inference-using-tensorflow-onnx-and-tensorrt/)

@fg-mindee fg-mindee added type: enhancement Improvement help wanted Extra attention is needed module: models Related to doctr.models labels Jan 11, 2021
@fg-mindee fg-mindee added this to the 0.1.0 milestone Jan 11, 2021
@fg-mindee fg-mindee self-assigned this Jan 12, 2021
@fg-mindee fg-mindee removed the type: enhancement Improvement label Jan 12, 2021
@fg-mindee
Copy link
Contributor Author

fg-mindee commented Feb 2, 2021

ONNX conversion seems to be incompatible with TF 2.4.* as per onnx/keras-onnx#662. I tried on my end and encountered the same problem.
Moving this to the next release until this gets fixed!

@fg-mindee fg-mindee modified the milestones: 0.1.0, 0.2.0 Feb 2, 2021
@fg-mindee
Copy link
Contributor Author

A good lead for ONNX support would be to use https://github.com/onnx/tensorflow-onnx (might have to create a savemodel to use it but it's worth a look)

@charlesmindee charlesmindee added the critical High priority label Apr 2, 2021
@fg-mindee fg-mindee modified the milestones: 0.2.0, 0.2.1 May 7, 2021
@fg-mindee fg-mindee modified the milestones: 0.2.1, 0.3.0 May 27, 2021
@fg-mindee fg-mindee modified the milestones: 0.3.0, 0.3.1 Jul 1, 2021
@fg-mindee fg-mindee added framework: pytorch Related to PyTorch backend framework: tensorflow Related to TensorFlow backend labels Jul 6, 2021
@fg-mindee fg-mindee modified the milestones: 0.3.1, 0.4.0 Aug 26, 2021
@fg-mindee fg-mindee removed the critical High priority label Aug 26, 2021
@fg-mindee fg-mindee modified the milestones: 0.4.0, 0.4.1 Sep 28, 2021
@fg-mindee fg-mindee modified the milestones: 0.4.1, 1.0.0 Nov 12, 2021
@felixdittrich92
Copy link
Contributor

@frgfm i think we can remove the tensorrt point If we support onnx wdyt ?

@frgfm
Copy link
Collaborator

frgfm commented Sep 5, 2022

Yes sure! We'll need to take a look at pruning at some point

@felixdittrich92
Copy link
Contributor

yeah pruning is fine but tensorrt is a bit to much (should do the user on his own side if we can provide onnx this should be not so tricky)

@felixdittrich92 felixdittrich92 modified the milestones: 1.0.0, 2.0.0 Jun 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
framework: pytorch Related to PyTorch backend framework: tensorflow Related to TensorFlow backend help wanted Extra attention is needed module: models Related to doctr.models
Projects
None yet
Development

No branches or pull requests

4 participants