-
Notifications
You must be signed in to change notification settings - Fork 248
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Why do you freeze batch norm parameters when training? #18
Comments
leao1995
changed the title
Why freeze batch norm parameters when training?
Why do you freeze batch norm parameters when training?
Nov 9, 2017
it's a common practice.First, Because the pretrain network's bn layers have been trained. Second,Object Detection 's batchsize is small, hard to make bn parameter stable. |
use Group norm instead of batch norm . it is more stable. |
Use synchronized batch normalization |
Using sync batch norm does not help with single GPU training and low batch sizes though. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
would it be better to let batch norm parameters adapt to your current data?
The text was updated successfully, but these errors were encountered: