Skip to content

updating batch norm after ema or checkpoint averaging #403

Answered by rwightman
shairoz-deci asked this question in Q&A
Discussion options

You must be logged in to vote

@shairoz-deci there are multiple q about that in past issues, discuss... but in my experience, no it is not necessary and I feel it works better not to. Many of the training recipes for popular models in Tensorflow, such as EfficientNet, etc average the BN stats as well... it really just gives you a longer time const for the stats since they are already ema'd with the momentum param. I don't see why it would cause them to deviate enough to be a problem.

Replies: 1 comment 3 replies

Comment options

You must be logged in to vote
3 replies
@shairoz-deci
Comment options

@rwightman
Comment options

@shairoz-deci
Comment options

Answer selected by shairoz-deci
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants