-
Notifications
You must be signed in to change notification settings - Fork 99
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
What's meaning of wh_offset_base ? #34
Comments
The wh_offset_base enables smaller, more easily predicted logits for wh. Wh logits are more similar to hm logits using this method. The loss values after weighting are balanced for wh and hm. TTFNet uses a different wh loss than centernet requiring a different loss re-weighting to keep the two losses in approximate balance. |
|
Logits refer to the feature maps fed into the final activation functions. Wh logits --> relu --> wh_offset_base --> wh prediction Hm logits --> sigmoid --> hm prediction The sigmoid curve yields most of its error with values greater than -5 for 0 targets and less than 5 for 1 targets so -5 to 5 is the typical interval the logits will fall into. TTFNet predicts the wh offsets at 512 x 512 image scale directly using relu as the activation function. E.g. a 142 length offset would require a logit of 142 . Using a wh_offset_base of 16 reduces the 142 logit to 8.9. Maintaining similar magnitude for both hm and wh logits will ease convergence. |
Thanks for your patience! |
It seems that the model will converge much more slower if wh_offset_base =1, especially the wh_loss. Besides, compared with centernet, the loss weight of wh_loss is much more larger than hm_loss(centernet: wh_loss_weight=0.1, hm_loss=1),why?
The text was updated successfully, but these errors were encountered: