Skip to content

ResNet strikes back: And what about fine tuning? #901

Discussion options

You must be logged in to vote

@alexander-soare a lot of the ideas have some amount of transfer, but one usually (not always) scales back the degree of augmentation, regularization, learning rate, epochs for transfer learning. I feel fine-tuning is usually easier than training from scratch but you can certainly see big differences across hparam choices. The transfer settings here were similar to the A3 from-scratch settings, LR lower but not drastically so.

The transfer runs were done late in the process when Hugo and Herve had some free cycles on their training infra. It was an important sanity check to do but not a focus here so extensive search wasn't done and no effort to target different transfer for different sou…

Replies: 1 comment 4 replies

Comment options

You must be logged in to vote
4 replies
@AntixK
Comment options

@rwightman
Comment options

@mert-kurttutan
Comment options

@luckyhug
Comment options

Answer selected by alexander-soare
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
5 participants