Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

At what time should i stop the searching and begin retraining my model? #88

Open
yuncheng97 opened this issue Jan 8, 2020 · 10 comments

Comments

@yuncheng97
Copy link

thanks lot for your wonderful efforts!
but I have some problems when implementing this project
I run the code for pascal and cityscapes with the default args, with one gpu RTX 2080Ti, but the result I got at epoch 40 is bad, pascal voc not even reach 20% miou and cityscapes also just around 20%. I'm so confused.
I saw your introduction that there should be 3 stages of searching, decoding and re-training. the re-training stage means I should start re-training until the search result reached ~79% or I should start re-training after ~40 epochs and re-train the model for the best result?
looking forward to your reply. thanks again.

@mrluin
Copy link

mrluin commented Jan 8, 2020

Just start to retrain after the ~40 epochs searching.
searching for 40 epochs -> retrain from scratch

@yuncheng97
Copy link
Author

Just start to retrain after the ~40 epochs searching.
searching for 40 epochs -> retrain from scratch

thanks! I will try this to see if it works

@cardwing
Copy link

@Oliver-jiang Can you share your retraining result of the searched architecture with us? Thanks a lot.

@cardwing
Copy link

@Oliver-jiang Since RTX 2080Ti only has 11GB memory, you may have banned global pooling in ASPP. That may be the reason for the bad performance.

@yuncheng97
Copy link
Author

yuncheng97 commented Feb 23, 2020 via email

@yuncheng97
Copy link
Author

@cardwing can you share some of your search_argparses with us? Thanks a lot.

@cardwing
Copy link

Currently, I have reduced the number of layers from 12 to 10 due to the limited memory and kept other hyperparameters fixed. The searching performance on the Cityscapes validation set is around 33% mIoU after 40 epochs.

@yuncheng97
Copy link
Author

@cardwing thanks for your advice, i tried your method but currently the best performance after 7 epochs is around 7% and there was a big fluctuation in some epochs. however, the performance at begin shown in the paper is already around 10% mIoU(1% in my experiment). did you have the same problem? looking forward to your reply

@cardwing
Copy link

cardwing commented Mar 3, 2020

I have not met that problem. You can just train the model for more epochs and check the result. However, it takes me around 7 days to train the searched architecture and the final performance is only around 65% mIoU. I have to admit that NAS really consumes resources and time.

@chendi23
Copy link

I have not met that problem. You can just train the model for more epochs and check the result. However, it takes me around 7 days to train the searched architecture and the final performance is only around 65% mIoU. I have to admit that NAS really consumes resources and time.

Same. around 65%

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants