Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

unet keeps producing nan during training #18

Open
EYcab opened this issue Jan 22, 2024 · 2 comments
Open

unet keeps producing nan during training #18

EYcab opened this issue Jan 22, 2024 · 2 comments

Comments

@EYcab
Copy link

EYcab commented Jan 22, 2024

1705914514278
Anyone knows why this unet process always produces nan results despite all the settings are done accordingly and all the other input variables are the same

@EYcab EYcab changed the title unet keep producing nan during training unet keeps producing nan during training Jan 22, 2024
@junyongyou
Copy link

Yes, I encountered the same here: loss becomes nan after some epochs. I tried different reward functions, and all the same.

@junyongyou
Copy link

I figured out the reason. You can change config.mixed_precision to "no" in base.py, such that full-precision can be enabled, and it should avoid that unet produces NaN.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants