Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Loss goes to NaN during training #14

Open
marshallllllll opened this issue Jun 7, 2018 · 1 comment
Open

Loss goes to NaN during training #14

marshallllllll opened this issue Jun 7, 2018 · 1 comment

Comments

@marshallllllll
Copy link

After a few hundred global steps, the loss goes to NaN. Any ideas as to why?

@parksunwoo
Copy link

parksunwoo commented Jun 8, 2019

Hi, I have a same problem...

8 random batch data selected by DataLoader... Done!
from graph, get pred_conf, pred_loc ... Done!
calculating (positive, negative, true_location) with jaccard overlap by DefaultBoxControl ... calculate_pos_neg_trueloc START
matching_dbboxes_gtboxes_in_batch_i START
get_top_confidences START
get_top_confidences END
matching_dbboxes_gtboxes_in_batch_i END
prepare_pos_neg_trueloc_in_matches START
prepare_pos_neg_trueloc_in_matches END
calculate_pos_neg_trueloc END
Done!
...
GLOBAL STEP : 29003 / LEARNING RATE : 0.001 / LOSS : nan ( 1.9297988414764404 secs)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants