-
Notifications
You must be signed in to change notification settings - Fork 42
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
the loss tend to be nan #5
Comments
@txiaoyun @d-acharya |
@xiaomingdaren123
def _cal_log_cov(features): But loss will grow bigger and fall again and again, also I did not reproduce the author's results. |
The gradient computation of tensorflow eigen decomposition is most likely producing NaNs. The proposed technique tensorflow_patch.txt worked previously on a different system (with occasional failures). Recently I tried it on a different system and it consistently produced NaNs too (on tensorflow 1.13 it produces nan after few epochs, where as in tensorflow 1.2 it produces nans after around 600 epochs). I will check if changing regularization and learning rate will avoid this. I will try to check this and update. Clipping is alternative solution and was actually used to train model4 and model2 mentioned in paper. However training again, I myself am unable to get same exact numbers. |
However, if you cannot get the numbers in the paper by using pretrained models, I would try following data: https://drive.google.com/open?id=1eh93I0ndg6X-liUJDYpWveIShLd0ao_x Different version of pickle or classifier was found to effect reported numbers. |
@d-acharya @txiaoyun @xiaomingdaren123 I didn't apply the patch suggested in tensorflow_patch. And what I use is python 3.5.
Could you give me some advice? Thank you. |
@txiaoyun @xiaomingdaren123 @d-acharya @YileYile |
@txiaoyun @xiaomingdaren123 @YileYile @fredlll |
Hi , how to solve the problem about 'without dlpcnn'? |
When I run the code, the train loss is nan. Could you give me some advice? Thank you.
The text was updated successfully, but these errors were encountered: