深度学习中遇到NAN loss
-
什么都不改,重新训练一下,有时也能解决问题
-
学习率减小
-
检查输入数据(x和y),如果是正常突然变为NAN,有可能是学习率策略导致,也可能是脏数据导致
-
If using integers as targets, makes sure they aren’t symmetrical at 0.
I.e., don’t use classes -1, 0, 1. Use instead 0, 1, 2. -
add a small epsilon value to the prediction to prevent this divergence
-
gradient clipping
参考:
- https://github.com/wudashuo/yolov5/blob/master/models/tf.py