DenseNetCaffe
DenseNetCaffe copied to clipboard
The loss equaled to 87.3365 during the training stage and didn't change
I followed the instruction and didn't change the settings in solver.prototxt, but the loss converged to 87.3365 soon. It's said that this is because the learning rate is too large and the feature before the softmax layer equals to inf. So I am wondering what settings should I use with this network. Thanks a lot!
Maybe first try a lower learning rate. If it doesn't work, see if other network architectures don't work either. Then make the decision on whether to use this network, or see if there's some bugs elsewhere.
@cengzy14 , same problem, have you found any solutions?
@zhaofenqiang 出现87.3365的原因是softmax之前的特征层出现的inf或者nan 如果是第一次测试时出现87.3365,原因是BN层方差初始为0,而eps初始为1e-5,除以根号下 eps导致feature map中数值越来越大,出现inf,所以第一次测试时一定会出现87.6635 如果是训练时出现可能是pooling层的stride不能整除输入feature map的尺寸,导致出现nan 解决这两个问题后我的accuracy=0了,最后也没有解决,于是就换了https://github.com/shicai/DenseNet-Caffe 里面提供的在imagenet上训练的代码和model
@cengzy14 谢谢,很有帮助~
@cengzy14 请问你微调的时候网络收敛快么