Loss jumps to 87.3365
The id_loss = 87.3365 usually means the softmax loss is computing -log(FLT_MIN) here, which implies the probability is smaller than FLT_MIN.However, this probability is computed as softmax(Wx), where W is initialized as zero matrix in the first iteration.
As I could not reproduce this situation on my machine, could you please kindly do me a favor by adding the following print function in caffe/src/caffe/layers/softmax_loss_layer.cu
as well as adding debug_print(prob_);
after this line.Then recompile the caffe, rerun the experiments, and check the output before the Iteration 0. Thank you very much in advance.
Last updated