I have an input image with greyscale values ranging from let's say 25000 - 35000. I'm doing binary pixel-wise classification, so the output ground truth is a matrix of either 0's or 1's.
Does anyone know what the default output activation function is? My question is, is it a ReLu? I want it to be a SoftMax function. In which case, each prediction value would be between 0 and 1(obviously close to my ground truth data).
I'm using example code from here that I have adjusted to make work for my data.
I have a working network that is training, but the minibatch loss is at about 425 right now and the accuracy at 0.0, and for the LSTM MNIST example code (linked) the minibatch loss was about 0.1 and the accuracy about 1.0. I hope that if I cahttps://smallseotools.com/asets/images/cleartext.svgn change the activation function to use the SoftMax function, I can improve results.