How to chose between ReLU and sigmoid activation for CNN output
I’m working on training a CNN model that takes an eye image as input and outputs the 5 coordinates of the ellipse representing the pupil (center x, center y, major axis, minor axis, angle). I started by training a small CNN with a linear activation function on the output layer. This worked, but the model was producing negative values, which while valid coordinates, ideally I’d prefer to have all positive outputs.