There's various activation functions: sigmoid, tanh, etc. And there's also a few initializer functions:
Nguyen and Widrow, random, normalized, constant, zero, etc. So do these have much effect on the outcome of a neural network specialising in face detection? Right now I'm using the Tanh activation function and just randomising all
…