Which of the following activation functions outputs values strictly between 0 and 1 for any real input?
What is the mathematical expression for the ReLU activation function?
If you input large negative values into the Tanh activation function, what is the approximate output?
Which activation function among ReLU, Sigmoid, and Tanh is non-linear but does not squash negative values to positive outputs?
For which input does the derivative of the Sigmoid activation function reach its maximum value?
If you want your activation function's output to cover both negative and positive ranges symmetrically, which should you use?
What value does ReLU return if the input is zero?
Which activation function is prone to the vanishing gradient problem because its output saturates for large positive or negative inputs?
If you need an output interpreted as a probability, which activation function is most suitable at an output layer?
When using the ReLU activation function, what is the output for an input value of -5?