[Fundamentals of Machine Learning and Neural Networks]
When comparing and contrasting the ReLU and sigmoid activation functions, which statement is true?
ReLU (Rectified Linear Unit) and sigmoid are activation functions used in neural networks. According to NVIDIA's deep learning documentation (e.g., cuDNN and TensorRT), ReLU, defined as f(x) = max(0, x), is computationally efficient because it involves simple thresholding, avoiding expensive exponential calculations required by sigmoid, f(x) = 1/(1 + e^(-x)). Sigmoid outputs values in the range
[0, 1], making it suitable for predicting probabilities in binary classification tasks. ReLU, with an unbounded positive range, is less suited for direct probability prediction but accelerates training by mitigating vanishing gradient issues. Option A is incorrect, as ReLU is non-linear (piecewise linear). Option B is false, as ReLU is more efficient and not inherently more accurate. Option C is wrong, as ReLU's range is
[0, ), not
[0, 1].
NVIDIA cuDNN Documentation: https://docs.nvidia.com/deeplearning/cudnn/developer-guide/index.html
Goodfellow, I., et al. (2016). 'Deep Learning.' MIT Press.
Judy
15 days agoJeff
25 days agoErick
29 days agoDean
1 months agoIzetta
14 days agoPeggie
15 days agoSilva
25 days agoAndra
1 months agoOnita
1 months agoDana
2 months agoRyan
2 months agoKatheryn
1 months agoWilliam
1 months agoMagda
2 months ago