[Fundamentals of Machine Learning and Neural Networks]
When comparing and contrasting the ReLU and sigmoid activation functions, which statement is true?
ReLU (Rectified Linear Unit) and sigmoid are activation functions used in neural networks. According to NVIDIA's deep learning documentation (e.g., cuDNN and TensorRT), ReLU, defined as f(x) = max(0, x), is computationally efficient because it involves simple thresholding, avoiding expensive exponential calculations required by sigmoid, f(x) = 1/(1 + e^(-x)). Sigmoid outputs values in the range
[0, 1], making it suitable for predicting probabilities in binary classification tasks. ReLU, with an unbounded positive range, is less suited for direct probability prediction but accelerates training by mitigating vanishing gradient issues. Option A is incorrect, as ReLU is non-linear (piecewise linear). Option B is false, as ReLU is more efficient and not inherently more accurate. Option C is wrong, as ReLU's range is
[0, ), not
[0, 1].
NVIDIA cuDNN Documentation: https://docs.nvidia.com/deeplearning/cudnn/developer-guide/index.html
Goodfellow, I., et al. (2016). 'Deep Learning.' MIT Press.
Tequila
4 months agoSalina
4 months agoErin
4 months agoNelida
4 months agoCornell
5 months agoFlo
5 months agoGregoria
5 months agoDylan
5 months agoThurman
5 months agoClarinda
6 months agoGearldine
6 months agoBeckie
6 months agoCorinne
6 months agoJudy
8 months agoMollie
7 months agoCaprice
8 months agoJeff
9 months agoErick
9 months agoDean
9 months agoJosue
7 months agoRachael
8 months agoIzetta
8 months agoPeggie
8 months agoSilva
9 months agoAndra
9 months agoOnita
9 months agoDana
10 months agoRyan
10 months agoKatheryn
9 months agoWilliam
9 months agoMagda
10 months ago