Question 5 (Intermediate):
In a neural network, what does the ReLU activation function return?
A) 1 / (1 + e^-x)
B) max(0, x)
C) x^2
D) e^x / (e^x + 1)
#NeuralNetworks #DeepLearning #ActivationFunctions #ReLU #AI
In a neural network, what does the ReLU activation function return?
A) 1 / (1 + e^-x)
B) max(0, x)
C) x^2
D) e^x / (e^x + 1)
#NeuralNetworks #DeepLearning #ActivationFunctions #ReLU #AI
❤1
Question 30 (Intermediate - PyTorch):
What is the purpose of
A) Disables model training
B) Speeds up computations by disabling gradient tracking
C) Forces GPU memory cleanup
D) Enables distributed training
#Python #PyTorch #DeepLearning #NeuralNetworks
✅ By: https://yangx.top/DataScienceQ
What is the purpose of
torch.no_grad()
context manager in PyTorch? A) Disables model training
B) Speeds up computations by disabling gradient tracking
C) Forces GPU memory cleanup
D) Enables distributed training
#Python #PyTorch #DeepLearning #NeuralNetworks
✅ By: https://yangx.top/DataScienceQ
🔥1