True or false 1) The learning rate has to stay the same throughout the learning algorithm. O True O False 2) The stochas
-
- Site Admin
- Posts: 899603
- Joined: Mon Aug 02, 2021 8:13 am
True or false 1) The learning rate has to stay the same throughout the learning algorithm. O True O False 2) The stochas
True or false 1) The learning rate has to stay the samethroughout the learning algorithm. O True O False 2) The stochasticgradient descent algorithm performs fewer weight updates per epochcomparing to the gradient descent algorithm. O True O False 3)Random Forests are an instance of bagging. O True O False 4) Theperceptron learning algorithm requires that the training data arelinearly separable. Otherwise, the model does not converge. O TrueO False 5)The learning rate is a hyperparameter of the logisticmodel. O True O False 6) The weight update rule in gradient descentmethods depends on the definition of the cost (error) function. OTrue O False 7)Dimensionality reduction with PC computes newattributes that are linear transformations of the originalattributes. O True O False 8) Stochastic gradient descent isguaranteed to always find the best possible model fit. O True OFalse 9)Shuffling the data generally improves the performance ofnon-stochastic gradient descent. O True O False 10)The weightupdate rule in gradient descent methods depends on the definitionof the cost (error) functions T or F