You are doing full batch gradient descent using the entire training set (not stochastic gradient descent). Is it necessa
Posted: Fri Jul 01, 2022 5:39 am
You are doing full batch gradient descent using the entire training set (not stochastic gradient descent). Is it necessary to shuffle the training data? Justify your answer.