List of questions
Related questions
Question 45 - H13-311_V3.5 discussion
Which of the following statements is false about gradient descent algorithms?
A.
Each time the global gradient updates its weight, all training samples need to be calculated.
B.
When GPUs are used for parallel computing, the mini-batch gradient descent (MBGD) takes less time than the stochastic gradient descent (SGD) to complete an epoch.
C.
The global gradient descent is relatively stable, which helps the model converge to the global extremum.
D.
When there are too many samples and GPUs are not used for parallel computing, the convergence process of the global gradient algorithm is time-consuming.
Your answer:
0 comments
Sorted by
Leave a comment first