I have a very basic question on the optimization algotithm, when I'm adjusting weights and biases in a NN, should I:
- Forward propagate and backpropagate to calculate gradient descent (DC) for each batch once and then repeat for
iterations_number
times.
or
- Forward propagate and backpropagate to calculate gradient descent (DC) one batch for
iterations_number
times and then continue with the next batch.
epochs_number
instead ofiterations_number
as the second is just $\frac{n}{batch size}$. – Oliver Mohr Bonometti Jan 16 '21 at 18:13iterations_number = (n / batchsize) * epochs_number
. – Jonathan Jan 17 '21 at 08:49