I have a very basic question on the optimization algotithm, when I'm adjusting weights and biases in a NN, should I:
- Forward propagate and backpropagate to calculate gradient descent (DC) for each batch once and then repeat for
iterations_numbertimes.
or
- Forward propagate and backpropagate to calculate gradient descent (DC) one batch for
iterations_numbertimes and then continue with the next batch.
epochs_numberinstead ofiterations_numberas the second is just $\frac{n}{batch size}$. – Oliver Mohr Bonometti Jan 16 '21 at 18:13iterations_number = (n / batchsize) * epochs_number. – Jonathan Jan 17 '21 at 08:49