N
N
numitus22016-06-21 22:46:38
Neural networks
numitus2, 2016-06-21 22:46:38

How is backpropagation different from stochastic gradient descent?

As far as I understood from the literature, they differ only in that the backpropagation of the error changes after processing the entire sample, and the stochastic descent after each image?

Answer the question

In order to leave comments, you need to log in

1 answer(s)
I
ivodopyanov, 2016-06-28
@ivodopyanov

In fact, they are slightly different concepts.
Backpropagation gives us a function to optimize ( = find min\max).
SGD finds the minimum/maximum of some arbitrary function.
In addition to stochastic descent, there is also full-batch gradient descent (this is the one, by default, when we change the coefficients after calculating the entire sample), mini-batch gradient descent (an intermediate option, when we change the coefficients after calculating N sample elements), rmsprop, ADAM , simulated annealing, etc.

Didn't find what you were looking for?

Ask your question

Ask a Question

731 491 924 answers to any question