#stochastic_gradient_descent
Stochastic gradient descent
Optimization algorithm
Stochastic gradient descent is an iterative method for optimizing an objective function with suitable smoothness properties. It can be regarded as a stochastic approximation of gradient descent optimization, since it replaces the actual gradient by an estimate thereof. Especially in high-dimensional optimization problems this reduces the very high computational burden, achieving faster iterations in exchange for a lower convergence rate.
Mon 9th
Provided by Wikipedia
This keyword could refer to multiple things. Here are some suggestions:
0 searches
This keyword has never been searched before
This keyword has never been searched for with any other keyword.