stochasticgradient
Stochastic gradient refers to an estimate of the gradient of a loss function computed from a random subset of data rather than the full dataset. It is a key concept in stochastic gradient descent and related optimization methods, where computing the exact gradient over all training examples can be prohibitively expensive for large-scale problems.
In practice, an iteration updates model parameters by moving in the direction opposite to the estimated gradient.
Mini-batch variants balance computational efficiency and gradient variance: smaller batches yield noisier updates but faster iterations,
Convergence properties depend on the problem class and learning-rate strategy. For convex objectives, stochastic gradient methods
Variants and related methods include momentum-based updates and adaptive learning-rate schemes (such as RMSprop, Adagrad, and