Package ai.djl.training.optimizer
Class Sgd
java.lang.Object
ai.djl.training.optimizer.Optimizer
ai.djl.training.optimizer.Sgd
Sgd
is a Stochastic Gradient Descent (SGD) optimizer.
If momentum is not set, it updates weights using the following update function:
\( weight = weight - learning_rate * (gradient + wd * weight) \).
If momentum is set, it updates weights using the following update function:
\( state = momentum * state + learning_rate * gradient \)
\( weight -= state \)
Momentum update has better convergence rates on neural networks.
- See Also:
-
Nested Class Summary
Nested ClassesNested classes/interfaces inherited from class ai.djl.training.optimizer.Optimizer
Optimizer.OptimizerBuilder<T extends Optimizer.OptimizerBuilder>
-
Field Summary
Fields inherited from class ai.djl.training.optimizer.Optimizer
clipGrad, rescaleGrad
-
Constructor Summary
Constructors -
Method Summary
Methods inherited from class ai.djl.training.optimizer.Optimizer
adadelta, adagrad, adam, adamW, getWeightDecay, nag, rmsprop, sgd, updateCount, withDefaultState
-
Constructor Details
-
Sgd
Creates a new instance ofSgd
.- Parameters:
builder
- the builder to create a new instance ofSgd
-
-
Method Details