Package ai.djl.training.optimizer
Class Adam
java.lang.Object
ai.djl.training.optimizer.Optimizer
ai.djl.training.optimizer.Adam
Adam
is a generalization of the AdaGrad Optimizer
. \( grad += weight_decay * w\)
\( m = beta1 * m + (1 - beta1) * grad\)
\( v = beta2 * v + (1 - beta2) * grad^2 \)
\( learning_rate_bias_correction = learning_rate / beta1**t * sqrt(beta2**t) \)
\( w -= learning_rate_bias_correction * m / (sqrt(v) + epsilon) \)
where g represents the gradient, and m/v are 1st and 2nd order moment estimates (mean and variance), t is the step.
- See Also:
-
Nested Class Summary
Nested ClassesNested classes/interfaces inherited from class ai.djl.training.optimizer.Optimizer
Optimizer.OptimizerBuilder<T extends Optimizer.OptimizerBuilder>
-
Field Summary
Fields inherited from class ai.djl.training.optimizer.Optimizer
clipGrad, rescaleGrad
-
Constructor Summary
ConstructorsModifierConstructorDescriptionprotected
Adam
(Adam.Builder builder) Creates a new instance ofAdam
optimizer. -
Method Summary
Methods inherited from class ai.djl.training.optimizer.Optimizer
adadelta, adagrad, adam, adamW, getWeightDecay, nag, rmsprop, sgd, updateCount, withDefaultState
-
Constructor Details
-
Adam
Creates a new instance ofAdam
optimizer.- Parameters:
builder
- the builder to create a new instance ofAdam
optimizer
-
-
Method Details