It is the default learning rate schedule.
It is the default learning rate schedule. For each iteration, the learning rate would update with the following formula:
l_{n + 1} = l / (1 + n * learning_rate_decay)
where l
is the initial learning rate
It is an epoch decay learning rate schedule The learning rate decays through a function argument on number of run epochs
It is an epoch decay learning rate schedule The learning rate decays through a function argument on number of run epochs
l_{n + 1} = l_{n} * 0.1 ^
decayType(epoch)
is a function with number of run epochs as the argument
EpochSchedule is a learning rate schedule which configure the learning rate according to some pre-defined Regime.
EpochSchedule is a learning rate schedule which configure the learning
rate according to some pre-defined Regime. If the running epoch is within
the interval of a regime r
[r.startEpoch, r.endEpoch], then the learning
rate will take the "learningRate" in r.config.
an array of pre-defined Regime.
EpochStep is a learning rate schedule, which rescale the learning rate by gamma
for each stepSize
epochs.
EpochStep is a learning rate schedule, which rescale the learning rate by gamma
for each stepSize
epochs.
For how many epochs to update the learning rate once
the rescale factor
Learning rate schedule for SGD
A learning rate decay policy, where the effective learning rate follows a polynomial decay, to be zero by the max_iteration.
A learning rate decay policy, where the effective learning rate follows a polynomial decay, to be zero by the max_iteration. Calculation: base_lr (1 - iter/maxIteration) ^ (power)
coeffient of decay, refer to calculation formula
max iteration when lr becomes zero
A structure to specify hyper parameters by start epoch and end epoch.
A structure to specify hyper parameters by start epoch and end epoch. Usually work with EpochSchedule.
start epoch
end epoch
config table contains hyper parameters
A learning rate decay policy, where the effective learning rate is calculated as base_lr * gamma ^ (floor(iter / stepSize))
A learning rate decay policy, where the effective learning rate is calculated as base_lr * gamma ^ (floor(iter / stepSize))
the inteval for lr decay
coefficient of decay, refer to calculation formula