Package

org.platanios.tensorflow.api.ops.training.optimizers

schedules

Permalink

package schedules

Linear Supertypes
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. schedules
  2. AnyRef
  3. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Type Members

  1. class ComposedSchedule extends Schedule

    Permalink

    Scheduling method helper for composing two existing learning rate scheduling methods.

    Scheduling method helper for composing two existing learning rate scheduling methods.

    The resulting learning rate is the initial learning rate after having applied schedule2 on it, and then schedule1.

  2. class CosineDecay extends Schedule

    Permalink

    Cosine decay method.

    Cosine decay method.

    This method applies a cosine decay function to a provided initial learning rate (i.e., value). It requires a step value to be provided in it's application function, in order to compute the decayed learning rate. You may simply pass a TensorFlow variable that you increment at each training step.

    The decayed value is computed as follows:

    cosineDecay = 0.5 * (1 + cos(pi * min(step, cycleSteps) / cycleSteps))
    decayed = value * ((1 - alpha) * cosineDecay + alpha)
  3. class CycleLinear10xDecay extends Schedule

    Permalink

    Cycle-linear 10x decay method.

    Cycle-linear 10x decay method.

    This method applies a cycle-linear decay function to a provided initial learning rate (i.e., value). It requires a step value to be provided in it's application function, in order to compute the decayed learning rate. You may simply pass a TensorFlow variable that you increment at each training step.

    The decayed value is computed as follows:

    cyclePosition = 1 - abs(((step % (2 * cycleSteps)) - cycleSteps) / cycleSteps)
    decayed = value * (0.1 + cyclePosition) * 3
  4. class ExponentialDecay extends Schedule

    Permalink

    Exponential decay method.

    Exponential decay method.

    This method applies an exponential decay function to a provided initial learning rate (i.e., value). It requires a step value to be provided in it's application function, in order to compute the decayed learning rate. You may simply pass a TensorFlow variable that you increment at each training step.

    The decayed value is computed as follows:

    decayed = value * decayRate ^ (step / decaySteps)

    where if staircase = true, then (step / decaySteps) is an integer division and the decayed learning rate follows a staircase function.

  5. class LuongExponentialDecay extends ExponentialDecay

    Permalink

    A particular instance of ExponentialDecay that was used in [Luong (2016)](https://github.com/lmthang/thesis).

  6. trait Schedule extends AnyRef

    Permalink

    Trait for implementing optimization learning rate scheduling methods.

    Trait for implementing optimization learning rate scheduling methods.

    When training a model, it is often recommended to lower the learning rate as the training progresses. Scheduling methods can be used for that purpose. They define ways in which to schedule the learning rate as training progresses.

  7. class SqrtDecay extends Schedule

    Permalink

    Square root decay method.

    Square root decay method.

    This method applies a square root decay function to a provided initial learning rate (i.e., value). It requires a step value to be provided in it's application function, in order to compute the decayed learning rate. You may simply pass a TensorFlow variable that you increment at each training step.

    The decayed value is computed as follows:

    decayed = value * decayFactor / sqrt(max(step, decayThreshold))
  8. class WarmUpExponentialSchedule extends Schedule

    Permalink

    Learning rate schedule that implements a warm-up scheme, similar to the one proposed in [Attention is All You Need (Section 5.3)](https://arxiv.org/pdf/1706.03762.pdf).

    Learning rate schedule that implements a warm-up scheme, similar to the one proposed in [Attention is All You Need (Section 5.3)](https://arxiv.org/pdf/1706.03762.pdf).

    For the first warmUpSteps steps the learning rate is multiplied by: exp(log(warmUpFactor) / step) ^ (warmUpSteps - step).

  9. class WarmUpLinearSchedule extends Schedule

    Permalink

    Learning rate schedule that implements a warm-up scheme, similar to the one proposed in [Attention is All You Need (Section 5.3)](https://arxiv.org/pdf/1706.03762.pdf).

    Learning rate schedule that implements a warm-up scheme, similar to the one proposed in [Attention is All You Need (Section 5.3)](https://arxiv.org/pdf/1706.03762.pdf).

    For the first warmUpSteps steps the learning rate is multiplied by: start + ((1.0f - start) / warmUpSteps) * step.

Value Members

  1. object ComposedSchedule

    Permalink
  2. object CosineDecay

    Permalink
  3. object CycleLinear10xDecay

    Permalink
  4. object ExponentialDecay

    Permalink
  5. object FixedSchedule extends Schedule with Product with Serializable

    Permalink

    Dummy scheduling method representing no schedule being used.

    Dummy scheduling method representing no schedule being used. Useful as a default value for Schedule-valued function arguments.

  6. object LuongExponentialDecay

    Permalink
  7. object SqrtDecay

    Permalink
  8. object WarmUpExponentialSchedule

    Permalink
  9. object WarmUpLinearSchedule

    Permalink

Inherited from AnyRef

Inherited from Any

Ungrouped