Package smile.deep.optimizer
Class Adam
java.lang.Object
smile.deep.optimizer.Adam
- All Implemented Interfaces:
Serializable
,Optimizer
Adaptive Moment optimizer. Adam computes adaptive learning rates for
each parameter. In addition to storing an exponentially decaying average
of past squared gradients RMSProp, Adam also keeps an exponentially
decaying average of past gradients, similar to momentum. Whereas momentum
can be seen as a ball running down a slope, Adam behaves like a heavy ball
with friction, which thus prefers flat minima in the error surface.
- See Also:
-
Constructor Summary
Constructors -
Method Summary
-
Constructor Details
-
Adam
public Adam()Constructor. -
Adam
public Adam(smile.math.TimeFunction learningRate) Constructor.- Parameters:
learningRate
- the learning rate.
-
Adam
public Adam(smile.math.TimeFunction learningRate, double beta1, double beta2) Constructor.- Parameters:
learningRate
- the learning rate.beta1
- the exponential decay rate for the 1st moment estimates.beta2
- the exponential decay rate for the 2nd moment estimates.
-
Adam
public Adam(smile.math.TimeFunction learningRate, double beta1, double beta2, double epsilon) Constructor.- Parameters:
learningRate
- the learning rate.beta1
- the exponential decay rate for the 1st moment estimates.beta2
- the exponential decay rate for the 2nd moment estimates.epsilon
- a small constant for numerical stability.
-
-
Method Details