Adam

class boulderopal.stochastic.Adam(learning_rate=0.01)

Adaptive moment estimation (Adam) optimizer for stochastic optimization.

For more detail on Adam see Adam on Wikipedia.

Parameters:

learning_rate (float, optional) – The learning rate for the Adam optimizer. If set, must be positive. Defaults to 0.01.

See also

boulderopal.run_stochastic_optimization

Perform gradient-based stochastic optimization of generic real-valued functions.