OptimizerState#
- class vulkpy.nn.OptimizerState#
Bases:
objectAbstract base class for Optimizer State
See also
vulkpy.nn.OptimizerOptimizer
vulkpy.nn.SGDStateOptimizerState subclass for SGD
vulkpy.nn.AdamStateOptimizerState subclass for Adam
Notes
Mutable per-parameter values are stored at this class instance, although static global parameters (e.g. learning rate) are stored at
Optimizerclass.Subclass of
OptimizerStateshould implementOptimizer.grad2diff(), which takes accumulated gradients and returns update difference.In standard design,
OptimizerStateholds a reference to its parentOptimizerin order to access global parameters.Methods Summary
grad2diff(grad)Compute update diff from gradient
Methods Documentation
- grad2diff(grad: Array) Array#
Compute update diff from gradient
- Parameters:
grad (vulkpy.Array) – Accumulated gradient
- Returns:
diff – Update diff. (
v += opt_state.grad2diff(grad))- Return type:
Notes
Subclass must implement this method.
- __init__()#