deepr.optimizers package
Submodules
deepr.optimizers.base module
Interface for Optimizers
deepr.optimizers.core module
Tensorflow Optimizers
- class deepr.optimizers.core.TensorflowOptimizer(optimizer, learning_rate, loss='loss', grad_norms=None, exclude_vars=None, clip=None, skip_vars=None, skip_steps=None, **kwargs)[source]
Bases:
Optimizer
Default Tensorflow Optimizers
- optimizer
Name of the optimizer. See TensorflowOptimizer.OPTIMIZERS for a description of available Tensorflow optimizers.
- Type:
- kwargs
Optional arguments for the Tensorflow optimizer.
- OPTIMIZERS = {'adagrad': <class 'tensorflow.python.training.adagrad.AdagradOptimizer'>, 'adam': <class 'tensorflow.python.training.adam.AdamOptimizer'>, 'lazyadam': <class 'tensorflow.contrib.opt.python.training.lazy_adam_optimizer.LazyAdamOptimizer'>, 'momentum': <class 'tensorflow.python.training.momentum.MomentumOptimizer'>, 'sgd': <class 'tensorflow.python.training.gradient_descent.GradientDescentOptimizer'>}