18#ifndef itkAdaptiveStochasticVarianceReducedGradientOptimizer_h
19#define itkAdaptiveStochasticVarianceReducedGradientOptimizer_h
89 using Superclass::MeasureType;
90 using Superclass::ParametersType;
91 using Superclass::DerivativeType;
92 using Superclass::CostFunctionType;
99 itkSetMacro(UseAdaptiveStepSizes,
bool);
100 itkGetConstMacro(UseAdaptiveStepSizes,
bool);
104 itkSetMacro(SigmoidMax,
double);
105 itkGetConstMacro(SigmoidMax,
double);
109 itkSetMacro(SigmoidMin,
double);
110 itkGetConstMacro(SigmoidMin,
double);
114 itkSetMacro(SigmoidScale,
double);
115 itkGetConstMacro(SigmoidScale,
double);
This class implements a gradient descent optimizer with adaptive gain.
ITK_DISALLOW_COPY_AND_MOVE(AdaptiveStochasticVarianceReducedGradientOptimizer)
~AdaptiveStochasticVarianceReducedGradientOptimizer() override=default
DerivativeType m_PreviousGradient
bool m_UseAdaptiveStepSizes
void UpdateCurrentTime() override
SmartPointer< const Self > ConstPointer
AdaptiveStochasticVarianceReducedGradientOptimizer()
A cost function that applies a scaling to another cost function.
NonLinearOptimizer::ScalesType ScalesType
This class implements a gradient descent optimizer with a decaying gain.
StandardStochasticVarianceReducedGradientOptimizer()