18#ifndef itkStandardStochasticVarianceReducedGradientOptimizer_h
19#define itkStandardStochasticVarianceReducedGradientOptimizer_h
75 using Superclass::MeasureType;
76 using Superclass::ParametersType;
77 using Superclass::DerivativeType;
78 using Superclass::CostFunctionType;
85 itkSetMacro(Param_a,
double);
86 itkGetConstMacro(Param_a,
double);
89 itkSetMacro(Param_beta,
double);
90 itkGetConstMacro(Param_beta,
double);
93 itkSetMacro(Param_A,
double);
94 itkGetConstMacro(Param_A,
double);
97 itkSetMacro(Param_alpha,
double);
98 itkGetConstMacro(Param_alpha,
double);
114 itkSetMacro(InitialTime,
double);
115 itkGetConstMacro(InitialTime,
double);
119 itkGetConstMacro(CurrentTime,
double);
A cost function that applies a scaling to another cost function.
This class implements a gradient descent optimizer with a decaying gain.
void AdvanceOneStep() override
virtual double Compute_a(double k) const
ITK_DISALLOW_COPY_AND_MOVE(StandardStochasticVarianceReducedGradientOptimizer)
~StandardStochasticVarianceReducedGradientOptimizer() override=default
StandardStochasticVarianceReducedGradientOptimizer()
virtual void UpdateCurrentTime()
virtual double Compute_beta(double k) const
SmartPointer< const Self > ConstPointer
void StartOptimization() override
virtual void ResetCurrentTimeToInitialTime()
Implement a gradient descent optimizer.
NonLinearOptimizer::ScalesType ScalesType
StochasticVarianceReducedGradientDescentOptimizer()