19#ifndef itkAdaptiveStochasticGradientDescentOptimizer_h
20#define itkAdaptiveStochasticGradientDescentOptimizer_h
22#include "../StandardGradientDescent/itkStandardGradientDescentOptimizer.h"
91 using Superclass::MeasureType;
92 using Superclass::ParametersType;
93 using Superclass::DerivativeType;
94 using Superclass::CostFunctionType;
101 itkSetMacro(UseAdaptiveStepSizes,
bool);
102 itkGetConstMacro(UseAdaptiveStepSizes,
bool);
106 itkSetMacro(SigmoidMax,
double);
107 itkGetConstMacro(SigmoidMax,
double);
111 itkSetMacro(SigmoidMin,
double);
112 itkGetConstMacro(SigmoidMin,
double);
116 itkSetMacro(SigmoidScale,
double);
117 itkGetConstMacro(SigmoidScale,
double);
This class implements a gradient descent optimizer with adaptive gain.
~AdaptiveStochasticGradientDescentOptimizer() override=default
DerivativeType m_PreviousGradient
ITK_DISALLOW_COPY_AND_MOVE(AdaptiveStochasticGradientDescentOptimizer)
bool m_UseAdaptiveStepSizes
void UpdateCurrentTime() override
SmartPointer< const Self > ConstPointer
AdaptiveStochasticGradientDescentOptimizer()
A cost function that applies a scaling to another cost function.
This class implements a gradient descent optimizer with a decaying gain.
StandardGradientDescentOptimizer()
NonLinearOptimizer::ScalesType ScalesType