StochasticOptimizer type

Defines values for StochasticOptimizer.
KnownStochasticOptimizer can be used interchangeably with StochasticOptimizer, this enum contains the known values that the service supports.

Known values supported by the service

None: No optimizer selected.
Sgd: Stochastic Gradient Descent optimizer.
Adam: Adam is algorithm the optimizes stochastic objective functions based on adaptive estimates of moments
Adamw: AdamW is a variant of the optimizer Adam that has an improved implementation of weight decay.

type StochasticOptimizer = string