Optimizers and schedulers supported by CAREamics.
SupportedOptimizer
Bases: str
, BaseEnum
Supported optimizers.
Attributes:
Name | Type | Description |
Adam | str | |
SGD | str | Stochastic Gradient Descent optimizer. |
Source code in src/careamics/config/support/supported_optimizers.py
| class SupportedOptimizer(str, BaseEnum):
"""Supported optimizers.
Attributes
----------
Adam : str
Adam optimizer.
SGD : str
Stochastic Gradient Descent optimizer.
"""
# ASGD = "ASGD"
# Adadelta = "Adadelta"
# Adagrad = "Adagrad"
ADAM = "Adam"
# AdamW = "AdamW"
ADAMAX = "Adamax"
# LBFGS = "LBFGS"
# NAdam = "NAdam"
# RAdam = "RAdam"
# RMSprop = "RMSprop"
# Rprop = "Rprop"
SGD = "SGD"
|
SupportedScheduler
Bases: str
, BaseEnum
Supported schedulers.
Attributes:
Name | Type | Description |
ReduceLROnPlateau | str | Reduce learning rate on plateau. |
StepLR | str | |
Source code in src/careamics/config/support/supported_optimizers.py
| class SupportedScheduler(str, BaseEnum):
"""Supported schedulers.
Attributes
----------
ReduceLROnPlateau : str
Reduce learning rate on plateau.
StepLR : str
Step learning rate.
"""
# ChainedScheduler = "ChainedScheduler"
# ConstantLR = "ConstantLR"
# CosineAnnealingLR = "CosineAnnealingLR"
# CosineAnnealingWarmRestarts = "CosineAnnealingWarmRestarts"
# CyclicLR = "CyclicLR"
# ExponentialLR = "ExponentialLR"
# LambdaLR = "LambdaLR"
# LinearLR = "LinearLR"
# MultiStepLR = "MultiStepLR"
# MultiplicativeLR = "MultiplicativeLR"
# OneCycleLR = "OneCycleLR"
# PolynomialLR = "PolynomialLR"
REDUCE_LR_ON_PLATEAU = "ReduceLROnPlateau"
# SequentialLR = "SequentialLR"
STEP_LR = "StepLR"
|