mirror of https://github.com/hpcaitech/ColossalAI
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
15 lines
623 B
15 lines
623 B
3 years ago
|
from .fp16_optimizer import FP16Optimizer
|
||
|
from .fused_adam import FusedAdam
|
||
|
from .fused_lamb import FusedLAMB
|
||
|
from .fused_sgd import FusedSGD
|
||
|
from .lamb import Lamb
|
||
|
from .lars import Lars
|
||
|
from .zero_redundancy_optimizer_level_1 import ZeroRedundancyOptimizer_Level_1
|
||
|
from .zero_redundancy_optimizer_level_2 import ZeroRedundancyOptimizer_Level_2
|
||
|
from .zero_redundancy_optimizer_level_3 import ZeroRedundancyOptimizer_Level_3
|
||
|
|
||
|
__all__ = [
|
||
|
'ZeroRedundancyOptimizer_Level_1', 'ZeroRedundancyOptimizer_Level_2', 'ZeroRedundancyOptimizer_Level_3',
|
||
|
'FusedLAMB', 'FusedAdam', 'FusedSGD', 'Lamb', 'FP16Optimizer', 'Lars'
|
||
|
]
|