mirror of https://github.com/hpcaitech/ColossalAI
[misc] fix dist logger (#5782)
parent
c46e09715c
commit
b9d646fe9e
|
@ -43,11 +43,11 @@ optim2DistOptim = {
|
|||
CAME: DistributedCAME,
|
||||
Adafactor: DistributedAdaFactor,
|
||||
}
|
||||
_logger = get_dist_logger()
|
||||
|
||||
|
||||
def cast_to_distributed(optim):
|
||||
if optim.__class__ in optim2DistOptim:
|
||||
_logger = get_dist_logger()
|
||||
_logger.info(f"Converting optimizer {optim.__class__.__name__} to its distributed version.", ranks=[0])
|
||||
|
||||
if isinstance(optim, GaLoreAdamW8bit):
|
||||
|
|
Loading…
Reference in New Issue