mirror of https://github.com/hpcaitech/ColossalAI
[misc] fix dist logger (#5782)
parent
c46e09715c
commit
b9d646fe9e
|
@ -43,11 +43,11 @@ optim2DistOptim = {
|
||||||
CAME: DistributedCAME,
|
CAME: DistributedCAME,
|
||||||
Adafactor: DistributedAdaFactor,
|
Adafactor: DistributedAdaFactor,
|
||||||
}
|
}
|
||||||
_logger = get_dist_logger()
|
|
||||||
|
|
||||||
|
|
||||||
def cast_to_distributed(optim):
|
def cast_to_distributed(optim):
|
||||||
if optim.__class__ in optim2DistOptim:
|
if optim.__class__ in optim2DistOptim:
|
||||||
|
_logger = get_dist_logger()
|
||||||
_logger.info(f"Converting optimizer {optim.__class__.__name__} to its distributed version.", ranks=[0])
|
_logger.info(f"Converting optimizer {optim.__class__.__name__} to its distributed version.", ranks=[0])
|
||||||
|
|
||||||
if isinstance(optim, GaLoreAdamW8bit):
|
if isinstance(optim, GaLoreAdamW8bit):
|
||||||
|
|
Loading…
Reference in New Issue