mirror of https://github.com/hpcaitech/ColossalAI
[install] run with out rich (#513)
parent
0f2d219162
commit
81145208d1
|
@ -5,12 +5,16 @@ import colossalai
|
||||||
import logging
|
import logging
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Union
|
from typing import Union
|
||||||
from rich.logging import RichHandler
|
|
||||||
|
|
||||||
from colossalai.context.parallel_mode import ParallelMode
|
from colossalai.context.parallel_mode import ParallelMode
|
||||||
|
|
||||||
_FORMAT = 'colossalai - %(name)s - %(asctime)s %(levelname)s: %(message)s'
|
try:
|
||||||
logging.basicConfig(level=logging.INFO, format=_FORMAT, handlers=[RichHandler()])
|
from rich.logging import RichHandler
|
||||||
|
_FORMAT = 'colossalai - %(name)s - %(asctime)s %(levelname)s: %(message)s'
|
||||||
|
logging.basicConfig(level=logging.INFO, format=_FORMAT, handlers=[RichHandler()])
|
||||||
|
except ImportError:
|
||||||
|
_FORMAT = 'colossalai - %(name)s - %(asctime)s %(levelname)s: %(message)s'
|
||||||
|
logging.basicConfig(level=logging.INFO, format=_FORMAT)
|
||||||
|
|
||||||
|
|
||||||
class DistributedLogger:
|
class DistributedLogger:
|
||||||
|
|
Loading…
Reference in New Issue