mirror of https://github.com/hpcaitech/ColossalAI
[pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.cicolossalchat
parent
46037c2ccd
commit
52d346f2a5
|
@ -13,7 +13,6 @@ from transformers.models.mixtral.modeling_mixtral import MixtralForCausalLM
|
||||||
import colossalai
|
import colossalai
|
||||||
from colossalai.booster import Booster
|
from colossalai.booster import Booster
|
||||||
from colossalai.booster.plugin.moe_hybrid_parallel_plugin import MoeHybridParallelPlugin
|
from colossalai.booster.plugin.moe_hybrid_parallel_plugin import MoeHybridParallelPlugin
|
||||||
from colossalai.tensor.moe_tensor.api import is_moe_tensor
|
|
||||||
from colossalai.testing import parameterize, spawn
|
from colossalai.testing import parameterize, spawn
|
||||||
from colossalai.testing.utils import spawn
|
from colossalai.testing.utils import spawn
|
||||||
from tests.test_moe.moe_utils import loose_close
|
from tests.test_moe.moe_utils import loose_close
|
||||||
|
|
Loading…
Reference in New Issue