mirror of https://github.com/hpcaitech/ColossalAI
[feat] update mixtral policy & bert policy for zerobubble
parent
80b04d7855
commit
b6d5e61809
|
@ -75,6 +75,8 @@ class BertPolicy(Policy):
|
||||||
|
|
||||||
sp_partial_derived = sp_mode == "split_gather"
|
sp_partial_derived = sp_mode == "split_gather"
|
||||||
|
|
||||||
|
use_zbv = self.pipeline_stage_manager is not None and self.pipeline_stage_manager.use_zbv
|
||||||
|
|
||||||
if self.shard_config.enable_tensor_parallelism:
|
if self.shard_config.enable_tensor_parallelism:
|
||||||
assert (
|
assert (
|
||||||
self.model.config.num_attention_heads % self.shard_config.tensor_parallel_size == 0
|
self.model.config.num_attention_heads % self.shard_config.tensor_parallel_size == 0
|
||||||
|
@ -97,6 +99,7 @@ class BertPolicy(Policy):
|
||||||
kwargs={
|
kwargs={
|
||||||
"seq_parallel_mode": sp_mode,
|
"seq_parallel_mode": sp_mode,
|
||||||
"fp8_communication": self.shard_config.fp8_communication,
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
},
|
},
|
||||||
),
|
),
|
||||||
SubModuleReplacementDescription(
|
SubModuleReplacementDescription(
|
||||||
|
@ -105,6 +108,7 @@ class BertPolicy(Policy):
|
||||||
kwargs={
|
kwargs={
|
||||||
"seq_parallel_mode": sp_mode,
|
"seq_parallel_mode": sp_mode,
|
||||||
"fp8_communication": self.shard_config.fp8_communication,
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
},
|
},
|
||||||
),
|
),
|
||||||
SubModuleReplacementDescription(
|
SubModuleReplacementDescription(
|
||||||
|
@ -113,6 +117,7 @@ class BertPolicy(Policy):
|
||||||
kwargs={
|
kwargs={
|
||||||
"seq_parallel_mode": sp_mode,
|
"seq_parallel_mode": sp_mode,
|
||||||
"fp8_communication": self.shard_config.fp8_communication,
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
},
|
},
|
||||||
),
|
),
|
||||||
SubModuleReplacementDescription(
|
SubModuleReplacementDescription(
|
||||||
|
@ -125,6 +130,7 @@ class BertPolicy(Policy):
|
||||||
kwargs={
|
kwargs={
|
||||||
"seq_parallel_mode": sp_mode,
|
"seq_parallel_mode": sp_mode,
|
||||||
"fp8_communication": self.shard_config.fp8_communication,
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
},
|
},
|
||||||
),
|
),
|
||||||
SubModuleReplacementDescription(
|
SubModuleReplacementDescription(
|
||||||
|
@ -138,6 +144,7 @@ class BertPolicy(Policy):
|
||||||
"seq_parallel_mode": sp_mode,
|
"seq_parallel_mode": sp_mode,
|
||||||
"skip_bias_add": self.enable_bias_gelu_fused,
|
"skip_bias_add": self.enable_bias_gelu_fused,
|
||||||
"fp8_communication": self.shard_config.fp8_communication,
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
},
|
},
|
||||||
),
|
),
|
||||||
SubModuleReplacementDescription(
|
SubModuleReplacementDescription(
|
||||||
|
@ -146,6 +153,97 @@ class BertPolicy(Policy):
|
||||||
kwargs={
|
kwargs={
|
||||||
"seq_parallel_mode": sp_mode,
|
"seq_parallel_mode": sp_mode,
|
||||||
"fp8_communication": self.shard_config.fp8_communication,
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="output.dropout",
|
||||||
|
target_module=col_nn.DropoutForParallelInput,
|
||||||
|
),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
|
||||||
|
policy[BertEmbeddings] = ModulePolicyDescription(
|
||||||
|
sub_module_replacement=[
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="dropout",
|
||||||
|
target_module=col_nn.DropoutForReplicatedInput,
|
||||||
|
),
|
||||||
|
]
|
||||||
|
)
|
||||||
|
if self.enable_bias_gelu_fused:
|
||||||
|
self.append_or_create_method_replacement(
|
||||||
|
description={
|
||||||
|
"forward": get_jit_fused_bert_intermediate_forward(),
|
||||||
|
},
|
||||||
|
policy=policy,
|
||||||
|
target_key=BertIntermediate,
|
||||||
|
)
|
||||||
|
|
||||||
|
elif use_zbv:
|
||||||
|
policy[BertLayer] = ModulePolicyDescription(
|
||||||
|
sub_module_replacement=[
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="attention.self.query",
|
||||||
|
target_module=col_nn.LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"seq_parallel_mode": sp_mode,
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="attention.self.key",
|
||||||
|
target_module=col_nn.LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"seq_parallel_mode": sp_mode,
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="attention.self.value",
|
||||||
|
target_module=col_nn.LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"seq_parallel_mode": sp_mode,
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="attention.self.dropout",
|
||||||
|
target_module=col_nn.DropoutForParallelInput,
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="attention.output.dense",
|
||||||
|
target_module=col_nn.LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"seq_parallel_mode": sp_mode,
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="attention.output.dropout",
|
||||||
|
target_module=col_nn.DropoutForParallelInput,
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="intermediate.dense",
|
||||||
|
target_module=col_nn.LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"seq_parallel_mode": sp_mode,
|
||||||
|
"skip_bias_add": self.enable_bias_gelu_fused,
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="output.dense",
|
||||||
|
target_module=col_nn.LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"seq_parallel_mode": sp_mode,
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
},
|
},
|
||||||
),
|
),
|
||||||
SubModuleReplacementDescription(
|
SubModuleReplacementDescription(
|
||||||
|
|
|
@ -7,9 +7,18 @@ from torch import Tensor
|
||||||
from torch.nn import Module
|
from torch.nn import Module
|
||||||
from transformers.models.mixtral.modeling_mixtral import MixtralForCausalLM, MixtralModel
|
from transformers.models.mixtral.modeling_mixtral import MixtralForCausalLM, MixtralModel
|
||||||
|
|
||||||
from colossalai.shardformer.layer import FusedRMSNorm, Linear1D_Col
|
from colossalai.shardformer.layer import (
|
||||||
from colossalai.shardformer.layer.embedding import PaddingEmbedding, VocabParallelEmbedding1D
|
FusedRMSNorm,
|
||||||
from colossalai.shardformer.layer.linear import Linear1D_Row
|
Linear1D_Col,
|
||||||
|
Linear1D_Row,
|
||||||
|
LinearWithGradAccum,
|
||||||
|
PaddingEmbedding,
|
||||||
|
VocabParallelEmbedding1D,
|
||||||
|
)
|
||||||
|
|
||||||
|
# from colossalai.shardformer.layer import FusedRMSNorm, Linear1D_Col
|
||||||
|
# from colossalai.shardformer.layer.embedding import PaddingEmbedding, VocabParallelEmbedding1D
|
||||||
|
# from colossalai.shardformer.layer.linear import Linear1D_Row
|
||||||
from colossalai.shardformer.modeling.mixtral import (
|
from colossalai.shardformer.modeling.mixtral import (
|
||||||
EPMixtralSparseMoeBlock,
|
EPMixtralSparseMoeBlock,
|
||||||
MixtralPipelineForwards,
|
MixtralPipelineForwards,
|
||||||
|
@ -166,6 +175,52 @@ class MixtralPolicy(Policy):
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
elif use_zbv:
|
||||||
|
policy[MixtralDecoderLayer] = ModulePolicyDescription(
|
||||||
|
sub_module_replacement=[
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="self_attn.q_proj",
|
||||||
|
target_module=LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="self_attn.k_proj",
|
||||||
|
target_module=LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="self_attn.v_proj",
|
||||||
|
target_module=LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="self_attn.o_proj",
|
||||||
|
target_module=LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="block_sparse_moe.gate",
|
||||||
|
target_module=LinearWithGradAccum,
|
||||||
|
kwargs={
|
||||||
|
"gather_output": True,
|
||||||
|
"fp8_communication": self.shard_config.fp8_communication,
|
||||||
|
"use_zbv": use_zbv,
|
||||||
|
},
|
||||||
|
),
|
||||||
|
],
|
||||||
|
)
|
||||||
if embedding_cls is not None:
|
if embedding_cls is not None:
|
||||||
self.append_or_create_submodule_replacement(
|
self.append_or_create_submodule_replacement(
|
||||||
description=SubModuleReplacementDescription(
|
description=SubModuleReplacementDescription(
|
||||||
|
@ -351,6 +406,23 @@ class MixtralForCausalLMPolicy(MixtralPolicy):
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
policy.update(new_item)
|
policy.update(new_item)
|
||||||
|
elif use_zbv:
|
||||||
|
new_item = {
|
||||||
|
MixtralForCausalLM: ModulePolicyDescription(
|
||||||
|
sub_module_replacement=[
|
||||||
|
SubModuleReplacementDescription(
|
||||||
|
suffix="lm_head",
|
||||||
|
target_module=LinearWithGradAccum,
|
||||||
|
kwargs=dict(
|
||||||
|
gather_output=True,
|
||||||
|
fp8_communication=self.shard_config.fp8_communication,
|
||||||
|
use_zbv=use_zbv,
|
||||||
|
),
|
||||||
|
)
|
||||||
|
],
|
||||||
|
)
|
||||||
|
}
|
||||||
|
policy.update(new_item)
|
||||||
|
|
||||||
if self.pipeline_stage_manager:
|
if self.pipeline_stage_manager:
|
||||||
# set None as default
|
# set None as default
|
||||||
|
|
Loading…
Reference in New Issue