mirror of https://github.com/hpcaitech/ColossalAI
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
48 lines
1.7 KiB
48 lines
1.7 KiB
from .colo_module import ColoModule
|
|
from colossalai.tensor import ComputePattern, distspec
|
|
from colossalai.core import global_context as gpc
|
|
from colossalai.context.parallel_mode import ParallelMode
|
|
|
|
|
|
class ColoLinear(ColoModule):
|
|
|
|
def __init__(self):
|
|
super(ColoLinear, self).__init__()
|
|
self._register_shard_params(['weight', 'bias'])
|
|
|
|
def register(self, compute_pattern):
|
|
if not compute_pattern in self._allowed_patterns:
|
|
if ComputePattern.TP1D == compute_pattern:
|
|
self._set_TP1D()
|
|
|
|
def _set_TP1D(self):
|
|
# TP1D Row Linear
|
|
_compute_pattern = ComputePattern.TP1D
|
|
self._register_allowed_patterns(
|
|
compute_pattern=_compute_pattern,
|
|
dist_specs={
|
|
'weight':
|
|
distspec.shard(gpc.get_group(ParallelMode.PARALLEL_1D), [-1],
|
|
[gpc.get_world_size(ParallelMode.PARALLEL_1D)]),
|
|
'bias':
|
|
None
|
|
},
|
|
mode='row',
|
|
)
|
|
|
|
# TP1D Col Linear
|
|
self._register_allowed_patterns(
|
|
compute_pattern=_compute_pattern,
|
|
dist_specs={
|
|
'weight':
|
|
distspec.shard(gpc.get_group(ParallelMode.PARALLEL_1D), [0],
|
|
[gpc.get_world_size(ParallelMode.PARALLEL_1D)]),
|
|
'bias':
|
|
distspec.shard(gpc.get_group(ParallelMode.PARALLEL_1D), [0],
|
|
[gpc.get_world_size(ParallelMode.PARALLEL_1D)])
|
|
},
|
|
mode='col',
|
|
)
|
|
|
|
self._set_default(compute_pattern=_compute_pattern, target_mode='row')
|