mirror of https://github.com/hpcaitech/ColossalAI
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
20 lines
751 B
20 lines
751 B
from torch import Tensor
|
|
|
|
from ..parallel_2d._operation import split_tensor_2d
|
|
from ..parallel_2p5d._operation import split_tensor_2p5d
|
|
from ..parallel_3d._operation import split_batch_3d
|
|
from ..utils import get_tensor_parallel_mode
|
|
|
|
_parallel_split_batch = {'2d': split_tensor_2d, '2.5d': split_tensor_2p5d, '3d': split_batch_3d}
|
|
|
|
|
|
def partition_batch(input_) -> Tensor:
|
|
tensor_parallel_mode = get_tensor_parallel_mode()
|
|
if tensor_parallel_mode in _parallel_split_batch:
|
|
if isinstance(input_, dict):
|
|
return {k: _parallel_split_batch[tensor_parallel_mode](v) for k, v in input_.items()}
|
|
else:
|
|
return _parallel_split_batch[tensor_parallel_mode](input_)
|
|
else:
|
|
return input_
|