You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/tensor
HELSON 260a55804a
[hotfix] fix shape error in backward when using ColoTensor (#1298)
2 years ago
..
__init__.py [tensor] a shorter shard and replicate spec (#1245) 2 years ago
colo_parameter.py [tensor] add zero_like colo op, important for Optimizer (#1236) 2 years ago
colo_tensor.py [hotfix] fix shape error in backward when using ColoTensor (#1298) 2 years ago
compute_spec.py [ColoTensor] rename APIs and add output_replicate to ComputeSpec (#1168) 2 years ago
const.py
dist_spec_mgr.py [hotfix] Dist Mgr gather torch version (#1284) 2 years ago
distspec.py [tensor] distributed checkpointing for parameters (#1240) 2 years ago
op_wrapper.py [tensor] refactor colo-tensor (#992) 3 years ago
param_op_hook.py [refactor] move process group from _DistSpec to ColoTensor. (#1203) 2 years ago
process_group.py [tensor] redistribute among different process groups (#1247) 2 years ago
tensor_spec.py [refactor] move process group from _DistSpec to ColoTensor. (#1203) 2 years ago
utils.py reorgnize colotensor directory (#1062) 3 years ago