mirror of https://github.com/hpcaitech/ColossalAI
polish sharded optim docstr and warning (#770)
parent
80e37eec42
commit
6e553748a7
|
@ -50,7 +50,7 @@ class ShardedOptimizerV2(ColossalaiOptimizer):
|
||||||
You must use ``ShardedOptimizerV2`` with ``ShardedModelV2``.
|
You must use ``ShardedOptimizerV2`` with ``ShardedModelV2``.
|
||||||
|
|
||||||
Note:
|
Note:
|
||||||
Make sure you enable ``use_memory_tracer`` in ``ShardedModelV2``,
|
Make sure you set ``tensor_placement_policy`` in ``ShardedModelV2`` to `"auto"`,
|
||||||
if you set ``gpu_margin_mem_ratio > 0``.
|
if you set ``gpu_margin_mem_ratio > 0``.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
|
@ -59,7 +59,6 @@ class ShardedOptimizerV2(ColossalaiOptimizer):
|
||||||
optimizer (Optimizer): An Optimizer instance.
|
optimizer (Optimizer): An Optimizer instance.
|
||||||
gpu_margin_mem_ratio (float, optional): The ratio of GPU remaining memory (after the first forward-backward)
|
gpu_margin_mem_ratio (float, optional): The ratio of GPU remaining memory (after the first forward-backward)
|
||||||
which will be used when using hybrid CPU optimizer.
|
which will be used when using hybrid CPU optimizer.
|
||||||
Make sure `reuse_fp16_shard` is enabled in `ShardedModelV2`, if `gpu_margin_mem_ratio` > `0.0`.
|
|
||||||
This argument is meaningless when `tensor_placement_policy` of `ShardedModelV2` is not "auto".
|
This argument is meaningless when `tensor_placement_policy` of `ShardedModelV2` is not "auto".
|
||||||
Defaults to 0.0.
|
Defaults to 0.0.
|
||||||
initial_scale (float, optional): Initial scale used by DynamicGradScaler. Defaults to 2**32.
|
initial_scale (float, optional): Initial scale used by DynamicGradScaler. Defaults to 2**32.
|
||||||
|
@ -119,7 +118,7 @@ class ShardedOptimizerV2(ColossalaiOptimizer):
|
||||||
|
|
||||||
# Store fp32 param shards
|
# Store fp32 param shards
|
||||||
self._register_master_weight()
|
self._register_master_weight()
|
||||||
if self.gpu_margin_mem_ratio != 0.0 and isinstance(sharded_model._tensor_placement_policy,
|
if self.gpu_margin_mem_ratio != 0.0 and not isinstance(sharded_model._tensor_placement_policy,
|
||||||
AutoTensorPlacementPolicy):
|
AutoTensorPlacementPolicy):
|
||||||
self._logger.warning(f'gpu_margin_mem_ratio is meaningless when tensor_placement_policy is not "auto"')
|
self._logger.warning(f'gpu_margin_mem_ratio is meaningless when tensor_placement_policy is not "auto"')
|
||||||
self._logger.debug(f"After init ShardedOptimizerV2 consumes {self.get_memory_usage()[0] / 1e6} MB CUDA Memory!",
|
self._logger.debug(f"After init ShardedOptimizerV2 consumes {self.get_memory_usage()[0] / 1e6} MB CUDA Memory!",
|
||||||
|
|
Loading…
Reference in New Issue