mirror of https://github.com/hpcaitech/ColossalAI
Jiarui Fang
2 years ago
committed by
GitHub
4 changed files with 44 additions and 72 deletions
@ -1,40 +1,5 @@
|
||||
import pytest |
||||
from colossalai.utils.model.colo_init_context import ColoInitContext |
||||
|
||||
import torch |
||||
|
||||
from colossalai.utils.cuda import get_current_device |
||||
|
||||
|
||||
@pytest.mark.skip |
||||
# FIXME(ver217): support lazy init |
||||
def test_lazy_init(): |
||||
in_dim = 4 |
||||
out_dim = 5 |
||||
|
||||
with ColoInitContext(lazy_memory_allocate=True) as ctx: |
||||
fc = torch.nn.Linear(in_dim, out_dim, bias=True) |
||||
|
||||
# lazy_memory_allocate=True, no payload is maintained |
||||
assert fc.weight._torch_tensor.numel() == 0 |
||||
|
||||
fc.weight.torch_tensor() |
||||
assert fc.weight._torch_tensor.numel() == in_dim * out_dim |
||||
|
||||
|
||||
@pytest.mark.skip |
||||
def test_device(): |
||||
in_dim = 4 |
||||
out_dim = 5 |
||||
|
||||
with ColoInitContext(lazy_memory_allocate=True, device=get_current_device()) as ctx: |
||||
fc = torch.nn.Linear(in_dim, out_dim, bias=True) |
||||
|
||||
# eval an lazy parameter |
||||
fc.weight.torch_tensor() |
||||
assert fc.weight.device == get_current_device() |
||||
|
||||
|
||||
if __name__ == '__main__': |
||||
test_lazy_init() |
||||
test_device() |
||||
from colossalai.utils.model.colo_init_context import ColoInitContext |
||||
|
Loading…
Reference in new issue