You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/tests/test_elixir/test_tools/test_registry.py

24 lines
576 B

import pytest
import torch
from colossalai.testing import run_on_environment_flag
from tests.test_elixir.utils import to_cuda
@run_on_environment_flag('ELX')
def test_registry():
from tests.test_elixir.utils.registry import TEST_MODELS
for name, model_tuple in TEST_MODELS:
torch.cuda.synchronize()
print(f'model `{name}` is in testing')
model_fn, data_fn = model_tuple
model = model_fn().cuda()
data = to_cuda(data_fn())
loss = model(**data)
loss.backward()
if __name__ == '__main__':
test_registry()