mirror of https://github.com/hpcaitech/ColossalAI
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
22 lines
502 B
22 lines
502 B
2 years ago
|
'''
|
||
|
torchrun --standalone --nproc_per_node=1 debug.py
|
||
|
'''
|
||
|
|
||
|
from diffusers import AutoencoderKL
|
||
|
|
||
|
import colossalai
|
||
|
from colossalai.utils.model.colo_init_context import ColoInitContext, post_process_colo_init_ctx
|
||
|
|
||
|
path = "/data/scratch/diffuser/stable-diffusion-v1-4"
|
||
|
|
||
|
colossalai.launch_from_torch(config={})
|
||
|
with ColoInitContext(device='cpu'):
|
||
|
vae = AutoencoderKL.from_pretrained(
|
||
|
path,
|
||
|
subfolder="vae",
|
||
|
revision=None,
|
||
|
)
|
||
|
|
||
|
for n, p in vae.named_parameters():
|
||
|
print(n)
|