ColossalAI icon indicating copy to clipboard operation
ColossalAI copied to clipboard

GPU memory used for stable diffusion

Open cyysc1998 opened this issue 3 years ago • 4 comments

I finetune the stable diffusion on one RTX3090 with config

python main.py --logdir /tmp -t --postfix test -b configs/train_colossalai.yaml

The config for lighting is as follows: image

image The batch size is set to 1. However, it is still out of memory. Is this normal? Thank you!

cyysc1998 avatar Nov 10 '22 12:11 cyysc1998

I can get it to start training by setting placement_policy to cpu or auto on my 2070 SUPER. But it crashes before it finishes training.

DonStroganotti avatar Nov 10 '22 19:11 DonStroganotti

@DonStroganotti Setting placement_policy to cpu works. But like you, is crashes after 5 epoches.

cyysc1998 avatar Nov 11 '22 02:11 cyysc1998

In addition to GPU memory, I'm curious about how much RAM memory is needed for training.

My device has 32GB of RAM and an RTX 3090 (24GB), but it runs out of memory after a few steps.

feisan avatar Nov 18 '22 09:11 feisan

I have 4*3090, I'll try it next week

Alfred-Duncan avatar Dec 16 '22 09:12 Alfred-Duncan

We have updated a lot. This issue was closed due to inactivity. Thanks.

binmakeswell avatar Apr 13 '23 10:04 binmakeswell