ColossalAI icon indicating copy to clipboard operation
ColossalAI copied to clipboard

[FEATURE]: Customize DDP arguments

Open ExtremeViscent opened this issue 4 years ago • 1 comments

Describe the feature

In the case when users need to specify DDP launching parameters (e.g. find_unused_parameters), keyword arguments should be passed to torch.nn.parallel.DistributedDataParallel at initialization.

ExtremeViscent avatar Apr 06 '22 00:04 ExtremeViscent

Thanks for the feedback, will add this soon.

FrankLeeeee avatar Apr 06 '22 06:04 FrankLeeeee

We have updated a lot. This issue was closed due to inactivity. Thanks.

binmakeswell avatar Apr 13 '23 03:04 binmakeswell

If I use zero2, and there are some parameters that should not be trained, how should I set them? If there are untrained parameters, you may encounter the following error: "AssertionError: fp32 param and grad have different shape".

shyoulala avatar May 17 '23 09:05 shyoulala