yukiyee
yukiyee
> Hi,what's your batch size on each gpu? The microbatch size is the unit of passing when using pipeline parallel. > > If your batch size is more than 1,...
Me too !! RuntimeError: Failed to replace input_layernorm of type LlamaRMSNorm with FusedRMSNorm with the exception: No module named 'fused_layer_norm_cuda'. Please check your model configuration or sharding policy, you can...
And I saw this prompt in `examples/language/llama2/scripts/benchmark_70B/3d.sh` ``` # TODO: fix this echo "3D parallel for LLaMA-2 is not ready yet" ``` Does it mean , even if I deployed...
> > And I saw this prompt in `examples/language/llama2/scripts/benchmark_70B/3d.sh` > > ``` > > # TODO: fix this > > echo "3D parallel for LLaMA-2 is not ready yet" >...
> ps -ef | grep > > then kill -9 the demon which bind to the port I have done this before, however it doesn't work. And I'm sure the...
Finally , I solve the problem as below : First, using `python xx.py` instead of `colossalai run --nproc_per_node 8 xx.py` works well. So the start command is ```shell srun -p...