[BUG]: LoRA does not support the training of reward models
π Describe the bug

Environment
env
OSοΌubuntu 20.04 GPUοΌ4 x A10 python==3.9.0 torch==1.13.1-cu116 colossalai==0.2.5
command
python train_reward_model.py --pretrain "bigscience/bloom-560m" --lora_rank 16
I am very confused why the training encounters errors after setting lora_rank, whether it is the wrong way of my useγ
@ht-zhou I read your modifications, I don't think these modifications can fix the problem, and I noticed that the latest main version of ColossalAI removed the use of LoRA during rm model training, and now I think ColossalAI does not support LoRA at all, right?

Bot detected the issue body's language is not English, translate it automatically. π―ππ»π§βπ€βπ§π«π§πΏβπ€βπ§π»π©πΎβπ€βπ¨πΏπ¬πΏ
