Error of fused_dense module of flash_attn
Hi author, I have compiled and installed fused_dense_lib successfully. But when I tried to run the finetuning code, I encountered this error: "RuntimeError: linear_act_forward failed." which is due to the line 291 in InternVideo2/single_modality/models/internvideo2.py. The complete error is as shown below:
Traceback (most recent call last):
File "
Could you please help me to fix this? Thank you very much.
I have not met this problem. Maybe you can refer to this issue https://github.com/Dao-AILab/flash-attention/issues/289#issuecomment-1620944622.
I have not met this problem. Maybe you can refer to this issue Dao-AILab/flash-attention#289 (comment).
Thank you for your reply. I follow the above comment to not use fused_mlp with DeepSpeed. This can fix this error.