Results 3 comments of Yuan Sui

@dgiofre @IcyFeather233 Have tried @dgiofre solution, but still failed. Then I tried to reinstall the flash_attn from the source releases using the following code: ```bash pip install https://github.com/Dao-AILab/flash-attention/releases/download/v2.8.3/flash_attn-2.8.3+cu12torch2.8cxx11abiTRUE-cp310-cp310-linux_x86_64.whl ``` which...

Have the same issue here. The error still exists not only for meta/meta-llama-3-70b-instruct, but also for model mixtral-8x7b-instruct-v0.1.