Chunyu

Results 9 comments of Chunyu

transformers v4.51.4 starts to support ASCEND NPU to directly enable ```flash_attention_2```. It seems that the transformers section in README needs to be adjusted.

[04.22] - OpenRLHF-NPU can enable ```--packing_samples```. - Update Q2 roadmap: ring_flash_attention will be supported in OpenRLHF-NPU.

Thanks for sharing this PR! I can successfully run DPO/RM on the NPU(Atlas 800T A2 Training Server) with ease. I am verifying the [Ray](https://github.com/ray-project/ray/pull/41256) part in OpenRLHF. Could you share...

https://github.com/OpenRLHF/OpenRLHF/issues/914 is a roadmap of the OpenRLHF-NPU workflow for reference.

> Will OpenRLHF support Huawei Ascend NPU (ShengTeng AI Processor)? https://github.com/OpenRLHF/OpenRLHF/issues/914 is a roadmap of the OpenRLHF-NPU workflow for reference.

> Hi [@zheliuyu](https://github.com/zheliuyu) ! Yes it totally makes sense, we can have a `use_local_kernels` flag inside `KernelConfig` to do that, do you want to open a PR for that ?...

WIP on https://github.com/zheliuyu/transformers-kernels

https://github.com/huggingface/transformers/pull/42800 PR merged. Documentation will be updated accordingly, but this issue can be closed. Thanks to everyone who followed this issue.❤

> [@zheliuyu](https://github.com/zheliuyu) Can you help me take a look? Thanks for the feedback. @FightingZhen will pay attention to this issue.