onesnep
Results
2
issues of
onesnep
I was glad to see [Flash Attention ported to ROCM](https://github.com/ROCmSoftwarePlatform/flash-attention), however currently compatibility is limited to gfx90a. I and many others would love to see this on other architectures. When...
Under Investigation
# Description The previous logic in `resolve_dtype` incorrectly configured the model's `torch_dtype` for Automatic Mixed Precision (AMP) training. When `fp16: true` or `bf16: true` was set for mixed precision, the...