onesnep

Results 2 issues of onesnep

I was glad to see [Flash Attention ported to ROCM](https://github.com/ROCmSoftwarePlatform/flash-attention), however currently compatibility is limited to gfx90a. I and many others would love to see this on other architectures. When...

Under Investigation

# Description The previous logic in `resolve_dtype` incorrectly configured the model's `torch_dtype` for Automatic Mixed Precision (AMP) training. When `fp16: true` or `bf16: true` was set for mixed precision, the...