FAdam_PyTorch icon indicating copy to clipboard operation
FAdam_PyTorch copied to clipboard

an implementation of FAdam (Fisher Adam) in PyTorch

FAdam_PyTorch

an implementation of FAdam (Fisher Adam) in PyTorch

Please see the official Arxiv paper:
FAdam: Adam is a natural gradient optimizer using diagonal empirical Fisher information

Schedule:
1 - impl in eager PyTorch --> Complete and working.
2 - add adaptive epsilon --> Complete and working.

3 - (if torch.compile not performant) - update to fused Cuda kernel (cpp extension)