MachRepo

Results 2 issues of MachRepo

Hello, I am trying to implement a mamba based model, whenever I try to increase d_model above 100 I get this error message. I am using torch.cuda.amp for mixed precision...

I was testing the module using this code : ``` from xLSTM.model import xLSTM import torch model = xLSTM(5, 8, 16, 5, 2, 0.1, True, 'slstm') inputs = torch.randint(low=0, high=5,size=(12,15000))...

bug