Vatshank Chaturvedi
Vatshank Chaturvedi
Hey @cathyzhyi / @dan-garvey, I am new to torch-mlir and if you folks aren't already working on this, I would like to take a stab at this. Do we first...
Hey @muellerzr, is there ongoing work for adding [XLA support](https://github.com/pytorch/xla/pull/3431) to FSDP? We, on the AWS SageMaker training compiler side, have started looking into XLA-FSDP and might be able to...
Okay cool @muellerzr! Although our focus is on GPUs, I am sure there will be significant overlap in the code for adding support for either device type. What do you...
@lilyq I had the same issue. I uninstalled transformers/tokenizers first and then pip reinstalled from source using `pip install git+https://github.com/huggingface/transformers` (all within my conda env). This installed the right version...
Hey @HamidShojanazeri, I am curious about your example too (going to attempt something very similar :) ). Did you get this to work, either with a different TP size or...
Hi @fduwjj, I tried this with TP size = 2 and still running into the same `RuntimeError: aten.add.Tensor: got mixed distributed and non-distributed tensors.`. Any ideas what could be going...
Hi @anj-s, is TorchDynamo support being worked on? I see dynamo mentioned in some of the code/comments under `spmd/compiler` and `pippy/fx` (and the `pippy_dynamo.py` example) but wasn't sure if the...