Léon
Léon
> I was able to run GPipe with lm.one_billion_wds.OneBWdsGPipeTransformerWPM in a single node with multiple GPUs. > > However, I am a little confused about how to run GPipe with...
> > Could you point me to (or share) instructions on how to run the one_billion_wds using GPipe on multiple local GPUs? > > You can run it using the...
> > > > Could you point me to (or share) instructions on how to run the one_billion_wds using GPipe on multiple local GPUs? > > > > > >...
> Hey, were you able to resolve this issue? I'm facing the same problem. Hi, have you found the solution to this issue?
> I did a lot of debugging and found out that there is an issue with input generation (for example, a tensor shaped (32, ) is being converted to (32,...
> You can find it [here](https://github.com/adis98/Lingvo_modified) Thanks
> You have a system with multiple gpus right? Could you let me know if it is running fine? If there are any bugs do let me know. sure, I...
> I've tested the code (the image processing model) with 2 GPUS (8GB, M60 Tesla) and it seems to be working fine. oh, nice! I have not got time to...
Hi, I have 4Gpus(v100) and I want to try to run this model. But I don't know what the number of _saver_max_to_keep_ & _worker_replicas_ means. Should I set the same...