lora
lora copied to clipboard
works but could use guidance
I seem to be one of the few to get it working
here are some thoughts and issues
- might require bleeding edge diffusers and transformers and accelerate - i always have them as current git head, others are probably using older or much older versions.
- for working on 6GB requires xformers and bnb
- I had to downgrade xformers due to latest having intermittent incompatibility (xformers cutlass backward isn't compatible with 30xx GPUs and maybe not 40xx GPUs, but it defaults to it sometimes to cutlass in current xformers for memory efficient attention)
- the learning rate has to be massively higher than expected (1e-4 !!!!) people are used to learning rates of 1e-6 or so and it learns essentially nothing.
- people don't understand what LORA is doing, might want to provide a brief explanation of why it can do dreambooth with such a small file difference.
Perhaps do a 1.4 example also since people are more familiar with it than 2.1.
You might want to do a tweaked example showing the corgi working (what I posted on reddit works, but isn't very good)
https://huggingface.co/docs/diffusers/training/dreambooth
https://drive.google.com/drive/folders/1BO_dyz-p65qhBRRMRA4TbZ8qW4rB99JZ
You might want to have the end of the example turn it into a .ckpt file so people get something they are used to using.
Also I'd recommend an example trained on a person.