LL3DA icon indicating copy to clipboard operation
LL3DA copied to clipboard

About training time&memory

Open hanxunyu opened this issue 1 year ago • 1 comments

Hi, thanks for your great work. May I ask what the training time and memory usage are when using the 7B parameter LLM?
Looking forward to your reply.

hanxunyu avatar Oct 24 '24 12:10 hanxunyu

For a 7B LLM, it takes about 1 - 2 days and approximately 40GB GPU mem. to achieve reasonable performance. With the help of flash attention, the training process may be even faster. You can make evaluation intervals longer for faster training.

ch3cook-fdu avatar Oct 24 '24 14:10 ch3cook-fdu