VLMEvalKit
VLMEvalKit copied to clipboard
how to run on multi-gpu with device_map='auto'
when I use 34B llm model, single-gpu will report OOM. so I set device_map='auto', but It seems cant use torchrun, It takes too much time to inference. how to solve this problem?
Hi, which VLM are you using?
Close the issue due to no response in weeks, plz reopen if needed
I use yi-34B LLM. It seems visual encoder cost too much time.(32 frames per video)