Matan Kleyman
Matan Kleyman
@JaydenKing32 Can you please describe the method that worked for you ? Keeping h.num_classes as coco default and only change h.label_map to your custom dataset ? Thanks
As far as the hotswap is still not implemented.. What is the current best way to run lora weights using vllm ? Should I merge the lora weights and the...
Hi guys, thanks for the great work. Do you know when is this planned to be released ? I see it was merged but no release was published yet.
This is the paramters config file: ` num_classes: 99 anchor_scale: 1.0 label_map: {*****, dict of 98 classes'}`
I'm interested in the ability to delete models and bentos as well. @eledhwen Did you find any workaround ? @parano Is it still on the roadmap ?
This PR would be super valuable for us. @pfldy2850 Do you plan to adjust it to the current master branch ? Because I see it is a bit outdated.
Any updates on how to fix that? facing the same issue when running mistral-7b quantized 4 bit with --backend vllm. @pavel1860 @aarnphm @1E04 @marijnbent @Bernsai @dudeperf3ct
@haotian-liu I try to run the 4-bit 34B on 24GB Ram but I'm pretty sure it offloads some of the weights to cpu, because of `low_cpu_mem_usage=True` which results in the...
Same issue here. Were you able to fix that ? @levi @iceman-p I suspect this is related to `device="auto"` and `low_cpu_mem_usage=True`