wsbagnsv1
wsbagnsv1
Will you implement some ncpu offload mechanism like in llama.cpp to only keep active parameters in vram and make the other offloadable to cpu/system ram? Not that you have to...
Ill do an gguf conversion then, which gguf you want exactly? Ill upload that one first
> > Ill do an gguf conversion then, which gguf you want exactly? Ill upload that one first > > The 14b version, thanks Ik that you want the 14b...
currently converting, ill upload the q8 first then (;
> q8 please Do you have your hugginface name? Ill do a private repo and add you for now, since i dont know it its actually working
You can test it when its online and tell me if the gguf works (;
You should have access now https://huggingface.co/QuantStack/Wan2.1-VACE-14B-GGUF
gonna take around 2h to upload probably, if the conversion is successful that is
I think there is native support for vace now, but only for the full model, so ggufs should work as well. Though I encountered some weird issues with 5d tensors...
Alright the Q8_0 is online, test it out and if it works tell me so ill publish the repo (;