wsbagnsv1

Results 47 comments of wsbagnsv1

Will you implement some ncpu offload mechanism like in llama.cpp to only keep active parameters in vram and make the other offloadable to cpu/system ram? Not that you have to...

Ill do an gguf conversion then, which gguf you want exactly? Ill upload that one first

> > Ill do an gguf conversion then, which gguf you want exactly? Ill upload that one first > > The 14b version, thanks Ik that you want the 14b...

currently converting, ill upload the q8 first then (;

> q8 please Do you have your hugginface name? Ill do a private repo and add you for now, since i dont know it its actually working

You can test it when its online and tell me if the gguf works (;

You should have access now https://huggingface.co/QuantStack/Wan2.1-VACE-14B-GGUF

gonna take around 2h to upload probably, if the conversion is successful that is

I think there is native support for vace now, but only for the full model, so ggufs should work as well. Though I encountered some weird issues with 5d tensors...

Alright the Q8_0 is online, test it out and if it works tell me so ill publish the repo (;