Wuzzooy

Results 13 comments of Wuzzooy

> > A tip: you likely need to use the reply template. See e.g. the help here. The model tries to continue (not reply) then instead. > > Help where?...

I've tried with llama.cpp python api server as the backend and i don't have the behavior i reported so i will close the issue.

Because the .bat launch app.py with the flag --rvc and --deepspeed. You need to add those arguments when you launch app.py manually app.py --rvc --deepspeed

Hello, No problem for the late reply, I'm really grateful for your work. I still had this issue with the updated version when trying to use rvc and found out...

Me neither i don't have the performance reported by RandomInternetPreson but when i use the realtimeTTS version installed in the same environment which use coqui engine i can generate few...

i'm able to run deepspeed on windows with python 3.9, i failed with python 3.10/11 I used this file to install it https://huggingface.co/Jmica/audiobook_maker/tree/main pydantic has to be under 2.0 or...

I've got the starting balance right but i'm still getting performance difference compared to the demo when i try other questions. I have more randomness via llama.cpp than the demo...

I just brought up v2.6 to say that i'm using the last llama.cpp with your PR in progress for v2.6 8967 but the model i use for this test is...

I've tried the 2.6 version and i didn't find any performance issue. I've tried to convert v2.5 with convert_hf_to_gguf.py but i have an error about the tokenizer ``` WARNING:hf-to-gguf:************************************************************************************** WARNING:hf-to-gguf:**...