scalar27
scalar27
I did leave out step 2 but then I get an error when I try to run: ModuleNotFoundError: No module named 'llama_cpp'
Thank you. I did get it to work following your comment. Like the other M1 person, I do get stuttering. It's a shame because the voice quality is excellent and...
I managed to get this working with the Gemma 2 model. However, I am having trouble setting the parameters. It's working but doesn't seem optimal. I see them in creation_params.json,...
Great. It seems like a more standard approach these days. I'd be happy to test whatever. As mentioned above I'm on a M1 Mac so this isn't the fastest setup...
I read that deepspeed only works with cuda (nvidia), thus not on a Mac.
+1, on Mac M1
Wow. I'm interested to learn more.
Not that I understand what he did, I think this person figured it out: https://github.com/DissonanceTK/MacReddy/commit/5560b21279e17f8c3947fc1c2de2914aa3af2609
I'd like to be able to use chat with X, and in particular, chat with PDF, using llama.cpp.
It also stops working if I start the code with wifi on and then turn it off. Also, the dev says running w/o internet works fine for him. Could it...