Andrew J. Holmes

Results 4 comments of Andrew J. Holmes

I have a similar issue as well. Hope it gets fixed in llama.cpp or wherever.

@hlhr202 u gotta hire us when you make it big :) LGTM

Got it running by using the .bin file from here: https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/tree/main Had no luck generating the q5_1 from here (via the instructions): https://github.com/ggerganov/llama.cpp#prepare-data--run If this is a common problem maybe...

You'd think this would already be included with most projects