Andrew J. Holmes
Andrew J. Holmes
I have a similar issue as well. Hope it gets fixed in llama.cpp or wherever.
@hlhr202 u gotta hire us when you make it big :) LGTM
Got it running by using the .bin file from here: https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/tree/main Had no luck generating the q5_1 from here (via the instructions): https://github.com/ggerganov/llama.cpp#prepare-data--run If this is a common problem maybe...
You'd think this would already be included with most projects