Matthieu Beaumont
Matthieu Beaumont
> > which OS are you running it on? i am trying to build on windows, although i am new to building windows binaries and haven't exactly done this before...
when i try with deepseek-coder-v2:16b-lite-instruct-q8_0 : llama_new_context_with_model: flash_attn requires n_embd_head_k == n_embd_head_v - forcing off llama_new_context_with_model: V cache quantization requires flash_attn llama_init_from_gpt_params: error: failed to create context with model '/mnt/AI_TEXT/OLLAMA/blobs/sha256-x'...
try pip install sentencepiece ?
qwen:14b is a very old model, try to use qwen2.5:14b