Patrick Devine

Results 426 comments of Patrick Devine

I'm going to close this since #5922 was merged, but we can reopen it if you're still hitting it.

You should now be able to get online help for each of the environment variables. e.g. `ollama serve -h`. There are several environment variables which are purposely _not_ exposed just...

Let's close this in favour of #4618 . You _can_ use Ollama to load GGLA based LoRA adapters (ggla being the old "gguf" file format specifically for LoRAs), but it's...

Let's close the issue. We can reopen if it's still a problem.

@ZhangZangQian can you update to the latest version of ollama? `0.1.39`. I believe it should fix the issue w/ `ollama run`. The problem I think is: ``` llama_model_load: error loading...

This should be fixed w/ `0.1.39` which now can parse a utf16 file (albeit _only_ w/ 8 bit characters). There's another PR coming for allowing utf16 characters in the modelfile...

Sorry about this one, guys. I believe there are two problems here: 1. we're not detecting the edge of the screen correctly with multi-byte chars; and 2. when we "brack...

Also, for now, a workaround is to include the `--nolinewrap` flag when you're calling `ollama run`.

The problem was that llama.cpp changed how the tokenizer worked because of changes w/ llama3 tokenization. This should be fixed in `0.1.39` though, so I'll go ahead and close the...