Kevin
Kevin
### Summary - Continue's edit mode fails to apply the correct indentation - But the API response contains the correct indentation - -> Bug is NOT caused by LLM -...
I can confirm inconsistent outputs with Ollama `v0.6.3-rc0` for several models I tested: `llama3.2:latest` `llama3.2-vision:11b` `gemma3:12b` I noticed that `llama3.2:latest` produces inconsistent results for identical inputs not only on the...
@rick-github Thank you for looking into this issue! I am able to reproduce the inconsistent embedding results by running Ollama compiled from source and setting `CUDA_VISIBLE_DEVICES` either to `0` or...
# Summary It looks like three different issues might cause different embeddings, logits, and text generation for the same inputs: 1) Generating the output on different operating systems 2) Generating...
> We do NOT do anything explicit or special rgd KV cache @flexorx Does the following correctly describe your issue when you submit multiple identical input queries? 1. The text...