llama-cpp-python
llama-cpp-python copied to clipboard
Python bindings for llama.cpp
850 return issubclass(cls.__origin__, self.__origin__) 851 if not isinstance(cls, _GenericAlias): --> 852 return issubclass(cls, self.__origin__) 853 return super().__subclasscheck__(cls) TypeError: issubclass() arg 1 must be a class The manual package works with...
Making the latest llama.cpp repo with `make LLAMA_OPENBLAS=1` works as it should. But running `LLAMA_OPENBLAS=1 pip install --force-reinstall --no-cache-dir --no-index --find-links . llama-cpp-python` (running on an offline server, its the...
Currently this comes up as an AssertionError which leads to a lot of confusion
* Add a feature with unlimited max_tokens that is enabled when max_tokens
Certain tokens in the vocabulary cannot be decoded to valid utf-8, I'm actually not sure if this is because they represent partial utf codepoints, but in any case they cause...
Edit : For now i've installed the wheel from "https://github.com/Loufe/llama-cpp-python/blob/main/wheels/llama_cpp_python-0.1.26-cp310-cp310-win_amd64.whl". The installation of the wheel works. So everything is fine for me. Got things working also in WSL with no...
Allow the user to alias their local models to OpenAI model names as many tools have those hard-coded. This may cause unexpected issues with tokenization mismatches.