llama.cpp icon indicating copy to clipboard operation
llama.cpp copied to clipboard

it’s so long a lime to wait while using a server

Open niubi-AI opened this issue 2 years ago • 3 comments

niubi-AI avatar Jun 01 '23 08:06 niubi-AI

Any way to run as quickly as using llama.cpp directly? I need to save each input and response

niubi-AI avatar Jun 01 '23 08:06 niubi-AI

@dansinboy are you using the default server binary that comes with llama.cpp or a binding?

jessejohnson avatar Jun 01 '23 14:06 jessejohnson

@dansinboy are you using the default server binary that comes with llama.cpp or a binding?

you get the point , at first I used a binding mode with llama_cpp_python, worked badly. and then change into default server, wow, it works well right now ...

niubi-AI avatar Jun 01 '23 14:06 niubi-AI

This issue was closed because it has been inactive for 14 days since being marked as stale.

github-actions[bot] avatar Apr 10 '24 01:04 github-actions[bot]