ggllm.cpp
ggllm.cpp copied to clipboard
Falcon LLM ggml framework with CPU and GPU support
ggml and llama.cpp support [Metal](https://github.com/ggerganov/llama.cpp/pull/1642), do Apple Silicon users need to use LLaMA.cpp or can they use gglm.cpp with Falcon?
I'm stuck with other work, I recently pushed half finished branch, containing a ton of fixes and changes but not finished. Also moved from falcon_main to "ggfalcon" which is meant...
I plan to PR today, though it depends on final progress. The computation speed is slow because we currently have no mulmat kernel with interleaving broadcast support yet, so tests...
Just a heads up, given it's more than a week since last release. I'm deep in a complete overhaul of a series of behavior and functions. The core focus is...
The parsing of the parameters is broken. Basically, parameters were only parsed correctly to arguments if it was the last argument of the command line. For most parameter errors in...
I have always been irritated (also in Llama.cpp) by the fact that the help text cannot be piped. Neither a `falcon-main --help |less` (the help is meanwhile 85 lines long)...
# Prerequisites - [X] I am running the latest code. Development is very rapid so there are no tagged versions as of now. - [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)....
# Prerequisites - [X ] I am running the latest code. Development is very rapid so there are no tagged versions as of now. - [X ] I carefully followed...
# Prerequisites - [X ] I am running the latest code. Development is very rapid so there are no tagged versions as of now. - [X ] I carefully followed...