ggllm.cpp icon indicating copy to clipboard operation
ggllm.cpp copied to clipboard

Falcon LLM ggml framework with CPU and GPU support

Results 28 ggllm.cpp issues
Sort by recently updated
recently updated
newest added

ggml and llama.cpp support [Metal](https://github.com/ggerganov/llama.cpp/pull/1642), do Apple Silicon users need to use LLaMA.cpp or can they use gglm.cpp with Falcon?

I'm stuck with other work, I recently pushed half finished branch, containing a ton of fixes and changes but not finished. Also moved from falcon_main to "ggfalcon" which is meant...

I plan to PR today, though it depends on final progress. The computation speed is slow because we currently have no mulmat kernel with interleaving broadcast support yet, so tests...

enhancement
finished

Just a heads up, given it's more than a week since last release. I'm deep in a complete overhaul of a series of behavior and functions. The core focus is...

enhancement

The parsing of the parameters is broken. Basically, parameters were only parsed correctly to arguments if it was the last argument of the command line. For most parameter errors in...

I have always been irritated (also in Llama.cpp) by the fact that the help text cannot be piped. Neither a `falcon-main --help |less` (the help is meanwhile 85 lines long)...

finished

# Prerequisites - [X] I am running the latest code. Development is very rapid so there are no tagged versions as of now. - [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)....

# Prerequisites - [X ] I am running the latest code. Development is very rapid so there are no tagged versions as of now. - [X ] I carefully followed...

# Prerequisites - [X ] I am running the latest code. Development is very rapid so there are no tagged versions as of now. - [X ] I carefully followed...