llama-stack icon indicating copy to clipboard operation
llama-stack copied to clipboard

Failing to build llama-stack on Mac OS X

Open genbit opened this issue 1 year ago • 3 comments

Followed Getting Started guide, tried to build locally on Mac OS 15.0 (M1):

  Using cached fairscale-0.4.13.tar.gz (266 kB)
  Installing build dependencies ... done
  Getting requirements to build wheel ... done
  Installing backend dependencies ... done
  Preparing metadata (pyproject.toml) ... done
ERROR: Could not find a version that satisfies the requirement fbgemm-gpu==0.8.0 (from versions: none)
ERROR: No matching distribution found for fbgemm-gpu==0.8.0
Failed to build target my-local-stack with return code 1

genbit avatar Sep 27 '24 19:09 genbit

+1. not sure if fbgemm-gpu is on macos

linsssbei avatar Sep 27 '24 20:09 linsssbei

I'm getting the same issue, no way macOS can't build this? What am I doing wrong?

californias66 avatar Sep 28 '24 01:09 californias66

The current inference supports only Cuda-based drivers, if you want to build on macos, you can use support with external inference like Ollama. It would work with that.

bhargavyagnik avatar Sep 28 '24 03:09 bhargavyagnik

Thanks @bhargavyagnik for pointers.

This is how it worked for me:

# shows me list of templates available for build command
$ llama stack build --list-templates

# use local-ollama inferencing
$ llama stack build --template local-ollama --name my-local-stack

This builds a stack using meta-reference for everything except using remote::ollama for inferencing.

genbit avatar Sep 28 '24 18:09 genbit

Thank you fam you're all GOATS fsho 🙏🏽🔥😮‍💨

californias66 avatar Sep 28 '24 18:09 californias66