MacOS15.3.2 ,Apple M4 Max, can not run, all the logs no error, but when run the commands show the error
python run_inference.py -m models/BitNet-b1.58-2B-4T/ggml-model-i2_s.gguf -p "You are a helpful assistant" -cnv
Traceback (most recent call last):
File "/Volumes/OWCPSSD/BitNet/run_inference.py", line 56, in
run_inference()
File "/Volumes/OWCPSSD/BitNet/run_inference.py", line 37, in run_inference
run_command(command)
File "/Volumes/OWCPSSD/BitNet/run_inference.py", line 11, in run_command
subprocess.run(command, shell=shell, check=True)
File "/Users/sunnymusk/miniconda3/envs/bitnet-cpp/lib/python3.9/subprocess.py", line 505, in run
with Popen(*popenargs, **kwargs) as process:
File "/Users/sunnymusk/miniconda3/envs/bitnet-cpp/lib/python3.9/subprocess.py", line 951, in init
self._execute_child(args, executable, preexec_fn, close_fds,
File "/Users/sunnymusk/miniconda3/envs/bitnet-cpp/lib/python3.9/subprocess.py", line 1837, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
FileNotFoundError: [Errno 2] No such file or directory: 'build/bin/llama-cli'
there are two files in build/bin ,but no llama-cli
ggml-common.h ggml-metal.metal
Try my related tip: https://github.com/microsoft/BitNet/issues/192#issuecomment-2818299330
build with -DLLAMA_BUILD_COMMON=ON -DLLAMA_BUILD_TOOLS=ON or just use provided llama.cpp submodule