automaticcat
automaticcat
@CameronNguyen130820 we postponed solving this issue since it might related to outdated GPU doesn't support some operation
i try to change the code just like the fix, but not working
https://www.npmjs.com/package/systeminformation
Intel optimization: Instructions set: AVX2 - Consumer grade AVX512 - Before gen 12 or Consumer grade AVX_VNNI - Consumer grade with OneAPI AVX512_VNNI - Server grade
Windows WSL https://github.com/janhq/jan/issues/912 oneAPI for Windows https://github.com/janhq/jan/issues/911 AMD GPU for Windows https://github.com/janhq/jan/issues/913 AMD CPU for Windows https://github.com/janhq/jan/issues/914 Intel GPU (maybe crossing with oneAPI) https://github.com/janhq/jan/issues/915
> I believe that Intel oneMKL should actually run on an Intel GPU: https://www.intel.com/content/www/us/en/docs/oneapi/optimization-guide-gpu/2023-0/offloading-onemkl-computations-onto-the-gpu.html I think we should bring this issue back, iGPU offloading at least the prompt eval is...
Notes on some principle to follow: - Should try to avoid system permission (infer hardware from a more generic way without high system permission) - Windows, Linux, MacOS is different...
I think this feature is important to make llama cpp usage spread even more
Re-open please, everyone is using 4-5bit quantization now
Hi, stop wasting earth energy and open source it already thanks