automaticcat

Results 98 comments of automaticcat

@CameronNguyen130820 we postponed solving this issue since it might related to outdated GPU doesn't support some operation

i try to change the code just like the fix, but not working

https://www.npmjs.com/package/systeminformation

Intel optimization: Instructions set: AVX2 - Consumer grade AVX512 - Before gen 12 or Consumer grade AVX_VNNI - Consumer grade with OneAPI AVX512_VNNI - Server grade

Windows WSL https://github.com/janhq/jan/issues/912 oneAPI for Windows https://github.com/janhq/jan/issues/911 AMD GPU for Windows https://github.com/janhq/jan/issues/913 AMD CPU for Windows https://github.com/janhq/jan/issues/914 Intel GPU (maybe crossing with oneAPI) https://github.com/janhq/jan/issues/915

> I believe that Intel oneMKL should actually run on an Intel GPU: https://www.intel.com/content/www/us/en/docs/oneapi/optimization-guide-gpu/2023-0/offloading-onemkl-computations-onto-the-gpu.html I think we should bring this issue back, iGPU offloading at least the prompt eval is...

Notes on some principle to follow: - Should try to avoid system permission (infer hardware from a more generic way without high system permission) - Windows, Linux, MacOS is different...

I think this feature is important to make llama cpp usage spread even more

Re-open please, everyone is using 4-5bit quantization now

Hi, stop wasting earth energy and open source it already thanks