Rick Zhou
Rick Zhou
cc: @MasterJH5574 Will need https://github.com/flashinfer-ai/flashinfer/pull/126 to be merged first:
@beaufortfrancois @tqchen Thanks a lot for putting the effort to support WebGPU in service worker. I was able to put up a sample Chrome extension running LLM in the service...
> please fix the jenkins here Should be addressed by https://github.com/mlc-ai/mlc-llm/pull/2292. I'm triggering a rebuild now
To fix CUDA error, https://github.com/apache/tvm/pull/16982
https://github.com/mlc-ai/mlc-llm/pull/2178
@tqchen PR to change this in JSONFFIEngine: https://github.com/mlc-ai/mlc-llm/pull/2225
Please take a look at existing operators at https://github.com/apache/tvm/blob/main/python/tvm/relax/frontend/nn/op.py. For example, torch.stack can be implemented with nn.op.unsqueeze + nn.op.concat
Hi @pchalasani @shahizat I was not able to reproduce the same error on my Mac. I suspect this is due to git configuration issue. Can you try directly running: ```...
@kidhan1234 Please make sure you're using the correct Python interpreter. Compare `pip show mlc-llm-nightly-cu121` and `python -c "import sys; print(sys.path)"` to make sure that the correct Python path is included.