Shubhadeep Das

Results 14 comments of Shubhadeep Das

@mohammedpithapur Thanks for fixing the error message! It should be either `model_engine: "triton-trt-llm"` OR `model_engine: "ai-playground"`

@mohammedpithapur we now have a dedicated example to showcase AI playground in latest codebase! No config file changes are needed (this avoids the confusion you faced). Please checkout the example...

@OlegSudakov Thanks for pointing this bug out. This should have been fixed with the latest v0.2.0 release. https://github.com/NVIDIA/GenerativeAIExamples/blob/main/notebooks/05_dataloader.ipynb Feel free to close this issue after checking.

@OlegSudakov Thanks for pointing this bug out. This should have been fixed with the latest v0.2.0 release. Can you please verify the same with latest codebase and close this issue?

Hi @dsbyprateekg thanks for reporting this! There seems to be some stability issues in the server backend which causes this. Are you able to reproduce this error consistently for all...

Hey @rbgo404 You can deploy the tensorRT-based LLM model by following the steps here https://nvidia.github.io/GenerativeAIExamples/latest/local-gpu.html#using-local-gpus-for-a-q-a-chatbot This notebook interacts with the model deployed behind `llm-inference-server` container which should get started up...

Thanks for reporting this. We are checking this and will get back to you shortly. @nealvaidya please help check this issue.

Hey @saivineethabvns, yes this is possible! As a prerequisite step to run this notebook, you need to deploy this example following the steps here https://nvidia.github.io/GenerativeAIExamples/latest/local-gpu.html#using-local-gpus-for-a-q-a-chatbot Now in this step where...

Thanks for reporting this bug. @mikemckiernan to help fix this.

@CaptainMcCrank here is the correct link until we fix this in our codebase https://nvidia.github.io/GenerativeAIExamples/latest/api-catalog.html#get-an-api-key-for-the-mixtral-8x7b-instruct-api-endpoint