Francesco Petrini
Francesco Petrini
Latest changes pass all L0_backend_python tests. I am planning a new section to the README.md file of this repo so I can include it in the same ticket and we...
This functionality is currently in development. When complete, there will be corresponding documentation detailing how to use the new endpoint.
Hello @AlexanderKomarov thank you for reporting an issue. I was unable to reproduce this behavior on my end using 4 copies of the same model. Is it possible to provide...
Hello @AlexanderKomarov, apologies for the late response! I was able to reproduce the issue using the repo you provided. I'll create a ticket for our team to investigate further.
Hello, First off, I apologize that this issue went unnoticed for so long. Thank you for bringing it to our attention. I have this as my top priority for this...
We're still investigating convenient packaging steps for a re-spin, however, from what I've found, I believe `libboost_filesystem.so.1.80.0` is the only missing item. Does executing the following steps solve the issue...
Hi @langong347, Thank you for submitting an issue. I notice your config does not set a different value for `intra_op_thread_count`, so yes, I believe the number of threads corresponds directly...
@langong347, Doing some testing: > 2. I wonder whether the 70-100 threads spun up in the main container (with Triton) is some sort of default thread pool used by Triton...
Hi folks, I've updated the 23.06 release [page](https://github.com/triton-inference-server/server/releases/tag/v2.35.0) with the new asset: [tritonserver2.35.0-jetpack5.1.2-update-2.tgz](https://github.com/triton-inference-server/server/releases/download/v2.35.0/tritonserver2.35.0-jetpack5.1.2-update-2.tgz). It proved too cumbersome to remove the boost dependency, so this new asset contains the same changes...
Can the stakeholders provide another round of reviews on this PR? We'd like to get these changes into a release asset this week.