Jiko
Jiko
Hi, I just wonder whether the splitting you implemented is from one of the paper in the reference?
## Description I tried to convert a mixed precision onnx model to mixed precision TensorRT engine. In my mixed precision onnx model, I have kept some ops (ReduceSum, Pow) to...
按照脚本预处理自有数据集,但发现内存爆炸,不足以支持pin_memory。 然后在config里关闭了pin_memory,但发现内存泄露。 按照我查看dataset和dataloader,按道理音频是在call的时候 才被读取,不应占有过多内存。从最初的1.2T慢慢增长到1.6T 请问有遇到过吗?
after converting onnx fp32 to int8 engine with custom calibration, the engine layers still show fp32
## Description I tried to follow the int8 custom calibration to build my int8 engine from onnx fp32 model. https://github.com/NVIDIA/TensorRT/tree/main/tools/Polygraphy/examples/cli/convert/01_int8_calibration_in_tensorrt After building the engine, I used the following to inspect...
### Please confirm before submission 在提交之前,请确认 - [x] I have searched for existing issues [search for existing issues](https://github.com/junjiem/dify-plugin-tools-mcp_sse/issues), including closed ones. 我已经搜索了现有问题[搜索现有问题](https://github.com/junjiem/dify-plugin-tools-mcp_sse/issues),包括已关闭的问题。" ### Dify version Dify版本 1.4.0 ### Plugin version...