TinyNeuralNetwork
TinyNeuralNetwork copied to clipboard
TinyNeuralNetwork is an efficient and easy-to-use deep learning model compression framework.
My use case: Apply post training quantization to a pth model and convert to tflite. The generated tflite model fails to pass benchmark test with following error message: STARTING! Log...
Hi author, For a certain model, oneshotpruner fails when there's channel padding with following error message: ERROR (tinynn.graph.modifier) All node's sparsity in one subgraph must be the same Please let...
请问这是什么原因呢? 
Major changes: 1. Patching `handle_torch_function` and `has_torch_functions` for unrelated funcs (e.g. `torch._assert`) creates an infinite recursion loop. Fixed by introducing a new lock `handle_func_lock `. 2. Previously, we used `module.register_forward_hook`...
According to user feedback, the current supported ops are not sufficient for model deployment. This requires a major update of the existing TFLite schema. Currently, we are using the schema...
I got a QAT int8 per-channel tflite model. To check the accuracy, I compare the inference results between it and the de-quantized onnx model. 1. python3.6 -m tf2onnx.convert --opset 11...
First off, thanks for this amazing repo. Im working on a ssd model in pytorch and i want to add post processing (NMS) into the tflite, how can i add...
现在移动端部署接入了TFLite,需要实现Torch到TFLite的转换;建议后端接入MNN, MNN后端性能领先于TFLite, 且目前提供了torchscript -> mnn模型的转换工具。
Sorry for bothering again, when I use QAT on yolov5, the conversion of the model will cause significant errors. I compare them layer by layer and find out it starts...