Long inference time
Hello, thank you for providing this excellent work. However, I have encountered an issue while using the model for inference. I have utilized fifteen images and generated a .pth file using the build_prototypes.ipynb script. I am loading the label names and test class weights from this file. Here are the rest of my settings: `config_file="configs/open-vocabulary/lvis/vitl.yaml", rpn_config_file="configs/RPN/mask_rcnn_R_50_FPN_1x.yaml", model_path="configs/open-vocabulary/lvis/vitl_0069999.pth",
image_dir='demo/plliquid_input',
output_dir='demo/output/mask_output',
category_space='liquid_dataset/plrealsense_liquid_prototypes.pth',
device='cpu',
overlapping_mode=True,
topk=1,
output_pth=False,
threshold=0.8`
Image size: 1280x720 Number of recognition categories: 4 Unfortunately, I have noticed that the inference time for each image exceeds 45 seconds. I would like to inquire if this speed is normal, or if there are any measures I can take to reduce the inference time. Any guidance or suggestions would be greatly appreciated.
I have the same problem. Did you manage to find a solution? @z7r7y7 I have set my K-value to 1 and have confirmed that it is running on my GPU. Setup:
- Nvidia RTX 4050 Laptop GPU
- 6 GB VRAM
- Inference time per image (HD): 20 - 32 seconds
I see that there are numerous open issues inquiring about slow inference time @mlzxy ?
I also have the same problem. Has anyone found a fix?
Thanks