-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MNN batch inference time not more efficient than single image #3184
Comments
|
I've run the modified test with dimensions closer to mine. The results are below. My image is a tensor of shape (3, 413, 413). I've also tried to quantize model, which reduced it size from 6.9MB to 1.8MB, but time increased from 7.5ms to 11ms which also seems strange to me. I used low precision in my model's BackendConfig. Do you have any different advices what can I use to reduce time of inference if I cannot use GPU? (base) daniel@Daniel-PC:~/Desktop/MNN/build$ ./run_test.out speed/MatMulBConst |
|
Originally posted by @mingyunzzu in #673
Can somebody explain why inference with batch isn't more efficient in MNN? When I run detection on single image it takes 7 miliseconds and when I run on batch of 32 images it takes 8 miliseconds per image. This is only the time of inference measured by time of runSession without preparing images and postprocessing. What can I use to reach better results?
The text was updated successfully, but these errors were encountered: