You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Can somebody explain why inference with batch isn't more efficient in MNN? When I run detection on single image it takes 7 miliseconds and when I run on batch of 32 images it takes 8 miliseconds per image. This is only the time of inference measured by time of runSession without preparing images and postprocessing. What can I use to reach better results?
The text was updated successfully, but these errors were encountered:
The issue's bug has been resolved. Now batch inference time depend on the compute flops of device. If one image has reach the peak flops, then batch image will not more efficient.
The get the device's compute peak flops, can use ./run_test.out speed/MatMulBConst
Normally GPU has more flops than CPU, you can use opencl instead of CPU to forward.
jxt1234
added
the
User
The user ask question about how to use. Or don't use MNN correctly and cause bug.
label
Jan 29, 2025
Originally posted by @mingyunzzu in #673
Can somebody explain why inference with batch isn't more efficient in MNN? When I run detection on single image it takes 7 miliseconds and when I run on batch of 32 images it takes 8 miliseconds per image. This is only the time of inference measured by time of runSession without preparing images and postprocessing. What can I use to reach better results?
The text was updated successfully, but these errors were encountered: