I've noticed that running my TFLite CNNs through the NNAPI on the Google Pixel 6 Pro (target device "google-edgetpu") while the camera is active causes large increases in inference runtimes (observed ~30-500% increase compared to inactive camera, depending on target model). I am trying to simultaneously display a camera preview to the screen and run frame-by-frame ML analysis on the camera feed, but this slowdown is a hindrance. I suspect the device is using the TPU for some image post-processing by default, but I have not been able to find a setting that can be turned off to fix this.
Any ideas on exactly what is happening and if/how it can be fixed?
Any ideas on exactly what is happening and if/how it can be fixed?
Last edited: