Mobile device : OnePlus 7T
APP version: 5.0.3.
TF version: 2.11
I have converted my pretrained model to an int8 quantized tflite model, but I encountered some problems as shown in following shotcuts. I suspect that the type of input data (maybe uint8/int8) conflicts with the required float32...