- Mobile device : OnePlus 7T
- APP version: 5.0.3.
- TF version: 2.11
I have converted my pretrained model to an int8 quantized tflite model, but I encountered some problems as shown in following shotcuts. I suspect that the type of input data (maybe uint8/int8) conflicts with the required float32 format for input data.
Can anyone provide some help or advice? Thank you in advance!
- Inputs and Outputs information of my tflite model: