CPU-int8 | CPU-FP16(ms) | CPU-FP32(ms) | GPU-int8(ms) | GPU-FP16(ms) | GPU-FP32(ms) | NNAPI-int8(ms) | NNAPI-FP16(ms) | NNAPI-FP32(ms) | |
TF-lite(float) | Unsupported | 218 | 220 | Unsupported | 30 | 31 | Unsupported | 22 | 41 |
TF-lite(weight quantized) | Unsupported | 169 | 163 | Unsupported | 19 | 32 | Unsupported | 194 | 156 |
TF-lite(full quantized) | 149 | Unsupported | Unsupported | 25 | Unsupported | Unsupported | 8710 | Unsupported | Unsupported |
I run a very simple model ESPCN for super resolution, it is just a 3-layer convolution network. The results are very strange. Is there something wrong with quantization or other things? If the reason is quantization, how to deal with it?
Last edited: