qiuzhangTiTi
New member
According to my current knowledge, the float32 or float16 TF Lite model should run faster than the uint8 model on a GPU device. So there is no need to perform int8 quantization for image denoising task which running on Exynos mail GPU. Is this correct? How about the image super-resolution task (NPU)? Thanks.