I also run the quantized fsrcnn from MAI-2021, the latency is still much higher than CPU, GPU. But the models come with AI benchmark can run very fast with NNAPI-int8
by the way, my input is big, a 245 * 530 image, and output a 980 * 2120 image.
And I use Netron to check the model to make sure the input, output and weights are int8 numbers.