Recent content by noodles

  1. N

    much higher latencies with NNAPI on Snapdragon 888

    I also run the quantized fsrcnn from MAI-2021, the latency is still much higher than CPU, GPU. But the models come with AI benchmark can run very fast with NNAPI-int8
  2. N

    much higher latencies with NNAPI on Snapdragon 888

    by the way, my input is big, a 245 * 530 image, and output a 980 * 2120 image. And I use Netron to check the model to make sure the input, output and weights are int8 numbers.
  3. N

    much higher latencies with NNAPI on Snapdragon 888

    CPU-int8 CPU-FP16(ms) CPU-FP32(ms) GPU-int8(ms) GPU-FP16(ms) GPU-FP32(ms) NNAPI-int8(ms) NNAPI-FP16(ms) NNAPI-FP32(ms) TF-lite(float) Unsupported 218 220 Unsupported 30 31 Unsupported 22 41 TF-lite(weight quantized) Unsupported 169 163 Unsupported 19 32 Unsupported 194 156...
Top