Recent content by Andrey Ignatov

  1. Andrey Ignatov

    What is the source repo/original model used for Section 23 Text Completion test?

    In terms of architecture, you can just check the TFLite models from AI Benchmark APK.
  2. Andrey Ignatov

    What is the source repo/original model used for Section 23 Text Completion test?

    No, that's not a typo - this model was adapted with some small modifications to the text completion task, being applied to word embeddings instead of images.
  3. Andrey Ignatov

    AI Benchmark 5.1.1 Mobile Released

    What's New: Updated Qualcomm QNN and MediaTek Neuron delegates. Enhanced stability and accuracy of the power consumption test. Various bug fixes and performance improvements. Download this release from the official website or from the Google Play store. Feel free to discuss AI Benchmark...
  4. Andrey Ignatov

    AI Benchmark V5 Scores Updates

    Detailed AI Benchmark V5 results were released for over 50 IoT, smartTV and automotive platforms: https://ai-benchmark.com/ranking_IoT https://ai-benchmark.com/ranking_IoT_detailed The results of the recently presented mobile chipsets including the Snapdragon 8 Gen 3, Dimensity 9300, Google...
  5. Andrey Ignatov

    Does aibenchmark update and adapt to AndroidT?

    Hi @Mountain, replied to you by email.
  6. Andrey Ignatov

    AI Benchmark 5.1.0 Mobile Released

    What's New: Added new NPU power consumption test. Updated TFLite runtime. Updated TFLite GPU, NNAPI, Qualcomm QNN, Hexagon NN and Samsung ENN delegates. Updated in-app ranking table. Various bug fixes and performance improvements. Download this release from the official website or from the...
  7. Andrey Ignatov

    Difference between the HTP and DSP delegate?

    This is a very brief answer, but the general idea is as follows: HTP = rebranded compute DSP (since Snapdragon 888 / Hexagon v68): contains HVX and HMX co-processors / modules. Note that both HVX and HMX modules are also present in other Hexagon DSPs without HTP. HTA = additional co-processor...
  8. Andrey Ignatov

    Is it possible to get source code of AI bench 5.0.3

    One can potentially extract all models directly from the benchmark APK file. Feel free to use this forum for sharing or comparing the results, such posts will not be deleted or banned.
  9. Andrey Ignatov

    More information about result consolidation

    Yes, average or median of the results after removing the outliers. For the majority of SoCs, the results are obtained based in phone measurements, but in some cases development kits are also used (e.g., when no actual devices have been released yet). No, the SoC ranking is not taking into...
  10. Andrey Ignatov

    Difference between the HTP and DSP delegate?

    Yes, partly: the Hexagon 6xx family is denoted as DSPs in QNN, while the Hexagon 7xx family - as HTPs. Here you can find the full list of Hexagon processors. There are also large architectural differences between these two families - the latest HTPs, for instance, are able to accelerate both...
  11. Andrey Ignatov

    Questions about Apple Benchmarks

    INT8 models were running with the TFLite GPU delegate. No, these are plain NPU/GPU runtime results. Because of the bug in the iOS TFLite implementation.
  12. Andrey Ignatov

    Power Efficiency Measurements & Inference Precision

    In the standard benchmark mode, only INT8 inference is tested. However, one can also check the results of FP16 inference in the PRO mode. You can switch between different NPU inference profiles in the settings, sustained speed is used by default.
  13. Andrey Ignatov

    Reference for error/accuracy

    Hi @bagofwater, Thank you for your suggestions. For FP16 inference, the targets are generated in FP32 mode that provides an accuracy of 7-8 digits after decimal point, so there are no issues here.
  14. Andrey Ignatov

    Can't run very a basic model with GPU delegate. Problem with my conversion?

    Yes, your model should have only one input layer in order to be executed successfully. The easiest workaround here would be to stack two input tensors together into a single input layer and then unstack them during inference.
  15. Andrey Ignatov

    How to access APU in Mediatek chipsets

    Right now, it is not possible to force enable delegates when running custom models, this functionality will be added in the next benchmark version.
Top