About this Learning Path

Who is this for?

This is an advanced topic for developers looking to leverage Arm's Scalable Matrix Extension 2 (SME2) instructions to accelerate LiteRT model inference on Android.

What will you learn?

Upon completion of this Learning Path, you will be able to:

  • Understand how KleidiAI integrates with LiteRT
  • Build the LiteRT benchmark tool and enable XNNPACK and KleidiAI with SME2 support in LiteRT
  • Create LiteRT models that can be accelerated by SME2 through KleidiAI
  • Use the benchmark tool to evaluate and validate the SME2 acceleration performance of LiteRT models

Prerequisites

Before starting, you will need the following:

Next