LiteRT on Android provides essentials for deploying high performance, custom ML features into your Android app.
LiteRT for ML runtime
Use LiteRT with Google Play services, Android's official ML inference runtime, to run high-performance ML inference in your app. Learn more
Hardware Acceleration with LiteRT Delegates
Use LiteRT Delegates distributed using Google Play services to run accelerated ML on specialized hardware such as GPUs or NPUs. This helps you deliver more fluid, lower latency user experiences to your users by accessing advanced on-device compute capabilities.
We provide support for GPU delegates, and we're working with partners to provide access to their custom delegates using Google Play services to support advanced use cases. Learn more
Enabled by Google Play services
Use Google Play services to access the LiteRT runtime and delegates. This ensures use of the latest stable versions while minimizing impact to your app's binary size. Learn more
Code samples
Review the LiteRT Android code samples and test ML features on your device. Learn more
Get started with LiteRT
Download the example code and get started with LiteRT and Android. Learn more
Acceleration Service
The Acceleration Service API lets you to safely pick the optimal hardware acceleration configuration at runtime without having to worry about the underlying device hardware and drivers. Learn more