Description: LiteRT, successor to TensorFlow Lite. is Google's On-device framework for high-performance ML & GenAI deployment on edge platforms, via efficient conversion, runtime, and optimization
View google-ai-edge/LiteRT on GitHub ↗
LiteRT, developed by Google, is an on-device framework designed for high-performance machine learning (ML) and generative AI (GenAI) deployment on edge platforms. It serves as the successor to TensorFlow Lite, building upon its legacy to provide a more efficient and optimized solution for running AI models directly on devices. The primary purpose of LiteRT is to enable developers to bring the power of AI to the edge, offering a streamlined and performant experience for various hardware platforms.
The core functionality of LiteRT revolves around efficient model conversion, runtime execution, and optimization. It facilitates the deployment of ML and GenAI models on a wide range of devices, including smartphones, tablets, and embedded systems. Key features include advanced GPU/NPU acceleration, superior ML and GenAI performance, and a simplified developer experience. The framework offers a new LiteRT Compiled Model API, which streamlines development through automated accelerator selection, asynchronous execution, and efficient I/O buffer handling. This API simplifies the process of integrating AI models into applications, reducing development time and complexity.
One of the significant advancements in LiteRT is its unified NPU acceleration. It provides seamless access to NPUs from major chipset providers, ensuring a consistent developer experience across different hardware. This allows developers to leverage the specialized processing capabilities of NPUs for faster and more energy-efficient inference. Furthermore, LiteRT boasts best-in-class GPU performance, utilizing state-of-the-art GPU acceleration techniques. The framework's buffer interoperability minimizes latency and enables zero-copy operations across various GPU buffer types, further enhancing performance.
LiteRT supports a broad range of platforms, including Android, iOS, Linux, macOS, Windows, Web, and IoT devices. This cross-platform compatibility allows developers to target a wide audience and deploy their AI-powered applications on various devices. The framework provides CPU, GPU, and NPU support on these platforms, enabling developers to choose the optimal hardware acceleration for their specific needs. The documentation highlights the supported hardware, including Google Tensor, Qualcomm, MediaTek, and other chipsets for NPU acceleration.
The repository provides clear guidance on getting started with LiteRT, including installation instructions and various "Choose Your Adventure" paths. These paths cater to different developer needs, such as converting PyTorch models, running pre-trained models, and maximizing performance. For developers working with PyTorch models, LiteRT offers tools like the LiteRT Torch Converter and LiteRT Generative Torch API to facilitate model conversion and deployment. For those new to on-device ML, the framework provides step-by-step instructions and sample applications to help them get started quickly. Developers seeking to maximize performance can leverage the LiteRT API to accelerate their existing models. For those working with GenAI models, LiteRT-LM is available for efficient deployment.
The repository also outlines the project's roadmap, which includes expanding hardware acceleration, optimizing for GenAI models, improving developer tools, and enhancing platform support. The project actively encourages contributions and provides resources for getting help, including GitHub issues and discussions. LiteRT is part of a larger ecosystem of Google AI Edge tools, including LiteRT Samples, LiteRT Torch Converter, LiteRT-LM, XNNPACK, and MediaPipe, providing developers with a comprehensive suite of resources for on-device ML development. The project is licensed under the Apache-2.0 License and adheres to a Code of Conduct to foster a welcoming and collaborative community.
Fetching additional details & charts...