LiteRT (Lite Runtime), formerly known as TensorFlow Lite, is Google's high-performance runtime for on-device AI. It enables developers to deploy machine learning models across various platforms and microcontrollers. LiteRT supports models from TensorFlow, PyTorch, and JAX, converting them into the efficient FlatBuffers format (.tflite) for optimized on-device inference. Key features include low latency, enhanced privacy by processing data locally, reduced model and binary sizes, and efficient power consumption. The runtime offers SDKs in multiple languages such as Java/Kotlin, Swift, Objective-C, C++, and Python, facilitating integration into diverse applications. Hardware acceleration is achieved through delegates like GPU and iOS Core ML, improving performance on supported devices. LiteRT Next, currently in alpha, introduces a new set of APIs that streamline on-device hardware acceleration.