We take another look at Executorch and KleidAI. The source discusses advancements in on-device AI, specifically focusing on Large Language Model (LLM) inference for Meta's Llama 3.2 quantized models. It highlights the collaboration between Arm and Meta to integrate Arm's KleidiAI software library into PyTorch's ExecuTorch framework. This integration significantly boosts AI workload performance on Arm mobile CPUs, enabling faster and more efficient deployment of AI models on edge devices. The article details performance improvements, including increased tokens per second and reduced memory footprint, making powerful AI accessible on a wider range of mobile devices.