What devices will support the on-device LLM?

Asked on 06/12/2025

1 search

The on-device large language model (LLM) is designed to run on Apple devices that are powered by Apple Silicon. This includes devices such as iPhones, iPads, and Macs. The model is optimized to take advantage of the CPU, GPU, and neural engine available on these devices, allowing for efficient and low-latency inference while keeping user data private. The on-device LLM is a device-scale model with around 3 billion parameters, making it suitable for tasks like summarization, extraction, and classification, but not for advanced reasoning or tasks requiring extensive world knowledge.