On device models
Asked on 2025-11-24
1 search
Apple's WWDC 2025 introduced the concept of on-device models, particularly focusing on large language models (LLMs) that are optimized to run directly on devices like iPhones. These models are designed to handle tasks such as summarization, extraction, and classification, but they are not intended for complex reasoning or tasks requiring extensive world knowledge, which are typically handled by larger server-based models.
The on-device model discussed is a large language model with 3 billion parameters, each quantized to 2 bits, making it significantly smaller than server-scale models. This allows for privacy benefits, as data does not need to be sent to a server, and the model can function offline. The Foundation Models framework, introduced at WWDC, provides developers with tools to integrate these models into their apps, leveraging Apple's machine learning APIs and tools.
For more detailed information on the on-device models, you can refer to the session Meet the Foundation Models framework (02:57) and Explore prompt design & safety for on-device foundation models (02:00).

Meet the Foundation Models framework
Learn how to tap into the on-device large language model behind Apple Intelligence! This high-level overview covers everything from guided generation for generating Swift data structures and streaming for responsive experiences, to tool calling for integrating data sources and sessions for context management. This session has no prerequisites.

Platforms State of the Union
Discover the newest advancements on Apple platforms.

Discover machine learning & AI frameworks on Apple platforms
Tour the latest updates to machine learning and AI frameworks available on Apple platforms. Whether you are an app developer ready to tap into Apple Intelligence, an ML engineer optimizing models for on-device deployment, or an AI enthusiast exploring the frontier of what is possible, we’ll offer guidance to help select the right tools for your needs.
